1
|
Pang L, He P, Han Y, Cui H, Feng P, Zhang C, Huang P, Tian S. Semantic Consistency Network with Edge Learner and Connectivity Enhancer for Cervical Tumor Segmentation from Histopathology Images. Interdiscip Sci 2025:10.1007/s12539-025-00691-w. [PMID: 40268829 DOI: 10.1007/s12539-025-00691-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2024] [Revised: 01/14/2025] [Accepted: 01/17/2025] [Indexed: 04/25/2025]
Abstract
Accurate tumor grading and regional identification of cervical tumors are important for diagnosis and prognosis. Traditional manual microscopy methods suffer from time-consuming, labor-intensive, and subjective bias problems, so tumor segmentation methods based on deep learning are gradually becoming a hotspot in current research. Cervical tumors have diverse morphologies, which leads to low similarity between the mask edge and ground-truth edge of existing semantic segmentation models. Moreover, the texture and geometric arrangement features of normal tissues and tumors are highly similar, which causes poor pixel connectivity in the mask of the segmentation model. To this end, we propose an end-to-end semantic consistency network with the edge learner and the connectivity enhancer, i.e., ERNet. First, the edge learner consists of a stacked shallow convolutional neural network, so it can effectively enhance the ability of ERNet to learn and represent polymorphic tumor edges. Second, the connectivity enhancer learns detailed information and contextual information of tumor images, so it can enhance the pixel connectivity of the masks. Finally, edge features and pixel-level features are adaptively coupled, and the segmentation results are additionally optimized by the tumor classification task as a whole. The results show that, compared with those of other state-of-the-art segmentation models, the structural similarity and the mean intersection over union of ERNet are improved to 88.17% and 83.22%, respectively, which reflects the excellent edge similarity and pixel connectivity of the proposed model. Finally, we conduct a generalization experiment on laryngeal tumor images. Therefore, the ERNet network has good clinical popularization and practical value.
Collapse
Affiliation(s)
- Lisha Pang
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China
| | - Peng He
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China.
| | - Yue Han
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China
| | - Hao Cui
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China
| | - Peng Feng
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China
| | - Chi Zhang
- School of Physics and Optoelectronic Engineering, Hangzhou Institute for Advanced Study, University of Chinese Academy of Sciences, Hangzhou, 310024, China.
| | - Pan Huang
- Key Laboratory of Optoelectronic Technology and Systems (Ministry of Education), College of Optoelectronic Engineering, Chongqing University, Chongqing, 400044, China
| | - Sukun Tian
- School and Hospital of Stomatology, Peking University, Beijing, 100089, China
| |
Collapse
|
2
|
Wenbo L, Tingting L, Xiao L. ModuCLIP: multi-scale CLIP framework for predicting foundation pit deformation in multi-modal robotic systems. Front Neurorobot 2025; 19:1544694. [PMID: 40236467 PMCID: PMC11996866 DOI: 10.3389/fnbot.2025.1544694] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2024] [Accepted: 02/17/2025] [Indexed: 04/17/2025] Open
Abstract
Introduction Foundation pit deformation prediction is a critical aspect of underground engineering safety assessment, influencing construction quality and personnel safety. However, due to complex geological conditions and numerous environmental interference factors, traditional prediction methods struggle to achieve precise modeling. Conventional approaches, including numerical simulations, empirical formulas, and machine learning models, suffer from limitations such as high computational costs, poor generalization, or excessive dependence on specific data distributions. Recently, deep learning models, particularly cross-modal architectures, have demonstrated great potential in engineering applications. However, effectively integrating multi-modal data for improved prediction accuracy remains a significant challenge. Methods This study proposes a Multi-Scale Contrastive Language-Image Pretraining (CLP) framework, ModuCLIP, designed for foundation pit deformation prediction in multi-modal robotic systems. The framework leverages a self-supervised contrastive learning mechanism to integrate multi-source information, including images, textual descriptions, and sensor data, while employing a multi-scale feature learning approach to enhance adaptability to complex conditions. Experiments conducted on multiple foundation pit engineering datasets demonstrate that ModuCLIP outperforms existing methods in terms of prediction accuracy, generalization, and robustness. Results and discussion The findings suggest that this framework provides an efficient and precise solution for foundation pit deformation prediction while offering new insights into multi-modal robotic perception and engineering monitoring applications.
Collapse
Affiliation(s)
- Lin Wenbo
- School of Geology, Gansu Industrial Vocational and Technical College, Tianshui, Gansu, China
| | - Li Tingting
- School of Electronic Information, Gansu Industrial Vocational and Technical College, Tianshui, Gansu, China
| | - Li Xiao
- Guangdong Nonferrous Industry Building Quality Inspection Co., Ltd., Guangzhou, Guangdong, China
| |
Collapse
|
3
|
Fang X, Chong CF, Wong KL, Simões M, Ng BK. Investigating the key principles in two-step heterogeneous transfer learning for early laryngeal cancer identification. Sci Rep 2025; 15:2146. [PMID: 39820368 PMCID: PMC11739633 DOI: 10.1038/s41598-024-84836-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2024] [Accepted: 12/27/2024] [Indexed: 01/19/2025] Open
Abstract
Data scarcity in medical images makes transfer learning a common approach in computer-aided diagnosis. Some disease classification tasks can rely on large homogeneous public datasets to train the transferred model, while others cannot, i.e., endoscopic laryngeal cancer image identification. Distinguished from most current works, this work pioneers exploring a two-step heterogeneous transfer learning (THTL) framework for laryngeal cancer identification and summarizing the fundamental principles for the intermediate domain selection. For heterogeneity and clear vascular representation, diabetic retinopathy images were chosen as THTL's intermediate domain. The experiment results reveal two vital principles in intermediate domain selection for future studies: 1) the size of the intermediate domain is not a sufficient condition to improve the transfer learning performance; 2) even distinct vascular features in the intermediate domain do not guarantee improved performance in the target domain. We observe that radial vascular patterns benefit benign classification, whereas twisted and tangled patterns align more with malignant classification. Additionally, to compensate for the absence of twisted patterns in the intermediate domains, we propose the Step-Wise Fine-Tuning (SWFT) technique, guided by the Layer Class Activate Map (LayerCAM) visualization result, getting 20.4% accuracy increases compared to accuracy from THTL's, even higher than fine-tune all layers.
Collapse
Affiliation(s)
- Xinyi Fang
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Chak Fong Chong
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Kei Long Wong
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China
- Department of Computer Science and Engineering, University of Bologna, Bologna, 40100, Italy
| | - Marco Simões
- Department of Informatics Engineering, Centre for Informatics and Systems of the University of Coimbra, University of Coimbra, Coimbra, 3000, Portugal
| | - Benjamin K Ng
- Faculty of Applied Sciences, Macao Polytechnic University, Macao, 999078, China.
| |
Collapse
|
4
|
Deng X, Luo J, Huang P, He P, Li J, Liu Y, Xiao H, Feng P. MCRANet: MTSL-based connectivity region attention network for PD-L1 status segmentation in H&E stained images. Comput Biol Med 2025; 184:109357. [PMID: 39531922 DOI: 10.1016/j.compbiomed.2024.109357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2024] [Revised: 11/03/2024] [Accepted: 11/04/2024] [Indexed: 11/16/2024]
Abstract
The quantitative analysis of Programmed death-ligand 1 (PD-L1) via Immunohistochemical (IHC) plays a crucial role in guiding immunotherapy. However, IHC faces challenges, including high costs, time consumption and result variability. Conversely, Hematoxylin-Eosin (H&E) staining offers cost-effectiveness, speed, and stable results. Nonetheless, H&E staining, which solely visualizes cellular morphological features, lacks clinical applicability in detecting biomarker expressions like PD-L1. Substituting H&E staining for IHC in determining PD-L1 status is a clinically significant and challenging task. Motivated by above observations, we propose a Multi-Task supervised learning (MTSL)-based connectivity region attention network (MCRANet) for PD-L1 status segmentation in H&E stained images. To reduce interference from non-tumor areas, the MTSL-based region attention is proposed to enhances the network's capability to distinguish between tumor and non-tumor regions. Consequently, this augmentation further improves the network's segmentation efficacy for PD-L1 positive and negative regions. Furthermore, the PD-L1 expression regions demonstrate interconnection throughout the tissue section. Leveraging this topological prior knowledge, we integrate a connectivity modeling module (CM module) within the MTSL-based region attention module (MRA module) to enhance the precision of MTSL-based region attention localization. This integration further improves the structural similarity between the segmentation results and the ground truth. Extensive visual and quantitative results demonstrate that our supervised-learning-guided MRA module produces more interpretable attention and the introduced CM module provides accurate positional attention to the MRA module. Compared to other state-of-the-art networks, MCRANet exhibits superior segmentation performance with a dice similarity coefficient (DSC) of 79.6 % on the lung squamous cell carcinoma (LUSC) PD-L1 status dataset.
Collapse
Affiliation(s)
- Xixiang Deng
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China
| | - Jiayang Luo
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China
| | - Pan Huang
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China.
| | - Peng He
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China.
| | - Jiahao Li
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China
| | - Yanan Liu
- Department of Electronic Information Engineering, Chongqing Technology and Business and Institute, 401520, Chongqing, China
| | - Hualiang Xiao
- Department of Pathology, Daping Hospital, Army Medical University, 400037, Chongqing, China
| | - Peng Feng
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, 400044, Chongqing, China.
| |
Collapse
|
5
|
Liu H, Cao R, Li S, Wang Y, Zhang X, Xu H, Sun X, Wang L, Qian P, Sun Z, Gao K, Li F. ViT-Based Face Diagnosis Images Analysis for Schizophrenia Detection. Brain Sci 2024; 15:30. [PMID: 39851398 PMCID: PMC11763813 DOI: 10.3390/brainsci15010030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2024] [Revised: 12/20/2024] [Accepted: 12/26/2024] [Indexed: 01/26/2025] Open
Abstract
OBJECTIVES Computer-aided schizophrenia (SZ) detection methods mainly depend on electroencephalogram and brain magnetic resonance images, which both capture physical signals from patients' brains. These inspection techniques take too much time and affect patients' compliance and cooperation, while difficult for clinicians to comprehend the principle of detection decisions. This study proposes a novel method using face diagnosis images based on traditional Chinese medicine principles, providing a non-invasive, efficient, and interpretable alternative for SZ detection. METHODS An innovative face diagnosis image analysis method for SZ detection, which learns feature representations based on Vision Transformer (ViT) directly from face diagnosis images. It provides a face features distribution visualization and quantitative importance of each facial region and is proposed to supplement interpretation and to increase efficiency in SZ detection while keeping a high detection accuracy. RESULTS A benchmarking platform comprising 921 face diagnostic images, 6 benchmark methods, and 4 evaluation metrics was established. The experimental results demonstrate that our method significantly improves SZ detection performance with a 3-10% increase in accuracy scores. Additionally, it is found that facial regions rank in descending order according to importance in SZ detection as eyes, mouth, forehead, cheeks, and nose, which is exactly consistent with the clinical traditional Chinese medicine experience. CONCLUSIONS Our method fully leverages semantic feature representations of first-introduced face diagnosis images in SZ, offering strong interpretability and visualization capabilities. It not only opens a new path for SZ detection but also brings new tools and concepts to the research and application in the field of mental illness.
Collapse
Affiliation(s)
- Huilin Liu
- School of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China; (H.L.); (L.W.); (P.Q.); (Z.S.)
| | - Runmin Cao
- State Key Laboratory of Intelligent Technology and Systems, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China; (R.C.); (S.L.); (Y.W.); (X.Z.)
| | - Songze Li
- State Key Laboratory of Intelligent Technology and Systems, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China; (R.C.); (S.L.); (Y.W.); (X.Z.)
- School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China;
| | - Yifan Wang
- State Key Laboratory of Intelligent Technology and Systems, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China; (R.C.); (S.L.); (Y.W.); (X.Z.)
- School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China;
| | - Xiaohan Zhang
- State Key Laboratory of Intelligent Technology and Systems, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China; (R.C.); (S.L.); (Y.W.); (X.Z.)
- School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China;
| | - Hua Xu
- State Key Laboratory of Intelligent Technology and Systems, Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China; (R.C.); (S.L.); (Y.W.); (X.Z.)
| | - Xirong Sun
- Shanghai Pudong New Area Mental Health Center, Tongji University, Shanghai 200124, China
| | - Lijuan Wang
- School of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China; (H.L.); (L.W.); (P.Q.); (Z.S.)
| | - Peng Qian
- School of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China; (H.L.); (L.W.); (P.Q.); (Z.S.)
| | - Zhumei Sun
- School of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China; (H.L.); (L.W.); (P.Q.); (Z.S.)
| | - Kai Gao
- School of Information Science and Engineering, Hebei University of Science and Technology, Shijiazhuang 050018, China;
| | - Fufeng Li
- School of Traditional Chinese Medicine, Shanghai University of Traditional Chinese Medicine, Shanghai 201203, China; (H.L.); (L.W.); (P.Q.); (Z.S.)
| |
Collapse
|
6
|
Wang Q, Deng X, Huang P, Ma Q, Zhao L, Feng Y, Wang Y, Zhao Y, Chen Y, Zhong P, He P, Ma M, Feng P, Xiao H. Prediction of PD-L1 tumor positive score in lung squamous cell carcinoma with H&E staining images and deep learning. Front Artif Intell 2024; 7:1452563. [PMID: 39759385 PMCID: PMC11695341 DOI: 10.3389/frai.2024.1452563] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2024] [Accepted: 12/10/2024] [Indexed: 01/07/2025] Open
Abstract
Background Detecting programmed death ligand 1 (PD-L1) expression based on immunohistochemical (IHC) staining is an important guide for the treatment of lung cancer with immune checkpoint inhibitors. However, this method has problems such as high staining costs, tumor heterogeneity, and subjective differences among pathologists. Therefore, the application of deep learning models to segment and quantitatively predict PD-L1 expression in digital sections of Hematoxylin and eosin (H&E) stained lung squamous cell carcinoma is of great significance. Methods We constructed a dataset comprising H&E-stained digital sections of lung squamous cell carcinoma and used a Transformer Unet (TransUnet) deep learning network with an encoder-decoder design to segment PD-L1 negative and positive regions and quantitatively predict the tumor cell positive score (TPS). Results The results showed that the dice similarity coefficient (DSC) and intersection overunion (IoU) of deep learning for PD-L1 expression segmentation of H&E-stained digital slides of lung squamous cell carcinoma were 80 and 72%, respectively, which were better than the other seven cutting-edge segmentation models. The root mean square error (RMSE) of quantitative prediction TPS was 26.8, and the intra-group correlation coefficients with the gold standard was 0.92 (95% CI: 0.90-0.93), which was better than the consistency between the results of five pathologists and the gold standard. Conclusion The deep learning model is capable of segmenting and quantitatively predicting PD-L1 expression in H&E-stained digital sections of lung squamous cell carcinoma, which has significant implications for the application and guidance of immune checkpoint inhibitor treatments. And the link to the code is https://github.com/Baron-Huang/PD-L1-prediction-via-HE-image.
Collapse
Affiliation(s)
- Qiushi Wang
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Xixiang Deng
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, Chongqing, China
| | - Pan Huang
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, Chongqing, China
| | - Qiang Ma
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Lianhua Zhao
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Yangyang Feng
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Yiying Wang
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Yuan Zhao
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Yan Chen
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Peng Zhong
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| | - Peng He
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, Chongqing, China
| | - Mingrui Ma
- Department of Information, Affiliated Tumor Hospital of Xinjiang Medical University, Urumchi, China
| | - Peng Feng
- The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, Chongqing, China
| | - Hualiang Xiao
- Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
| |
Collapse
|
7
|
Chang CW, Yang ST, Liu HH, Chang WH, Lee WL, Wang PH. Surgery-based radiation-free multimodality treatment for locally advanced cervical cancer. Taiwan J Obstet Gynecol 2024; 63:651-664. [PMID: 39266145 DOI: 10.1016/j.tjog.2024.07.014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/31/2024] [Indexed: 09/14/2024] Open
Abstract
The current review described a 55-year woman using 28 months to finish her surgery-based radiation-free multimodality treatment journey to fight International Federation of Gynaecology & Obstetrics (FIGO) 2018 clinical stage IIA2 (cT2aN0M0) squamous cell carcinoma (SCC) of the cervix. She received six cycles of perioperative adjuvant therapy, including three cycles of neoadjuvant therapy (NAT) and three cycles of postoperative adjuvant therapy by using combination of dose-dense chemotherapy (CT, weekly paclitaxel 80 mg/m2+triweekly cisplatin 40 mg/m2), immunotherapy (IO, triweekly pembrolizumab 200 mg) and half-dose anti-angiogenic agent (triweekly bevacizumab 7.5 mg/kg) plus interval radical surgery (radical hysterectomy + bilateral salpingo-oophorectomy + bilateral pelvic lymph node dissection + para-aortic lymph node sampling) and following maintenance therapy with monthly 22 cycles of half-dose of IO (pembrolizumab 100 mg) and concomitant 4 cycles of single-agent CT (paclitaxel 175 mg/m2) and 18 cycles of half-dose anti-angiogenic agent (bevacizumab 7.5 mg/kg). During the cervical SCC fighting journey, two unwanted adverse events (AEs) occurred. One was pseudo-progressive disease during the NAT treatment and pathology-confirmed upgrading FIGO stage IIIC1p (ypT2a1N1M0) after radical surgery and the other was the occurrence of hypothyroidism during the post operative adjuvant therapy. Based on this case we presented, we review the recent trend in the management of women with locally advanced cervical cancer (LACC) using the radiation-free but surgery-based multimodality strategy and highlight the strengths and limitations about perioperative adjuvant therapy with dose-dense CT + IO + half-dose anti-angiogenic agent and maintenance treatment of half-dose IO combining with short-term single agent CT and following long-term half-dose anti-angiogenic agent. All underscore the possibility that women with LACC have an opportunity to receive surgery-based RT-free multi-modality strategy to manage their diseases with satisfactory results. Additionally, the evolving role of IO plus CT with/without anti-angiogenic agent functioning as either primary treatment or adjuvant therapy for the treatment of advanced CC has been in process continuously. Moreover, the patient's positive response to IO, pembrolizumab as an example, both during the primary and maintenance therapy, highlights the importance of integrating IO into CT regimens for CC, especially in cases where conventional therapies, RT as an example, are insufficient or who do not want to receive RT-based treatment. The sustained disease-free status of the patient over several years reinforces the potential of IO to significantly increase long-term survival outcomes in CC patients, particularly for those with LACC.
Collapse
Affiliation(s)
- Che-Wei Chang
- Department of Obstetrics and Gynecology, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Szu-Ting Yang
- Department of Obstetrics and Gynecology, Taipei Veterans General Hospital, Taipei, Taiwan; Institute of Clinical Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Hung-Hsien Liu
- Department of Medical Imaging and Intervention, Tucheng Hospital, New Taipei City, Taiwan
| | - Wen-Hsun Chang
- Department of Obstetrics and Gynecology, Taipei Veterans General Hospital, Taipei, Taiwan; Female Cancer Foundation, Taipei, Taiwan; Department of Nursing, Taipei Veterans General Hospital, Taipei, Taiwan
| | - Wen-Ling Lee
- Institute of Clinical Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Female Cancer Foundation, Taipei, Taiwan; Department of Medicine, Cheng-Hsin General Hospital, Taipei, Taiwan.
| | - Peng-Hui Wang
- Department of Obstetrics and Gynecology, Taipei Veterans General Hospital, Taipei, Taiwan; Institute of Clinical Medicine, National Yang Ming Chiao Tung University, Taipei, Taiwan; Female Cancer Foundation, Taipei, Taiwan; Department of Medical Research, China Medical University Hospital, Taichung, Taiwan.
| |
Collapse
|
8
|
Alzakari SA, Maashi M, Alahmari S, Arasi MA, Alharbi AAK, Sayed A. Towards laryngeal cancer diagnosis using Dandelion Optimizer Algorithm with ensemble learning on biomedical throat region images. Sci Rep 2024; 14:19713. [PMID: 39181918 PMCID: PMC11344795 DOI: 10.1038/s41598-024-70525-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2024] [Accepted: 08/19/2024] [Indexed: 08/27/2024] Open
Abstract
Laryngeal cancer exhibits a notable global health burden, with later-stage detection contributing to a low mortality rate. Laryngeal cancer diagnosis on throat region images is a pivotal application of computer vision (CV) and medical image diagnoses in the medical sector. It includes detecting and analysing abnormal or cancerous tissue from the larynx, an integral part of the vocal and respiratory systems. The computer-aided system makes use of artificial intelligence (AI) through deep learning (DL) and machine learning (ML) models, including convolution neural networks (CNN), for automated disease diagnoses and detection. Various DL and ML approaches are executed to categorize the extraction feature as healthy and cancerous tissues. This article introduces an automated Laryngeal Cancer Diagnosis using the Dandelion Optimizer Algorithm with Ensemble Learning (LCD-DOAEL) method on Biomedical Throat Region Image. The LCD-DOAEL method aims to investigate the images of the throat region for the presence of laryngeal cancer. In the LCD-DOAEL method, the Gaussian filtering (GF) approach is applied to eliminate the noise in the biomedical images. Besides, the complex and intrinsic feature patterns can be extracted by the MobileNetv2 model. Meanwhile, the DOA model carries out the hyperparameter selection of MobileNetV2 architecture. Finally, the ensemble of three classifiers such as bidirectional long short-term memory (BiLSTM), regularized extreme learning machine (ELM), and backpropagation neural network (BPNN) models, are utilized for the classification process. A comprehensive set of simulations is conducted on the biomedical image dataset to highlight the efficient performance of the LCD-DOAEL technique. The comparison analysis of the LCD-DOAEL method exhibited a superior accuracy outcome of 97.54% over other existing techniques.
Collapse
Affiliation(s)
- Sarah A Alzakari
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, 11671, Riyadh, Saudi Arabia
| | - Mashael Maashi
- Department of Software Engineering, College of Computer and Information Sciences, King Saud University, Po Box 103786, 11543, Riyadh, Saudi Arabia
| | - Saad Alahmari
- Department of Computer Science, Applied College, Northern Border University, Arar, Saudi Arabia.
| | - Munya A Arasi
- Department of Computer Science, Applied College at RijalAlmaa, King Khalid University, Abha, Saudi Arabia
| | - Abeer A K Alharbi
- Department Information Systems, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), 11432, Riyadh, Saudi Arabia
| | - Ahmed Sayed
- Research Center, Future University in Egypt, New Cairo, 11835, Egypt
| |
Collapse
|
9
|
Ekta, Bhatia V. Auto-BCS: A Hybrid System for Real-Time Breast Cancer Screening from Pathological Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:1752-1766. [PMID: 38429562 PMCID: PMC11300416 DOI: 10.1007/s10278-024-01056-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/14/2023] [Revised: 12/24/2023] [Accepted: 01/14/2024] [Indexed: 03/03/2024]
Abstract
Breast cancer is recognized as a prominent cause of cancer-related mortality among women globally, emphasizing the critical need for early diagnosis resulting improvement in survival rates. Current breast cancer diagnostic procedures depend on manual assessments of pathological images by medical professionals. However, in remote or underserved regions, the scarcity of expert healthcare resources often compromised the diagnostic accuracy. Machine learning holds great promise for early detection, yet existing breast cancer screening algorithms are frequently characterized by significant computational demands, rendering them unsuitable for deployment on low-processing-power mobile devices. In this paper, a real-time automated system "Auto-BCS" is introduced that significantly enhances the efficiency of early breast cancer screening. The system is structured into three distinct phases. In the initial phase, images undergo a pre-processing stage aimed at noise reduction. Subsequently, feature extraction is carried out using a lightweight and optimized deep learning model followed by extreme gradient boosting classifier, strategically employed to optimize the overall performance and prevent overfitting in the deep learning model. The system's performance is gauged through essential metrics, including accuracy, precision, recall, F1 score, and inference time. Comparative evaluations against state-of-the-art algorithms affirm that Auto-BCS outperforms existing models, excelling in both efficiency and processing speed. Computational efficiency is prioritized by Auto-BCS, making it particularly adaptable to low-processing-power mobile devices. Comparative assessments confirm the superior performance of Auto-BCS, signifying its potential to advance breast cancer screening technology.
Collapse
Affiliation(s)
- Ekta
- Netaji Subhas University of Technology, Delhi, India
| | | |
Collapse
|
10
|
Shi J, Shu T, Wu K, Jiang Z, Zheng L, Wang W, Wu H, Zheng Y. Masked hypergraph learning for weakly supervised histopathology whole slide image classification. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 253:108237. [PMID: 38820715 DOI: 10.1016/j.cmpb.2024.108237] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/06/2024] [Revised: 05/16/2024] [Accepted: 05/20/2024] [Indexed: 06/02/2024]
Abstract
BACKGROUND AND OBJECTIVES Graph neural network (GNN) has been extensively used in histopathology whole slide image (WSI) analysis due to the efficiency and flexibility in modelling relationships among entities. However, most existing GNN-based WSI analysis methods only consider the pairwise correlation of patches from one single perspective (e.g. spatial affinity or embedding similarity) yet ignore the intrinsic non-pairwise relationships present in gigapixel WSI, which are likely to contribute to feature learning and downstream tasks. The objective of this study is therefore to explore the non-pairwise relationships in histopathology WSI and exploit them to guide the learning of slide-level representations for better classification performance. METHODS In this paper, we propose a novel Masked HyperGraph Learning (MaskHGL) framework for weakly supervised histopathology WSI classification. Compared with most GNN-based WSI classification methods, MaskHGL exploits the non-pairwise correlations between patches with hypergraph and global message passing conducted by hypergraph convolution. Concretely, multi-perspective hypergraphs are first built for each WSI, then hypergraph attention is introduced into the jointed hypergraph to propagate the non-pairwise relationships and thus yield more discriminative node representation. More importantly, a masked hypergraph reconstruction module is devised to guide the hypergraph learning which can generate more powerful robustness and generalization than the method only using hypergraph modelling. Additionally, a self-attention-based node aggregator is also applied to explore the global correlation of patches in WSI and produce the slide-level representation for classification. RESULTS The proposed method is evaluated on two public TCGA benchmark datasets and one in-house dataset. On the public TCGA-LUNG (1494 WSIs) and TCGA-EGFR (696 WSIs) test set, the area under receiver operating characteristic (ROC) curve (AUC) were 0.9752±0.0024 and 0.7421±0.0380, respectively. On the USTC-EGFR (754 WSIs) dataset, MaskHGL achieved significantly better performance with an AUC of 0.8745±0.0100, which surpassed the second-best state-of-the-art method SlideGraph+ 2.64%. CONCLUSIONS MaskHGL shows a great improvement, brought by considering the intrinsic non-pairwise relationships within WSI, in multiple downstream WSI classification tasks. In particular, the designed masked hypergraph reconstruction module promisingly alleviates the data scarcity and greatly enhances the robustness and classification ability of our MaskHGL. Notably, it has shown great potential in cancer subtyping and fine-grained lung cancer gene mutation prediction from hematoxylin and eosin (H&E) stained WSIs.
Collapse
Affiliation(s)
- Jun Shi
- School of Software, Hefei University of Technology, Hefei, 230601, Anhui Province, China
| | - Tong Shu
- School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, 230601, Anhui Province, China
| | - Kun Wu
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 102206, China
| | - Zhiguo Jiang
- Image Processing Center, School of Astronautics, Beihang University, Beijing, 102206, China; Tianmushan Laboratory, Hangzhou, 311115, Zhejiang Province, China
| | - Liping Zheng
- School of Software, Hefei University of Technology, Hefei, 230601, Anhui Province, China
| | - Wei Wang
- Department of Pathology, the First Affiliated Hospital of USTC, Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230036, Anhui Province, China; Intelligent Pathology Institute, Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230036, Anhui Province, China
| | - Haibo Wu
- Department of Pathology, the First Affiliated Hospital of USTC, Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230036, Anhui Province, China; Intelligent Pathology Institute, Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230036, Anhui Province, China
| | - Yushan Zheng
- School of Engineering Medicine, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China.
| |
Collapse
|
11
|
Fang J. A semi-supervised segmentation method for microscopic hyperspectral pathological images based on multi-consistency learning. Front Oncol 2024; 14:1396887. [PMID: 38962265 PMCID: PMC11220190 DOI: 10.3389/fonc.2024.1396887] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Accepted: 04/15/2024] [Indexed: 07/05/2024] Open
Abstract
Pathological images are considered the gold standard for clinical diagnosis and cancer grading. Automatic segmentation of pathological images is a fundamental and crucial step in constructing powerful computer-aided diagnostic systems. Medical microscopic hyperspectral pathological images can provide additional spectral information, further distinguishing different chemical components of biological tissues, offering new insights for accurate segmentation of pathological images. However, hyperspectral pathological images have higher resolution and larger area, and their annotation requires more time and clinical experience. The lack of precise annotations limits the progress of research in pathological image segmentation. In this paper, we propose a novel semi-supervised segmentation method for microscopic hyperspectral pathological images based on multi-consistency learning (MCL-Net), which combines consistency regularization methods with pseudo-labeling techniques. The MCL-Net architecture employs a shared encoder and multiple independent decoders. We introduce a Soft-Hard pseudo-label generation strategy in MCL-Net to generate pseudo-labels that are closer to real labels for pathological images. Furthermore, we propose a multi-consistency learning strategy, treating pseudo-labels generated by the Soft-Hard process as real labels, by promoting consistency between predictions of different decoders, enabling the model to learn more sample features. Extensive experiments in this paper demonstrate the effectiveness of the proposed method, providing new insights for the segmentation of microscopic hyperspectral tissue pathology images.
Collapse
Affiliation(s)
- Jinghui Fang
- College of Information Science and Engineering, Hohai University, Nanjing, China
| |
Collapse
|
12
|
Velimirovici MD, Feier CVI, Vonica RC, Faur AM, Muntean C. Efficacy and Safety of Atezolizumab as a PD-L1 Inhibitor in the Treatment of Cervical Cancer: A Systematic Review. Biomedicines 2024; 12:1291. [PMID: 38927498 PMCID: PMC11200956 DOI: 10.3390/biomedicines12061291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Revised: 06/07/2024] [Accepted: 06/10/2024] [Indexed: 06/28/2024] Open
Abstract
BACKGROUND AND OBJECTIVES The efficacy and safety of PD-L1 inhibitors in the treatment of cervical cancer is an ongoing research question. This review aims to establish a clear profile of atezolizumab, examining its impact on survival outcomes, response rates, and safety measured by serious adverse events (SAEs). MATERIALS AND METHODS A literature search was conducted using PubMed, Scopus, and Web of Science, focusing on articles published up to February 2024. The review followed the PRISMA guidelines and synthesized outcomes from four randomized trial studies involving atezolizumab administered at 1200 mg IV every three weeks, alone or in combination with chemoradiotherapy. RESULTS A total of 284 patients received atezolizumab, the majority being advanced stage cervical cancer (IVA-IVB). Median follow-up times ranged from 9 weeks to 32.9 months. It was found that combining atezolizumab with standard therapies extended median progression-free survival (PFS) from 10.4 to 13.7 months and overall survival (OS) from 22.8 to 32.1 months, according to the phase III trial. Monotherapy and initial treatment settings with atezolizumab also showed promising efficacy, with disease-free survival rates at 24 months reaching 79% compared to 52% with standard therapy alone. However, the treatment was associated with high rates of SAEs, reaching up to 79% in more intensive treatment combinations. CONCLUSIONS Atezolizumab demonstrates significant potential in improving PFS and OS in patients with cervical cancer, supporting its inclusion as a first-line treatment option. Despite the efficacy benefits, the high incidence of SAEs necessitates careful patient selection and management strategies to mitigate risks. This systematic review supports the continued evaluation of atezolizumab in broader clinical trials to refine its therapeutic profile and safety measures in the context of cervical cancer treatment.
Collapse
Affiliation(s)
- Milan Daniel Velimirovici
- Doctoral School, “Victor Babes” University of Medicine and Pharmacy, 2 E. Murgu Sq., 300041 Timisoara, Romania;
| | - Catalin Vladut Ionut Feier
- First Discipline of Surgery, Department X-Surgery, “Victor Babes” University of Medicine and Pharmacy, 2 E. Murgu Sq., 300041 Timisoara, Romania;
- First Surgery Clinic, “Pius Brinzeu” Clinical Emergency Hospital, 300723 Timisoara, Romania
| | - Razvan Constantin Vonica
- Preclinical Department, Discipline of Physiology, Faculty of General Medicine, “Lucian Blaga” University of Sibiu, 550169 Sibiu, Romania
| | - Alaviana Monique Faur
- Faculty of Medicine, Victor Babes” University of Medicine and Pharmacy, 2 E. Murgu Sq., 300041 Timisoara, Romania;
| | - Calin Muntean
- Medical Informatics and Biostatistics, Department III-Functional Sciences, “Victor Babes” University of Medicine and Pharmacy, 2 E. Murgu Sq., 300041 Timisoara, Romania;
| |
Collapse
|
13
|
Wang CH, Chang W, Lee MR, Tay J, Wu CY, Wu MC, Roth HR, Yang D, Zhao C, Wang W, Huang CH. Deep Learning-based Diagnosis of Pulmonary Tuberculosis on Chest X-ray in the Emergency Department: A Retrospective Study. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:589-600. [PMID: 38343228 PMCID: PMC11031502 DOI: 10.1007/s10278-023-00952-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Revised: 10/22/2023] [Accepted: 10/23/2023] [Indexed: 04/20/2024]
Abstract
Prompt and correct detection of pulmonary tuberculosis (PTB) is critical in preventing its spread. We aimed to develop a deep learning-based algorithm for detecting PTB on chest X-ray (CXRs) in the emergency department. This retrospective study included 3498 CXRs acquired from the National Taiwan University Hospital (NTUH). The images were chronologically split into a training dataset, NTUH-1519 (images acquired during the years 2015 to 2019; n = 2144), and a testing dataset, NTUH-20 (images acquired during the year 2020; n = 1354). Public databases, including the NIH ChestX-ray14 dataset (model training; 112,120 images), Montgomery County (model testing; 138 images), and Shenzhen (model testing; 662 images), were also used in model development. EfficientNetV2 was the basic architecture of the algorithm. Images from ChestX-ray14 were employed for pseudo-labelling to perform semi-supervised learning. The algorithm demonstrated excellent performance in detecting PTB (area under the receiver operating characteristic curve [AUC] 0.878, 95% confidence interval [CI] 0.854-0.900) in NTUH-20. The algorithm showed significantly better performance in posterior-anterior (PA) CXR (AUC 0.940, 95% CI 0.912-0.965, p-value < 0.001) compared with anterior-posterior (AUC 0.782, 95% CI 0.644-0.897) or portable anterior-posterior (AUC 0.869, 95% CI 0.814-0.918) CXR. The algorithm accurately detected cases of bacteriologically confirmed PTB (AUC 0.854, 95% CI 0.823-0.883). Finally, the algorithm tested favourably in Montgomery County (AUC 0.838, 95% CI 0.765-0.904) and Shenzhen (AUC 0.806, 95% CI 0.771-0.839). A deep learning-based algorithm could detect PTB on CXR with excellent performance, which may help shorten the interval between detection and airborne isolation for patients with PTB.
Collapse
Affiliation(s)
- Chih-Hung Wang
- Department of Emergency Medicine, College of Medicine, National Taiwan University, Taipei, Taiwan
- Department of Emergency Medicine, National Taiwan University Hospital, No. 7, Zhongshan S. Rd, Zhongzheng Dist., Taipei City, 100, Taiwan
| | - Weishan Chang
- Department of Mathematics, National Taiwan University, Taipei, Taiwan
| | - Meng-Rui Lee
- Department of Internal Medicine, National Taiwan University Hospital, Taipei, Taiwan
| | - Joyce Tay
- Department of Emergency Medicine, National Taiwan University Hospital, No. 7, Zhongshan S. Rd, Zhongzheng Dist., Taipei City, 100, Taiwan
| | - Cheng-Yi Wu
- Department of Emergency Medicine, National Taiwan University Hospital, No. 7, Zhongshan S. Rd, Zhongzheng Dist., Taipei City, 100, Taiwan
| | - Meng-Che Wu
- Department of Emergency Medicine, National Taiwan University Hospital, No. 7, Zhongshan S. Rd, Zhongzheng Dist., Taipei City, 100, Taiwan
| | | | - Dong Yang
- NVIDIA Corporation, Bethesda, MD, USA
| | - Can Zhao
- NVIDIA Corporation, Bethesda, MD, USA
| | - Weichung Wang
- Institute of Applied Mathematical Sciences, National Taiwan University, No. 1, Sec. 4, Roosevelt Rd., Taipei, 106, Taiwan.
| | - Chien-Hua Huang
- Department of Emergency Medicine, College of Medicine, National Taiwan University, Taipei, Taiwan.
- Department of Emergency Medicine, National Taiwan University Hospital, No. 7, Zhongshan S. Rd, Zhongzheng Dist., Taipei City, 100, Taiwan.
| |
Collapse
|
14
|
Sierra-Jerez F, Martinez F. A non-aligned translation with a neoplastic classifier regularization to include vascular NBI patterns in standard colonoscopies. Comput Biol Med 2024; 170:108008. [PMID: 38277922 DOI: 10.1016/j.compbiomed.2024.108008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 12/21/2023] [Accepted: 01/13/2024] [Indexed: 01/28/2024]
Abstract
Polyp vascular patterns are key to categorizing colorectal cancer malignancy. These patterns are typically observed in situ from specialized narrow-band images (NBI). Nonetheless, such vascular characterization is lost from standard colonoscopies (the primary attention mechanism). Besides, even for NBI observations, the categorization remains biased for expert observations, reporting errors in classification from 59.5% to 84.2%. This work introduces an end-to-end computational strategy to enhance in situ standard colonoscopy observations, including vascular patterns typically observed from NBI mechanisms. These retrieved synthetic images are achieved by adjusting a deep representation under a non-aligned translation task from optical colonoscopy (OC) to NBI. The introduced scheme includes an architecture to discriminate enhanced neoplastic patterns achieving a remarkable separation into the embedding representation. The proposed approach was validated in a public dataset with a total of 76 sequences, including standard optical sequences and the respective NBI observations. The enhanced optical sequences were automatically classified among adenomas and hyperplastic samples achieving an F1-score of 0.86%. To measure the sensibility capability of the proposed approach, serrated samples were projected to the trained architecture. In this experiment, statistical differences from three classes with a ρ-value <0.05 were reported, following a Mann-Whitney U test. This work showed remarkable polyp discrimination results in enhancing OC sequences regarding typical NBI patterns. This method also learns polyp class distributions under the unpaired criteria (close to real practice), with the capability to separate serrated samples from adenomas and hyperplastic ones.
Collapse
Affiliation(s)
- Franklin Sierra-Jerez
- Biomedical Imaging, Vision and Learning Laboratory (BIVL(2)ab), Universidad Industrial de Santander (UIS), Colombia
| | - Fabio Martinez
- Biomedical Imaging, Vision and Learning Laboratory (BIVL(2)ab), Universidad Industrial de Santander (UIS), Colombia.
| |
Collapse
|
15
|
Mohamed N, Almutairi RL, Abdelrahim S, Alharbi R, Alhomayani FM, Elamin Elnaim BM, Elhag AA, Dhakal R. Automated Laryngeal Cancer Detection and Classification Using Dwarf Mongoose Optimization Algorithm with Deep Learning. Cancers (Basel) 2023; 16:181. [PMID: 38201608 PMCID: PMC10778016 DOI: 10.3390/cancers16010181] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 12/25/2023] [Accepted: 12/26/2023] [Indexed: 01/12/2024] Open
Abstract
Laryngeal cancer (LCA) is a serious disease with a concerning global rise in incidence. Accurate treatment for LCA is particularly challenging in later stages, due to its complex nature as a head and neck malignancy. To address this challenge, researchers have been actively developing various analysis methods and tools to assist medical professionals in efficient LCA identification. However, existing tools and methods often suffer from various limitations, including low accuracy in early-stage LCA detection, high computational complexity, and lengthy patient screening times. With this motivation, this study presents an Automated Laryngeal Cancer Detection and Classification using a Dwarf Mongoose Optimization Algorithm with Deep Learning (ALCAD-DMODL) technique. The main objective of the ALCAD-DMODL method is to recognize the existence of LCA using the DL model. In the presented ALCAD-DMODL technique, a median filtering (MF)-based noise removal process takes place to get rid of the noise. Additionally, the ALCAD-DMODL technique involves the EfficientNet-B0 model for deriving feature vectors from the pre-processed images. For optimal hyperparameter tuning of the EfficientNet-B0 model, the DMO algorithm can be applied to select the parameters. Finally, the multi-head bidirectional gated recurrent unit (MBGRU) model is applied for the recognition and classification of LCA. The simulation result analysis of the ALCAD-DMODL technique is carried out on the throat region image dataset. The comparison study stated the supremacy of the ALCAD-DMODL technique in terms of distinct measures.
Collapse
Affiliation(s)
- Nuzaiha Mohamed
- Department of Public Health, College of Public Health and Health Informatics, University of Hail, Ha’il 81451, Saudi Arabia; (N.M.); (R.L.A.); (S.A.)
| | - Reem Lafi Almutairi
- Department of Public Health, College of Public Health and Health Informatics, University of Hail, Ha’il 81451, Saudi Arabia; (N.M.); (R.L.A.); (S.A.)
| | - Sayda Abdelrahim
- Department of Public Health, College of Public Health and Health Informatics, University of Hail, Ha’il 81451, Saudi Arabia; (N.M.); (R.L.A.); (S.A.)
| | - Randa Alharbi
- Department of Statistics, Faculty of Science, University of Tabuk, Tabuk 71491, Saudi Arabia;
| | - Fahad Mohammed Alhomayani
- College of Computers and Information Technology, Taif University, Taif 21944, Saudi Arabia;
- Applied College, Taif University, Taif 21944, Saudi Arabia
| | - Bushra M. Elamin Elnaim
- Department of Computer Science, College of Science and Humanities in Al-Sulail, Prince Sattam Bin Abdulaziz University, Al-Kharj 16278, Saudi Arabia;
| | - Azhari A. Elhag
- Department of Mathematics and Statistics, College of Science, Taif University, Taif 21944, Saudi Arabia;
| | - Rajendra Dhakal
- Department of Computer Science and Engineering, Sejong University, Seoul 05006, Republic of Korea
| |
Collapse
|
16
|
Su Z, Adam A, Nasrudin MF, Ayob M, Punganan G. Skeletal Fracture Detection with Deep Learning: A Comprehensive Review. Diagnostics (Basel) 2023; 13:3245. [PMID: 37892066 PMCID: PMC10606060 DOI: 10.3390/diagnostics13203245] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2023] [Revised: 10/12/2023] [Accepted: 10/13/2023] [Indexed: 10/29/2023] Open
Abstract
Deep learning models have shown great promise in diagnosing skeletal fractures from X-ray images. However, challenges remain that hinder progress in this field. Firstly, a lack of clear definitions for recognition, classification, detection, and localization tasks hampers the consistent development and comparison of methodologies. The existing reviews often lack technical depth or have limited scope. Additionally, the absence of explainable facilities undermines the clinical application and expert confidence in results. To address these issues, this comprehensive review analyzes and evaluates 40 out of 337 recent papers identified in prestigious databases, including WOS, Scopus, and EI. The objectives of this review are threefold. Firstly, precise definitions are established for the bone fracture recognition, classification, detection, and localization tasks within deep learning. Secondly, each study is summarized based on key aspects such as the bones involved, research objectives, dataset sizes, methods employed, results obtained, and concluding remarks. This process distills the diverse approaches into a generalized processing framework or workflow. Moreover, this review identifies the crucial areas for future research in deep learning models for bone fracture diagnosis. These include enhancing the network interpretability, integrating multimodal clinical information, providing therapeutic schedule recommendations, and developing advanced visualization methods for clinical application. By addressing these challenges, deep learning models can be made more intelligent and specialized in this domain. In conclusion, this review fills the gap in precise task definitions within deep learning for bone fracture diagnosis and provides a comprehensive analysis of the recent research. The findings serve as a foundation for future advancements, enabling improved interpretability, multimodal integration, clinical decision support, and advanced visualization techniques.
Collapse
Affiliation(s)
- Zhihao Su
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.); (M.A.)
| | - Afzan Adam
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.); (M.A.)
| | - Mohammad Faidzul Nasrudin
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.); (M.A.)
| | - Masri Ayob
- Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi 43600, Selangor, Malaysia; (Z.S.); (M.F.N.); (M.A.)
| | - Gauthamen Punganan
- Department of Orthopedics and Traumatology, Hospital Raja Permaisuri Bainun, Ipoh 30450, Perak, Malaysia;
| |
Collapse
|
17
|
Zheng Y, Huang D, Hao X, Wei J, Lu H, Liu Y. UniVisNet: A Unified Visualization and Classification Network for accurate grading of gliomas from MRI. Comput Biol Med 2023; 165:107332. [PMID: 37598632 DOI: 10.1016/j.compbiomed.2023.107332] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Revised: 07/30/2023] [Accepted: 08/07/2023] [Indexed: 08/22/2023]
Abstract
Accurate grading of brain tumors plays a crucial role in the diagnosis and treatment of glioma. While convolutional neural networks (CNNs) have shown promising performance in this task, their clinical applicability is still constrained by the interpretability and robustness of the models. In the conventional framework, the classification model is trained first, and then visual explanations are generated. However, this approach often leads to models that prioritize classification performance or complexity, making it difficult to achieve a precise visual explanation. Motivated by these challenges, we propose the Unified Visualization and Classification Network (UniVisNet), a novel framework that aims to improve both the classification performance and the generation of high-resolution visual explanations. UniVisNet addresses attention misalignment by introducing a subregion-based attention mechanism, which replaces traditional down-sampling operations. Additionally, multiscale feature maps are fused to achieve higher resolution, enabling the generation of detailed visual explanations. To streamline the process, we introduce the Unified Visualization and Classification head (UniVisHead), which directly generates visual explanations without the need for additional separation steps. Through extensive experiments, our proposed UniVisNet consistently outperforms strong baseline classification models and prevalent visualization methods. Notably, UniVisNet achieves remarkable results on the glioma grading task, including an AUC of 94.7%, an accuracy of 89.3%, a sensitivity of 90.4%, and a specificity of 85.3%. Moreover, UniVisNet provides visually interpretable explanations that surpass existing approaches. In conclusion, UniVisNet innovatively generates visual explanations in brain tumor grading by simultaneously improving the classification performance and generating high-resolution visual explanations. This work contributes to the clinical application of deep learning, empowering clinicians with comprehensive insights into the spatial heterogeneity of glioma.
Collapse
Affiliation(s)
- Yao Zheng
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China
| | - Dong Huang
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China; Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China
| | - Xiaoshuo Hao
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China
| | - Jie Wei
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China
| | - Hongbing Lu
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China; Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China.
| | - Yang Liu
- Air Force Medical University, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China; Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, No. 169 Changle West Road, Xi'an, 710032, ShaanXi, China.
| |
Collapse
|
18
|
Atabansi CC, Nie J, Liu H, Song Q, Yan L, Zhou X. A survey of Transformer applications for histopathological image analysis: New developments and future directions. Biomed Eng Online 2023; 22:96. [PMID: 37749595 PMCID: PMC10518923 DOI: 10.1186/s12938-023-01157-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2023] [Accepted: 09/15/2023] [Indexed: 09/27/2023] Open
Abstract
Transformers have been widely used in many computer vision challenges and have shown the capability of producing better results than convolutional neural networks (CNNs). Taking advantage of capturing long-range contextual information and learning more complex relations in the image data, Transformers have been used and applied to histopathological image processing tasks. In this survey, we make an effort to present a thorough analysis of the uses of Transformers in histopathological image analysis, covering several topics, from the newly built Transformer models to unresolved challenges. To be more precise, we first begin by outlining the fundamental principles of the attention mechanism included in Transformer models and other key frameworks. Second, we analyze Transformer-based applications in the histopathological imaging domain and provide a thorough evaluation of more than 100 research publications across different downstream tasks to cover the most recent innovations, including survival analysis and prediction, segmentation, classification, detection, and representation. Within this survey work, we also compare the performance of CNN-based techniques to Transformers based on recently published papers, highlight major challenges, and provide interesting future research directions. Despite the outstanding performance of the Transformer-based architectures in a number of papers reviewed in this survey, we anticipate that further improvements and exploration of Transformers in the histopathological imaging domain are still required in the future. We hope that this survey paper will give readers in this field of study a thorough understanding of Transformer-based techniques in histopathological image analysis, and an up-to-date paper list summary will be provided at https://github.com/S-domain/Survey-Paper .
Collapse
Affiliation(s)
| | - Jing Nie
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China.
| | - Haijun Liu
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China
| | - Qianqian Song
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China
| | - Lingfeng Yan
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China
| | - Xichuan Zhou
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing, 400044, China.
| |
Collapse
|
19
|
Oliveira SP, Montezuma D, Moreira A, Oliveira D, Neto PC, Monteiro A, Monteiro J, Ribeiro L, Gonçalves S, Pinto IM, Cardoso JS. A CAD system for automatic dysplasia grading on H&E cervical whole-slide images. Sci Rep 2023; 13:3970. [PMID: 36894572 PMCID: PMC9998461 DOI: 10.1038/s41598-023-30497-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2022] [Accepted: 02/24/2023] [Indexed: 03/11/2023] Open
Abstract
Cervical cancer is the fourth most common female cancer worldwide and the fourth leading cause of cancer-related death in women. Nonetheless, it is also among the most successfully preventable and treatable types of cancer, provided it is early identified and properly managed. As such, the detection of pre-cancerous lesions is crucial. These lesions are detected in the squamous epithelium of the uterine cervix and are graded as low- or high-grade intraepithelial squamous lesions, known as LSIL and HSIL, respectively. Due to their complex nature, this classification can become very subjective. Therefore, the development of machine learning models, particularly directly on whole-slide images (WSI), can assist pathologists in this task. In this work, we propose a weakly-supervised methodology for grading cervical dysplasia, using different levels of training supervision, in an effort to gather a bigger dataset without the need of having all samples fully annotated. The framework comprises an epithelium segmentation step followed by a dysplasia classifier (non-neoplastic, LSIL, HSIL), making the slide assessment completely automatic, without the need for manual identification of epithelial areas. The proposed classification approach achieved a balanced accuracy of 71.07% and sensitivity of 72.18%, at the slide-level testing on 600 independent samples, which are publicly available upon reasonable request.
Collapse
Affiliation(s)
- Sara P Oliveira
- INESCTEC, 4200-465, Porto, Portugal.,FEUP, University of Porto, 4200-465, Porto, Portugal
| | - Diana Montezuma
- IMP Diagnostics, 4150-146, Porto, Portugal. .,ICBAS, University of Porto, 4050-313, Porto, Portugal.
| | - Ana Moreira
- FEUP, University of Porto, 4200-465, Porto, Portugal
| | | | - Pedro C Neto
- INESCTEC, 4200-465, Porto, Portugal.,FEUP, University of Porto, 4200-465, Porto, Portugal
| | | | | | | | | | | | - Jaime S Cardoso
- INESCTEC, 4200-465, Porto, Portugal.,FEUP, University of Porto, 4200-465, Porto, Portugal
| |
Collapse
|
20
|
Nagaraju S, Kumar KV, Rani BP, Lydia EL, Ishak MK, Filali I, Karim FK, Mostafa SM. Automated Diabetic Foot Ulcer Detection and Classification Using Deep Learning. IEEE ACCESS 2023; 11:127578-127588. [DOI: 10.1109/access.2023.3332292] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/06/2025]
Affiliation(s)
- Sunnam Nagaraju
- Department of Mechanical Engineering, MLR Institute of Technology, Hyderabad, India
| | - Kollati Vijaya Kumar
- Department of Computer Science and Engineering, GITAM School of Technology, Vishakhapatnam Campus, GITAM (Deemed to be a University), Visakhapatnam, India
| | - B. Prameela Rani
- Department of CSE-AIML, Aditya College of Engineering, Surampalem, Andhra Pradesh, India
| | - E. Laxmi Lydia
- Department of Computer Science and Engineering, Vignan's Institute of Information Technology, Visakhapatnam, India
| | - Mohamad Khairi Ishak
- Department of Electrical and Computer Engineering, College of Engineering and Information Technology, Ajman University, Ajman, United Arab Emirates
| | - Imen Filali
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Faten Khalid Karim
- Department of Computer Sciences, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Samih M. Mostafa
- Computer Science Department, Faculty of Computers and Information, South Valley University, Qena, Egypt
| |
Collapse
|