1
|
Sampieri C, Peretti G. Democratizing cancer detection: artificial intelligence-enhanced endoscopy could address global disparities in head and neck cancer outcomes. Eur Arch Otorhinolaryngol 2025; 282:2739-2743. [PMID: 39966156 DOI: 10.1007/s00405-025-09257-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2024] [Accepted: 01/30/2025] [Indexed: 02/20/2025]
Abstract
INTRODUCTION This article explores the potential role of artificial intelligence (AI) in enhancing the early detection and diagnosis of head and neck squamous cell carcinoma (HNSCC). DISCUSSION The latest data reflecting the growing disparity in HNSCC incidence and mortality across different regions and socioeconomic groups are reviewed, stressing the importance of early detection in improving outcomes. The potential role of AI in improving the accuracy and consistency of endoscopic procedures, particularly in resource-limited settings with limited access to specialized healthcare is discussed. By analyzing current technologies and case studies, the review highlights how AI can democratize access to early cancer detection, improve survival rates, and reduce the financial burden on patients. The integration of AI into healthcare systems can lead to significant cost savings and enhanced accessibility, particularly in underserved populations, where HNSCC is projected to rise dramatically in the coming decades.
Collapse
Affiliation(s)
- Claudio Sampieri
- Department of Otolaryngology, Hospital Clínic, Barcelona, Spain.
- Department of Experimental Medicine (DIMES), University of Genoa, Genoa, Italy.
- Department of Otorhinolaryngology, Hospital Clínic, C. de Villarroel, 170, Barcelona, 08029, Spain.
| | - Giorgio Peretti
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| |
Collapse
|
2
|
Boscolo Nata F, Tirelli G. Comments on "Achieving negative superficial resection margins with NBI and white light in carcinoma oral cavity: Could it be a norm?". Oral Oncol 2025; 163:107226. [PMID: 40015146 DOI: 10.1016/j.oraloncology.2025.107226] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2025] [Accepted: 02/23/2025] [Indexed: 03/01/2025]
Affiliation(s)
- F Boscolo Nata
- ENT Clinic, Head and Neck Department, University of Trieste, Strada di Fiume 447, Trieste, Italy.
| | - G Tirelli
- ENT Clinic, Head and Neck Department, University of Trieste, Strada di Fiume 447, Trieste, Italy
| |
Collapse
|
3
|
Baldini C, Migliorelli L, Berardini D, Azam MA, Sampieri C, Ioppi A, Srivastava R, Peretti G, Mattos LS. Improving real-time detection of laryngeal lesions in endoscopic images using a decoupled super-resolution enhanced YOLO. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 260:108539. [PMID: 39689500 DOI: 10.1016/j.cmpb.2024.108539] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 11/08/2024] [Accepted: 11/27/2024] [Indexed: 12/19/2024]
Abstract
BACKGROUND AND OBJECTIVE Laryngeal Cancer (LC) constitutes approximately one third of head and neck cancers. Detecting early-stage lesions in this anatomical region is crucial for achieving a high survival rate. However, it poses significant diagnostic challenges owing to the varied appearance of lesions and the need for precise characterization for appropriate clinical management. Conventional diagnostic approaches rely heavily on endoscopic examination, which often requires expert interpretation and may be limited by subjective assessment. Deep learning (DL) approaches offer promising opportunities for automating lesion detection, but their efficacy in handling multi-modal imaging data and accurately localizing small lesions remains a subject of investigation. Furthermore, the clinical domain may largely benefit from the deployment of efficient DL methods that can ensure equitable access to advanced technologies, regardless of the availability of resources that can often be limited. In this study, a DL-based approach, named SRE-YOLO, was introduced to provide real-time assistance to less-experienced personnel during laryngeal assessment, by automatically detecting lesions at different scales from endoscopic White Light (WL) and Narrow-Band Imaging (NBI) images. METHODS During the training, the SRE-YOLO integrates a YOLOv8 nano (YOLOv8n) baseline with a Super-Resolution (SR) branch to enhance lesion detection. This last component is decoupled during inference to preserve the low computational demand of the YOLOv8n baseline. The evaluation was conducted on a multi-center dataset, encompassing diverse laryngeal pathologies and acquisition modalities. RESULTS The SRE-YOLO method improved the Average Precision (AP@IoU=0.5) in lesion detection by 5% with respect to the YOLOv8n baseline, while maintaining the inference speed of 58.8 Frames Per Second (FPS). Comparative analyses against state-of-the-art DL methods highlighted the efficacy of the SRE-YOLO approach in balancing detection accuracy, computational efficiency, and real-time applicability. CONCLUSIONS This research underscores the potential of SRE-YOLO in developing efficient DL-driven decision support systems for real-time detection of laryngeal lesions at different scales from both WL and NBI endoscopic data.
Collapse
Affiliation(s)
- Chiara Baldini
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genova, Italy; Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genova, Italy.
| | - Lucia Migliorelli
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genova, Italy; Department of Information Engineering, Università Politecnica delle Marche, Ancona, Italy
| | - Daniele Berardini
- Department of Information Engineering, Università Politecnica delle Marche, Ancona, Italy
| | - Muhammad Adeel Azam
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genova, Italy; Department of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genova, Italy
| | - Claudio Sampieri
- Department of Experimental Medicine (DIMES), University of Genoa, Genova, Italy; Department of Otolaryngology, Hospital Clínic, Barcelona, Spain
| | - Alessandro Ioppi
- Department of Otorhinolaryngology-Head and Neck Surgery, S. Chiara Hospital, Azienda Provinciale per i Servizi Sanitari (APSS), Trento, Italy
| | - Rakesh Srivastava
- Sushrut Institute of Plastic Surgery & Super specialty Hospital, Lucknow, India
| | - Giorgio Peretti
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genova, Italy; Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genova, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genova, Italy
| |
Collapse
|
4
|
Zhang X, Zhao J, Zong D, Ren H, Gao C. Taming vision transformers for clinical laryngoscopy assessment. J Biomed Inform 2025; 162:104766. [PMID: 39827999 DOI: 10.1016/j.jbi.2024.104766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2024] [Revised: 12/09/2024] [Accepted: 12/26/2024] [Indexed: 01/22/2025]
Abstract
OBJECTIVE Laryngoscopy, essential for diagnosing laryngeal cancer (LCA), faces challenges due to high inter-observer variability and the reliance on endoscopist expertise. Distinguishing precancerous from early-stage cancerous lesions is particularly challenging, even for experienced practitioners, given their similar appearances. This study aims to enhance laryngoscopic image analysis to improve early screening/detection of cancer or precancerous conditions. METHODS We propose MedFormer, a laryngeal cancer classification method based on the Vision Transformer (ViT). To address data scarcity, MedFormer employs a customized transfer learning approach that leverages the representational power of pre-trained transformers. This method enables robust out-of-domain generalization by fine-tuning a minimal set of additional parameters. RESULTS MedFormer exhibits sensitivity-specificity values of 98%-89% for identifying precancerous lesions (leukoplakia) and 89%-97% for detecting cancer, surpassing CNN counterparts significantly. Additionally, when compared to the two selected ViT-based models, MedFormer also demonstrates superior performance. It also outperforms physician visual evaluations (PVE) in certain scenarios and matches PVE performance in all cases. Visualizations using class activation maps (CAM) and deformable patches demonstrate MedFormer's interpretability, aiding clinicians in understanding the model's predictions. CONCLUSION We highlight the potential of visual transformers in clinical laryngoscopic assessments, presenting MedFormer as an effective method for the early detection of laryngeal cancer.
Collapse
Affiliation(s)
- Xinzhu Zhang
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China
| | - Jing Zhao
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China.
| | - Daoming Zong
- School of Computer Science and Technology, East China Normal University, North Zhongshan Road 3663, Shanghai, 200062, China
| | - Henglei Ren
- Eye & ENT Hospital of Fudan University, Fenyang Road 83, Shanghai, 200000, China
| | - Chunli Gao
- Eye & ENT Hospital of Fudan University, Fenyang Road 83, Shanghai, 200000, China.
| |
Collapse
|
5
|
Kavak ÖT, Gündüz Ş, Vural C, Enver N. Artificial intelligence based diagnosis of sulcus: assesment of videostroboscopy via deep learning. Eur Arch Otorhinolaryngol 2024; 281:6083-6091. [PMID: 39001913 PMCID: PMC11512876 DOI: 10.1007/s00405-024-08801-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2024] [Accepted: 06/19/2024] [Indexed: 07/15/2024]
Abstract
PURPOSE To develop a convolutional neural network (CNN)-based model for classifying videostroboscopic images of patients with sulcus, benign vocal fold (VF) lesions, and healthy VFs to improve clinicians' accuracy in diagnosis during videostroboscopies when evaluating sulcus. MATERIALS AND METHODS Videostroboscopies of 433 individuals who were diagnosed with sulcus (91), who were diagnosed with benign VF diseases (i.e., polyp, nodule, papilloma, cyst, or pseudocyst [311]), or who were healthy (33) were analyzed. After extracting 91,159 frames from videostroboscopies, a CNN-based model was created and tested. The healthy and sulcus groups underwent binary classification. In the second phase of the study, benign VF lesions were added to the training set, and multiclassification was executed across all groups. The proposed CNN-based model results were compared with five laryngology experts' assessments. RESULTS In the binary classification phase, the CNN-based model achieved 98% accuracy, 98% recall, 97% precision, and a 97% F1 score for classifying sulcus and healthy VFs. During the multiclassification phase, when evaluated on a subset of frames encompassing all included groups, the CNN-based model demonstrated greater accuracy when compared with that of the five laryngologists (%76 versus 72%, 68%, 72%, 63%, and 72%). CONCLUSION The utilization of a CNN-based model serves as a significant aid in the diagnosis of sulcus, a VF disease that presents notable challenges in the diagnostic process. Further research could be undertaken to assess the practicality of implementing this approach in real-time application in clinical practice.
Collapse
Affiliation(s)
- Ömer Tarık Kavak
- Department of Otorhinolaryngology, Marmara University Faculty of Medicine, Pendik Training and Research Hospital, Fevzi Çakmak Muhsin Yazıcıoğlu Street, İstanbul, 34899, Turkey.
| | - Şevket Gündüz
- VRLab Academy, 32 Willoughby Rd, Harringay Ladder, London, N8 0JG, UK
| | - Cabir Vural
- Marmara University Faculty of Engineering, Electrical and Electronics Engineering, Başıbüyük, RTE Campus, İstanbul, 34854, Turkey
| | - Necati Enver
- Department of Otorhinolaryngology, Marmara University Faculty of Medicine, Pendik Training and Research Hospital, Fevzi Çakmak Muhsin Yazıcıoğlu Street, İstanbul, 34899, Turkey
| |
Collapse
|
6
|
Kim S, Chang Y, An S, Kim D, Cho J, Oh K, Baek S, Choi BK. Enhanced WGAN Model for Diagnosing Laryngeal Carcinoma. Cancers (Basel) 2024; 16:3482. [PMID: 39456576 PMCID: PMC11506071 DOI: 10.3390/cancers16203482] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Revised: 10/03/2024] [Accepted: 10/09/2024] [Indexed: 10/28/2024] Open
Abstract
This study modifies the U-Net architecture for pixel-based segmentation to automatically classify lesions in laryngeal endoscopic images. The advanced U-Net incorporates five-level encoders and decoders, with an autoencoder layer to derive latent vectors representing the image characteristics. To enhance performance, a WGAN was implemented to address common issues such as mode collapse and gradient explosion found in traditional GANs. The dataset consisted of 8171 images labeled with polygons in seven colors. Evaluation metrics, including the F1 score and intersection over union, revealed that benign tumors were detected with lower accuracy compared to other lesions, while cancers achieved notably high accuracy. The model demonstrated an overall accuracy rate of 99%. This enhanced U-Net model shows strong potential in improving cancer detection, reducing diagnostic errors, and enhancing early diagnosis in medical applications.
Collapse
Affiliation(s)
- Sungjin Kim
- Department of Artificial Intelligence, Cheju Halla University, Jeju 63092, Republic of Korea; (S.K.); (Y.C.); (S.A.)
| | - Yongjun Chang
- Department of Artificial Intelligence, Cheju Halla University, Jeju 63092, Republic of Korea; (S.K.); (Y.C.); (S.A.)
| | - Sungjun An
- Department of Artificial Intelligence, Cheju Halla University, Jeju 63092, Republic of Korea; (S.K.); (Y.C.); (S.A.)
| | - Deokseok Kim
- Research Lab, MTEG, Seoul 03920, Republic of Korea;
| | - Jaegu Cho
- Department of Otolaryngology-Head and Neck Surgery, Korea University College of Medicine, Seoul 02841, Republic of Korea; (J.C.); (K.O.); (S.B.)
| | - Kyungho Oh
- Department of Otolaryngology-Head and Neck Surgery, Korea University College of Medicine, Seoul 02841, Republic of Korea; (J.C.); (K.O.); (S.B.)
| | - Seungkuk Baek
- Department of Otolaryngology-Head and Neck Surgery, Korea University College of Medicine, Seoul 02841, Republic of Korea; (J.C.); (K.O.); (S.B.)
| | - Bo K. Choi
- Research Lab, MTEG, Seoul 03920, Republic of Korea;
| |
Collapse
|
7
|
Nwosu OI, Naunheim MR. Artificial Intelligence in Laryngology, Broncho-Esophagology, and Sleep Surgery. Otolaryngol Clin North Am 2024; 57:821-829. [PMID: 38719714 DOI: 10.1016/j.otc.2024.04.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/06/2024]
Abstract
Technological advancements in laryngology, broncho-esophagology, and sleep surgery have enabled the collection of increasing amounts of complex data for diagnosis and treatment of voice, swallowing, and sleep disorders. Clinicians face challenges in efficiently synthesizing these data for personalized patient care. Artificial intelligence (AI), specifically machine learning and deep learning, offers innovative solutions for processing and interpreting these data, revolutionizing diagnosis and management in these fields, and making care more efficient and effective. In this study, we review recent AI-based innovations in the fields of laryngology, broncho-esophagology, and sleep surgery.
Collapse
Affiliation(s)
- Obinna I Nwosu
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, MA, USA; Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, MA, USA
| | - Matthew R Naunheim
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, MA, USA; Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, MA, USA.
| |
Collapse
|
8
|
Xiong M, Luo JW, Ren J, Hu JJ, Lan L, Zhang Y, Lv D, Zhou XB, Yang H. Applying Deep Learning with Convolutional Neural Networks to Laryngoscopic Imaging for Automated Segmentation and Classification of Vocal Cord Leukoplakia. EAR, NOSE & THROAT JOURNAL 2024:1455613241275341. [PMID: 39302102 DOI: 10.1177/01455613241275341] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/22/2024] Open
Abstract
Objectives: Vocal cord leukoplakia is clinically described as a white plaque or patch on the vocal cords observed during macroscopic examination, which does not take into account histological features or prognosis. A clinical challenge in managing vocal cord leukoplakia is to assess the potential malignant transformation of the lesion. This study aims to investigate the potential of deep learning (DL) for the simultaneous segmentation and classification of vocal cord leukoplakia using narrow band imaging (NBI) and white light imaging (WLI). The primary objective is to assess the model's accuracy in detecting and classifying lesions, comparing its performance in WLI and NBI. Methods: We applied DL to segment and classify NBI and WLI of vocal cord leukoplakia, and used pathological diagnosis as the gold standard. Results: The DL model autonomously detected lesions with an average intersection-over-union (IoU) >70%. In classification tasks, the model differentiated between lesions in the surgical group with a sensitivity of 93% and a specificity of 94% for WLI, and a sensitivity of 99% and a specificity of 97% for NBI. In addition, the model achieved a mean average precision of 81% in WLI and 92% in NBI, with an IoU threshold >0.5. Conclusions: The model proposed by us is helpful in assisting in accurate diagnosis of vocal cord leukoplakia from NBI and WLI.
Collapse
Affiliation(s)
- Ming Xiong
- Department of Otolaryngology, Head & Neck Surgery, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Jia-Wei Luo
- West China Biomedical Big Data Center, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Jia Ren
- Department of Otolaryngology, Head & Neck Surgery, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Juan-Juan Hu
- Department of Otolaryngology, Head & Neck Surgery, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Lan Lan
- West China Biomedical Big Data Center, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Ying Zhang
- Department of Pathology, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Dan Lv
- Department of Otolaryngology, Head & Neck Surgery, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| | - Xiao-Bo Zhou
- School of Biomedical Informatics, University of Texas Health Science Center at Houston, Houston, TX, USA
| | - Hui Yang
- Department of Otolaryngology, Head & Neck Surgery, West China Hospital of Sichuan University/West China School of Medicine, Sichuan University, Chengdu, Sichuan, China
| |
Collapse
|
9
|
Wang D, Li N, Guo R, Pang J, Zhang L, Zhang F, Zhang J, Yang X. Clinical Value of Narrow Band Imaging Endoscopy in the Early Diagnosis and Staging Assessment of Laryngeal and Hypopharyngeal Cancer. J Voice 2024:S0892-1997(24)00211-X. [PMID: 39147689 DOI: 10.1016/j.jvoice.2024.07.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2024] [Revised: 06/28/2024] [Accepted: 07/01/2024] [Indexed: 08/17/2024]
Abstract
OBJECTIVE To explore the clinical value of narrow band imaging (NBI) endoscopy in the early diagnosis and staging assessment of laryngeal and hypopharyngeal cancer. METHODS A total of 78 patients with lesions in the hypopharynx or larynx were examined using endoscopy, observed under both white light and NBI modes, and graded using NBI. Using Lugol's iodine solution, laryngeal and hypopharyngeal lesions were graded using iodine staining. Using histopathological examination or postoperative pathological results as the diagnostic criteria, the sensitivity, specificity, and accuracy of endoscopy and iodine staining in diagnosing early cancer and precancerous lesions were evaluated. RESULTS Multiple lesions were identified by both methods, and pathological examination confirmed 86 lesions, including early squamous cell carcinoma and precancerous lesions, such as early esophageal cancer, high-grade esophageal intraepithelial neoplasia, and hypopharyngeal cancer. Endoscopy showed significantly higher accuracy, detection rate, sensitivity, and specificity in NBI mode than in white light mode (96.12%, 86.05%, 97.37%, 86.67% vs 86.05%, 76.74%, 86.84%, 80%, respectively; P < 0.05). NBI grading and iodine staining grading showed good consistency with pathological diagnosis, with a Kappa value of 0.684 and 0.622, respectively. CONCLUSIONS NBI endoscopy allows for better observation of subtle structural changes on the surface of lesions compared to white light endoscopy. It provides high accuracy in detecting early laryngeal and hypopharyngeal cancer and precancerous lesions, determining biopsy sites, facilitating early diagnosis, and establishing safe surgical margins. NBI endoscopy offers a viable alternative for non-invasive screening and early diagnosis of laryngeal and hypopharyngeal cancer, showing great potential for clinical advancement.
Collapse
Affiliation(s)
- Dapeng Wang
- Department of Radiation Oncology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Ning Li
- Department of Radiation Oncology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Ruyuan Guo
- Department of Radiation Oncology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Jing Pang
- Endoscopy Center, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Li Zhang
- Department of Head and Neck Surgery, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Fuli Zhang
- Department of Radiation Oncology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Junjie Zhang
- Department of Radiology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Xiaotang Yang
- Department of Radiology, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China.
| |
Collapse
|
10
|
Baldini C, Azam MA, Sampieri C, Ioppi A, Ruiz-Sevilla L, Vilaseca I, Alegre B, Tirrito A, Pennacchi A, Peretti G, Moccia S, Mattos LS. An automated approach for real-time informative frames classification in laryngeal endoscopy using deep learning. Eur Arch Otorhinolaryngol 2024; 281:4255-4264. [PMID: 38698163 PMCID: PMC11266252 DOI: 10.1007/s00405-024-08676-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2024] [Accepted: 04/08/2024] [Indexed: 05/05/2024]
Abstract
PURPOSE Informative image selection in laryngoscopy has the potential for improving automatic data extraction alone, for selective data storage and a faster review process, or in combination with other artificial intelligence (AI) detection or diagnosis models. This paper aims to demonstrate the feasibility of AI in providing automatic informative laryngoscopy frame selection also capable of working in real-time providing visual feedback to guide the otolaryngologist during the examination. METHODS Several deep learning models were trained and tested on an internal dataset (n = 5147 images) and then tested on an external test set (n = 646 images) composed of both white light and narrow band images. Four videos were used to assess the real-time performance of the best-performing model. RESULTS ResNet-50, pre-trained with the pretext strategy, reached a precision = 95% vs. 97%, recall = 97% vs, 89%, and the F1-score = 96% vs. 93% on the internal and external test set respectively (p = 0.062). The four testing videos are provided in the supplemental materials. CONCLUSION The deep learning model demonstrated excellent performance in identifying diagnostically relevant frames within laryngoscopic videos. With its solid accuracy and real-time capabilities, the system is promising for its development in a clinical setting, either autonomously for objective quality control or in conjunction with other algorithms within a comprehensive AI toolset aimed at enhancing tumor detection and diagnosis.
Collapse
Affiliation(s)
- Chiara Baldini
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Departement of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| | - Muhammad Adeel Azam
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Departement of Informatics, Bioengineering, Robotics and Systems Engineering, University of Genoa, Genoa, Italy
| | - Claudio Sampieri
- Department of Experimental Medicine (DIMES), University of Genoa, Genoa, Italy.
- Department of Otolaryngology, Hospital Clínic, C. de Villarroel, 170, 08029, Barcelona, Spain.
- Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain.
| | | | - Laura Ruiz-Sevilla
- Otorhinolaryngology Head-Neck Surgery Department, Hospital Universitari Joan XXIII de Tarragona, Tarragona, Spain
| | - Isabel Vilaseca
- Department of Otolaryngology, Hospital Clínic, C. de Villarroel, 170, 08029, Barcelona, Spain
- Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain
- Translational Genomics and Target Therapies in Solid Tumors Group, Institut d́Investigacions Biomèdiques August Pi i Sunyer, IDIBAPS, Barcelona, Spain
- Faculty of Medicine, University of Barcelona, Barcelona, Spain
| | - Berta Alegre
- Department of Otolaryngology, Hospital Clínic, C. de Villarroel, 170, 08029, Barcelona, Spain
- Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain
| | - Alessandro Tirrito
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Alessia Pennacchi
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Giorgio Peretti
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Sara Moccia
- The BioRobotics Institute and Department of Excellence in Robotics and AI, Scuola Superiore Sant'Anna, Pisa, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| |
Collapse
|
11
|
杜 玉, 杨 相, 韩 薇, 刘 吉. [The effect of narrow band imaging in CO2 laser therapy in early-stage glottic cancer]. LIN CHUANG ER BI YAN HOU TOU JING WAI KE ZA ZHI = JOURNAL OF CLINICAL OTORHINOLARYNGOLOGY HEAD AND NECK SURGERY 2024; 38:646-650. [PMID: 38973047 PMCID: PMC11599962 DOI: 10.13201/j.issn.2096-7993.2024.07.017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 11/16/2023] [Revised: 01/09/2024] [Indexed: 07/09/2024]
Abstract
Objective:To explore efficacy of narrow band imaging(NBI) technique in CO2laser therapy in Early-Stage Glottic cancer. Methods:The clinical data of patients with Early-Stage Glottic cancer who underwent CO2laser vocal cord resection from June 2011 to August 2022 were retrospectively analyzed. Among these, 27 patients who underwent surgery assisted by NBI were assigned to the observation group, while 25 patients who underwent conventional CO2 laser microsurgery with a suspension laryngoscope were assigned to the control group. The differences between the two groups were analyzed in terms of intraoperative frozen pathology results, postoperative recurrence rates, 5-year cumulative disease-free survival rates, complications, and voice recovery. Results:All 52 patients were operated successfully. Temporary tracheostomy and serious complications did not occur during the operation. The postoperative patient's pronunciation was satisfactory. One patient experienced vocal cord adhesion, but there were no severe complications such as breathing difficulties or bleeding, with an overall complication rate of 1.92%. Postoperative follow-up was 1-5 years. The 5 years recurrence free survival in the general group was 77.90%, and the 5 years recurrence free survival in the NBI group was 100%, the difference was statistically significant(P<0.05). NBI endoscopy is safer and more accurate than the general group in determining the safe margin of tumor mucosal resection(P<0.05). Among the patients who accepted the voice analysis, the difference was no statistically significant(P>0.05). Conclusion:Compared with conventional CO2laser surgery under microscope, NBI guided laser resection of Early-Stage Glottic cancer is more accurate. NBI guided laser resection could improve 5 years recurrence free survival rate. In a word, narrow-band imaging endoscopy can has very high value in clinical application.
Collapse
Affiliation(s)
- 玉晓 杜
- 天津市人民医院耳鼻咽喉头颈外科(天津,300122)Department of Otorhinolaryngology Head and Neck Surgery, Tianjin Union Medical Center, Tianjin, 300122, China
| | - 相立 杨
- 天津市人民医院耳鼻咽喉头颈外科(天津,300122)Department of Otorhinolaryngology Head and Neck Surgery, Tianjin Union Medical Center, Tianjin, 300122, China
| | - 薇薇 韩
- 天津市人民医院耳鼻咽喉头颈外科(天津,300122)Department of Otorhinolaryngology Head and Neck Surgery, Tianjin Union Medical Center, Tianjin, 300122, China
| | - 吉祥 刘
- 天津市人民医院耳鼻咽喉头颈外科(天津,300122)Department of Otorhinolaryngology Head and Neck Surgery, Tianjin Union Medical Center, Tianjin, 300122, China
| |
Collapse
|
12
|
Wang ML, Tie CW, Wang JH, Zhu JQ, Chen BH, Li Y, Zhang S, Liu L, Guo L, Yang L, Yang LQ, Wei J, Jiang F, Zhao ZQ, Wang GQ, Zhang W, Zhang QM, Ni XG. Multi-instance learning based artificial intelligence model to assist vocal fold leukoplakia diagnosis: A multicentre diagnostic study. Am J Otolaryngol 2024; 45:104342. [PMID: 38703609 DOI: 10.1016/j.amjoto.2024.104342] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2024] [Accepted: 04/23/2024] [Indexed: 05/06/2024]
Abstract
OBJECTIVE To develop a multi-instance learning (MIL) based artificial intelligence (AI)-assisted diagnosis models by using laryngoscopic images to differentiate benign and malignant vocal fold leukoplakia (VFL). METHODS The AI system was developed, trained and validated on 5362 images of 551 patients from three hospitals. Automated regions of interest (ROI) segmentation algorithm was utilized to construct image-level features. MIL was used to fusion image level results to patient level features, then the extracted features were modeled by seven machine learning algorithms. Finally, we evaluated the image level and patient level results. Additionally, 50 videos of VFL were prospectively gathered to assess the system's real-time diagnostic capabilities. A human-machine comparison database was also constructed to compare the diagnostic performance of otolaryngologists with and without AI assistance. RESULTS In internal and external validation sets, the maximum area under the curve (AUC) for image level segmentation models was 0.775 (95 % CI 0.740-0.811) and 0.720 (95 % CI 0.684-0.756), respectively. Utilizing a MIL-based fusion strategy, the AUC at the patient level increased to 0.869 (95 % CI 0.798-0.940) and 0.851 (95 % CI 0.756-0.945). For real-time video diagnosis, the maximum AUC at the patient level reached 0.850 (95 % CI, 0.743-0.957). With AI assistance, the AUC improved from 0.720 (95 % CI 0.682-0.755) to 0.808 (95 % CI 0.775-0.839) for senior otolaryngologists and from 0.647 (95 % CI 0.608-0.686) to 0.807 (95 % CI 0.773-0.837) for junior otolaryngologists. CONCLUSIONS The MIL based AI-assisted diagnosis system can significantly improve the diagnostic performance of otolaryngologists for VFL and help to make proper clinical decisions.
Collapse
Affiliation(s)
- Mei-Ling Wang
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Shenzhen, China
| | - Cheng-Wei Tie
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Beijing, China
| | - Jian-Hui Wang
- Department of Endoscopy, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China
| | - Ji-Qing Zhu
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Beijing, China
| | - Bing-Hong Chen
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Shenzhen, China
| | - Ying Li
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Shenzhen, China
| | - Sen Zhang
- Department of Otolaryngology Head and Neck Surgery, The First Hospital, Shanxi Medical University, Taiyuan, China
| | - Lin Liu
- Department of Otolaryngology Head and Neck Surgery, Dalian Friendship Hospital, Dalian, China
| | - Li Guo
- Department of Otolaryngology Head and Neck Surgery, the First Affiliated Hospital, College of Clinical Medicine of Henan University of Science and Technology, Luoyang, China
| | - Long Yang
- Department of Otolaryngology, The Second People's Hospital of Baoshan City, Baoshan, China
| | - Li-Qun Yang
- Department of Otolaryngology, The Second People's Hospital of Baoshan City, Baoshan, China
| | - Jiao Wei
- Department of Otolaryngology, Qujing Second People's Hospital of Yunnan Province, Qujing, China
| | - Feng Jiang
- Department of Otolaryngology, Kunming First People's Hospital, Kunming, China
| | - Zhi-Qiang Zhao
- Department of Otolaryngology, Baoshan People's Hospital, Baoshan, China
| | - Gui-Qi Wang
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Beijing, China.
| | - Wei Zhang
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Shenzhen, China.
| | - Quan-Mao Zhang
- Department of Endoscopy, Shanxi Province Cancer Hospital, Shanxi Hospital Affiliated to Cancer Hospital, Chinese Academy of Medical Sciences, Cancer Hospital Affiliated to Shanxi Medical University, Taiyuan, China.
| | - Xiao-Guang Ni
- Department of Endoscopy, National Cancer Center, National Clinical Research Center for Cancer, Cancer Hospital, Chinese Academy of Medical Sciences, Peking Union Medical College, Beijing, China.
| |
Collapse
|
13
|
Sampieri C, Azam MA, Ioppi A, Baldini C, Moccia S, Kim D, Tirrito A, Paderno A, Piazza C, Mattos LS, Peretti G. Real-Time Laryngeal Cancer Boundaries Delineation on White Light and Narrow-Band Imaging Laryngoscopy with Deep Learning. Laryngoscope 2024; 134:2826-2834. [PMID: 38174772 DOI: 10.1002/lary.31255] [Citation(s) in RCA: 4] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Revised: 12/05/2023] [Accepted: 12/11/2023] [Indexed: 01/05/2024]
Abstract
OBJECTIVE To investigate the potential of deep learning for automatically delineating (segmenting) laryngeal cancer superficial extent on endoscopic images and videos. METHODS A retrospective study was conducted extracting and annotating white light (WL) and Narrow-Band Imaging (NBI) frames to train a segmentation model (SegMENT-Plus). Two external datasets were used for validation. The model's performances were compared with those of two otolaryngology residents. In addition, the model was tested on real intraoperative laryngoscopy videos. RESULTS A total of 3933 images of laryngeal cancer from 557 patients were used. The model achieved the following median values (interquartile range): Dice Similarity Coefficient (DSC) = 0.83 (0.70-0.90), Intersection over Union (IoU) = 0.83 (0.73-0.90), Accuracy = 0.97 (0.95-0.99), Inference Speed = 25.6 (25.1-26.1) frames per second. The external testing cohorts comprised 156 and 200 images. SegMENT-Plus performed similarly on all three datasets for DSC (p = 0.05) and IoU (p = 0.07). No significant differences were noticed when separately analyzing WL and NBI test images on DSC (p = 0.06) and IoU (p = 0.78) and when analyzing the model versus the two residents on DSC (p = 0.06) and IoU (Senior vs. SegMENT-Plus, p = 0.13; Junior vs. SegMENT-Plus, p = 1.00). The model was then tested on real intraoperative laryngoscopy videos. CONCLUSION SegMENT-Plus can accurately delineate laryngeal cancer boundaries in endoscopic images, with performances equal to those of two otolaryngology residents. The results on the two external datasets demonstrate excellent generalization capabilities. The computation speed of the model allowed its application on videolaryngoscopies simulating real-time use. Clinical trials are needed to evaluate the role of this technology in surgical practice and resection margin improvement. LEVEL OF EVIDENCE III Laryngoscope, 134:2826-2834, 2024.
Collapse
Affiliation(s)
- Claudio Sampieri
- Department of Experimental Medicine (DIMES), University of Genova, Genoa, Italy
- Functional Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain
- Otorhinolaryngology Department, Hospital Clínic, Barcelona, Spain
| | - Muhammad Adeel Azam
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Dipartimento di Informatica, Bioingegneria, Robotica e Ingegneria dei Sistemi (DIBRIS), University of Genova, Genoa, Italy
| | - Alessandro Ioppi
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genova, Genoa, Italy
- Department of Otorhinolaryngology-Head and Neck Surgery, "S. Chiara" Hospital, Azienda Provinciale per i Servizi Sanitari (APSS), Trento, Italy
| | - Chiara Baldini
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Dipartimento di Informatica, Bioingegneria, Robotica e Ingegneria dei Sistemi (DIBRIS), University of Genova, Genoa, Italy
| | - Sara Moccia
- The BioRobotics Institute and Department of Excellence in Robotics and AI, Scuola Superiore Sant'Anna, Pisa, Italy
| | - Dahee Kim
- Department of Otorhinolaryngology, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Alessandro Tirrito
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genova, Genoa, Italy
| | - Alberto Paderno
- Unit of Otorhinolaryngology - Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, Brescia, Italy
| | - Cesare Piazza
- Unit of Otorhinolaryngology - Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, Brescia, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Giorgio Peretti
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genova, Genoa, Italy
| |
Collapse
|
14
|
Wang Y, Yang C, Yang Q, Zhong R, Wang K, Shen H. Diagnosis of cervical lymphoma using a YOLO-v7-based model with transfer learning. Sci Rep 2024; 14:11073. [PMID: 38744888 PMCID: PMC11094110 DOI: 10.1038/s41598-024-61955-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2023] [Accepted: 05/12/2024] [Indexed: 05/16/2024] Open
Abstract
To investigate the ability of an auxiliary diagnostic model based on the YOLO-v7-based model in the classification of cervical lymphadenopathy images and compare its performance against qualitative visual evaluation by experienced radiologists. Three types of lymph nodes were sampled randomly but not uniformly. The dataset was randomly divided into for training, validation, and testing. The model was constructed with PyTorch. It was trained and weighting parameters were tuned on the validation set. Diagnostic performance was compared with that of the radiologists on the testing set. The mAP of the model was 96.4% at the 50% intersection-over-union threshold. The accuracy values of it were 0.962 for benign lymph nodes, 0.982 for lymphomas, and 0.960 for metastatic lymph nodes. The precision values of it were 0.928 for benign lymph nodes, 0.975 for lymphomas, and 0.927 for metastatic lymph nodes. The accuracy values of radiologists were 0.659 for benign lymph nodes, 0.836 for lymphomas, and 0.580 for metastatic lymph nodes. The precision values of radiologists were 0.478 for benign lymph nodes, 0.329 for lymphomas, and 0.596 for metastatic lymph nodes. The model effectively classifies lymphadenopathies from ultrasound images and outperforms qualitative visual evaluation by experienced radiologists in differential diagnosis.
Collapse
Affiliation(s)
- Yuegui Wang
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China
| | - Caiyun Yang
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China
| | - Qiuting Yang
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China
| | - Rong Zhong
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China
| | - Kangjian Wang
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China
| | - Haolin Shen
- Department of Ultrasound, Zhangzhou Affiliated Hospital to Fujian Medical University, No. 59 North Shengli Road, Zhangzhou, 363000, Fujian, China.
| |
Collapse
|
15
|
Tao X, Zhao X, Liu H, Wang J, Tian C, Liu L, Ding Y, Chen X, Liu Y. Automatic Recognition of Concealed Fish Bones under Laryngoscopy: A Practical AI Model Based on YOLO-V5. Laryngoscope 2024; 134:2162-2169. [PMID: 37983879 DOI: 10.1002/lary.31175] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Revised: 09/27/2023] [Accepted: 10/24/2023] [Indexed: 11/22/2023]
Abstract
BACKGROUND Fish bone impaction is one of the most common problems encountered in otolaryngology emergencies. Due to their small and transparent nature, as well as the complexity of pharyngeal anatomy, identifying fish bones efficiently under laryngoscopy requires substantial clinical experience. This study aims to create an AI model to assist clinicians in detecting pharyngeal fish bones more efficiently under laryngoscopy. METHODS Totally 3133 laryngoscopic images related to fish bones were collected for model training and validation. The images in the training dataset were trained using the YOLO-V5 algorithm model. After training, the model was validated and its performance was evaluated using a test dataset. The model's predictions were compared to those of human experts. Seven laryngoscopic videos related to fish bone were used to validate real-time target detection by the model. RESULTS The model trained in YOLO-V5 demonstrated good generalization and performance, with an average precision of 0.857 when the intersection over union (IOU) threshold was set to 0.5. The precision, recall rate, and F1 scores of the model are 0.909, 0.818, and 0.87, respectively. The overall accuracy of the model in the validation set was 0.821, comparable to that of ENT specialists. The model processed each image in 0.012 s, significantly faster than human processing (p < 0.001). Furthermore, the model exhibited outstanding performance in video recognition. CONCLUSION Our AI model based on YOLO-V5 effectively identifies and localizes fish bone foreign bodies in static laryngoscopic images and dynamic videos. It shows great potential for clinical application. LEVEL OF EVIDENCE 3 Laryngoscope, 134:2162-2169, 2024.
Collapse
Affiliation(s)
- Xiaoyao Tao
- Otorhinolaryngology Head and Neck Surgery Department, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | - Xu Zhao
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Hairui Liu
- School of Information Engineering, China University of Geosciences, Beijing, China
| | - Jinqiao Wang
- Institute of Automation, Chinese Academy of Sciences, Beijing, China
| | - Chunhui Tian
- Otolaryngology-Head and Neck Surgery Department, Suzhou Hospital of Anhui Medical University, Suzhou, China
| | - Longsheng Liu
- Otolaryngology-Head and Neck Surgery Department, Chaohu Hospital of Anhui Medical University, Hefei, China
| | - Yujie Ding
- Otolaryngology-Head and Neck Surgery Department, Feixi County People's Hospital, Hefei, China
| | - Xue Chen
- Otolaryngology-Head and Neck Surgery Department, Feidong County People's Hospital, Hefei, China
| | - Yehai Liu
- Otorhinolaryngology Head and Neck Surgery Department, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| |
Collapse
|
16
|
You Z, Han B, Shi Z, Zhao M, Du S, Liu H, Hei X, Ren X, Yan Y. Vocal Cord Leukoplakia Classification Using Siamese Network Under Small Samples of White Light Endoscopy Images. Otolaryngol Head Neck Surg 2024; 170:1099-1108. [PMID: 38037413 DOI: 10.1002/ohn.591] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Revised: 10/25/2023] [Accepted: 11/02/2023] [Indexed: 12/02/2023]
Abstract
OBJECTIVE Accurate vocal cord leukoplakia classification is instructive for clinical diagnosis and surgical treatment. This article introduces a reliable very deep Siamese network for accurate vocal cord leukoplakia classification. STUDY DESIGN A study of a classification network based on a retrospective database. SETTING Academic university and hospital. METHODS The white light image datasets of vocal cord leukoplakia used in this article were classified into 6 classes: normal tissues, inflammatory keratosis, mild dysplasia, moderate dysplasia, severe dysplasia, and squamous cell carcinoma. The classification performance was assessed by comparing it with 6 classical deep learning models, including AlexNet, VGG Net, Google Inception, ResNet, DenseNet, and Vision Transformer. RESULTS Experiments show the superior classification performance of our proposed network compared to state-of-the-art methods. The overall accuracy is 0.9756. The values of sensitivity and specificity are very high as well. The confusion matrix provides information for the 6-class classification task and demonstrates the superiority of our proposed network. CONCLUSION Our very deep Siamese network can provide accurate classification results of vocal cord leukoplakia, which facilitates early detection, clinical diagnosis, and surgical treatment. The excellent performance obtained in white light images can reduce the cost for patients, especially those living in developing countries.
Collapse
Affiliation(s)
- Zhenzhen You
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Botao Han
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Zhenghao Shi
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Minghua Zhao
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Shuangli Du
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Haiqin Liu
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| | - Xinhong Hei
- Scool of Computer Science and Engineering, Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Xiaoyong Ren
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| | - Yan Yan
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| |
Collapse
|
17
|
Marchi F, Bellini E, Iandelli A, Sampieri C, Peretti G. Exploring the landscape of AI-assisted decision-making in head and neck cancer treatment: a comparative analysis of NCCN guidelines and ChatGPT responses. Eur Arch Otorhinolaryngol 2024; 281:2123-2136. [PMID: 38421392 DOI: 10.1007/s00405-024-08525-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2023] [Accepted: 02/02/2024] [Indexed: 03/02/2024]
Abstract
PURPOSE Recent breakthroughs in natural language processing and machine learning, exemplified by ChatGPT, have spurred a paradigm shift in healthcare. Released by OpenAI in November 2022, ChatGPT rapidly gained global attention. Trained on massive text datasets, this large language model holds immense potential to revolutionize healthcare. However, existing literature often overlooks the need for rigorous validation and real-world applicability. METHODS This head-to-head comparative study assesses ChatGPT's capabilities in providing therapeutic recommendations for head and neck cancers. Simulating every NCCN Guidelines scenarios. ChatGPT is queried on primary treatments, adjuvant treatment, and follow-up, with responses compared to the NCCN Guidelines. Performance metrics, including sensitivity, specificity, and F1 score, are employed for assessment. RESULTS The study includes 68 hypothetical cases and 204 clinical scenarios. ChatGPT exhibits promising capabilities in addressing NCCN-related queries, achieving high sensitivity and overall accuracy across primary treatment, adjuvant treatment, and follow-up. The study's metrics showcase robustness in providing relevant suggestions. However, a few inaccuracies are noted, especially in primary treatment scenarios. CONCLUSION Our study highlights the proficiency of ChatGPT in providing treatment suggestions. The model's alignment with the NCCN Guidelines sets the stage for a nuanced exploration of AI's evolving role in oncological decision support. However, challenges related to the interpretability of AI in clinical decision-making and the importance of clinicians understanding the underlying principles of AI models remain unexplored. As AI continues to advance, collaborative efforts between models and medical experts are deemed essential for unlocking new frontiers in personalized cancer care.
Collapse
Affiliation(s)
- Filippo Marchi
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Largo Rosanna Benzi, 10, 16132, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, 16132, Genoa, Italy
| | - Elisa Bellini
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Largo Rosanna Benzi, 10, 16132, Genoa, Italy.
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, 16132, Genoa, Italy.
| | - Andrea Iandelli
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Largo Rosanna Benzi, 10, 16132, Genoa, Italy
| | - Claudio Sampieri
- Department of Experimental Medicine (DIMES), University of Genoa, Genoa, Italy
- Department of Otolaryngology-Hospital Cliníc, Barcelona, Spain
- Functional Unit of Head and Neck Tumors-Hospital Cliníc, Barcelona, Spain
| | - Giorgio Peretti
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Largo Rosanna Benzi, 10, 16132, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, 16132, Genoa, Italy
| |
Collapse
|
18
|
Chen TH, Wang YT, Wu CH, Kuo CF, Cheng HT, Huang SW, Lee C. A colonial serrated polyp classification model using white-light ordinary endoscopy images with an artificial intelligence model and TensorFlow chart. BMC Gastroenterol 2024; 24:99. [PMID: 38443794 PMCID: PMC10913269 DOI: 10.1186/s12876-024-03181-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Accepted: 02/19/2024] [Indexed: 03/07/2024] Open
Abstract
In this study, we implemented a combination of data augmentation and artificial intelligence (AI) model-Convolutional Neural Network (CNN)-to help physicians classify colonic polyps into traditional adenoma (TA), sessile serrated adenoma (SSA), and hyperplastic polyp (HP). We collected ordinary endoscopy images under both white and NBI lights. Under white light, we collected 257 images of HP, 423 images of SSA, and 60 images of TA. Under NBI light, were collected 238 images of HP, 284 images of SSA, and 71 images of TA. We implemented the CNN-based artificial intelligence model, Inception V4, to build a classification model for the types of colon polyps. Our final AI classification model with data augmentation process is constructed only with white light images. Our classification prediction accuracy of colon polyp type is 94%, and the discriminability of the model (area under the curve) was 98%. Thus, we can conclude that our model can help physicians distinguish between TA, SSA, and HPs and correctly identify precancerous lesions such as TA and SSA.
Collapse
Affiliation(s)
- Tsung-Hsing Chen
- Department of Gastroenterology and Hepatology, Linkou Medical Center, Chang Gung Memorial Hospital, Taoyuan, Taiwan
- College of Medicine, Chang Gung University, Taoyuan, Taiwan
| | | | - Chi-Huan Wu
- Department of Gastroenterology and Hepatology, Linkou Medical Center, Chang Gung Memorial Hospital, Taoyuan, Taiwan
- College of Medicine, Chang Gung University, Taoyuan, Taiwan
| | - Chang-Fu Kuo
- Division of Rheumatology, Allergy, and Immunology, Chang Gung Memorial Hospital- Linkou and Chang Gung University College of Medicine, Taoyuan, Taiwan, ROC
- Center for Artificial Intelligence in Medicine, Chang Gung Memorial Hospital, Taoyuan, Taiwan, ROC
| | - Hao-Tsai Cheng
- Department of Gastroenterology and Hepatology, Linkou Medical Center, Chang Gung Memorial Hospital, Taoyuan, Taiwan
- College of Medicine, Chang Gung University, Taoyuan, Taiwan
- Division of Gastroenterology and Hepatology, Department of Internal Medicine, New Taipei Municipal TuCheng Hospital, New Taipei City, Taiwan
- Graduate Institute of Clinical Medicine, College of Medicine, Chang Gung University, Taoyuan City, Taiwan
| | - Shu-Wei Huang
- Department of Gastroenterology and Hepatology, Linkou Medical Center, Chang Gung Memorial Hospital, Taoyuan, Taiwan
- College of Medicine, Chang Gung University, Taoyuan, Taiwan
- Division of Gastroenterology and Hepatology, Department of Internal Medicine, New Taipei Municipal TuCheng Hospital, New Taipei City, Taiwan
| | - Chieh Lee
- Department of Information and Management, College of Business, National Sun Yat-sen University, Kaohsiung city, Taiwan.
| |
Collapse
|
19
|
Staníková L, Kántor P, Fedorová K, Zeleník K, Komínek P. Clinical significance of type IV vascularization of laryngeal lesions according to the Ni classification. Front Oncol 2024; 14:1222827. [PMID: 38333687 PMCID: PMC10851150 DOI: 10.3389/fonc.2024.1222827] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 01/09/2024] [Indexed: 02/10/2024] Open
Abstract
Background Scattered, small, dot-like intraepithelial papillary capillary loops (IPCLs) represent type IV epithelial vascularization according to "Ni classification" and are considered to be nonmalignant. According to the European Laryngological Society classification, these loops are malignant vascular changes. This contradiction has high clinical importance; therefore, clarification of the clinical significance of type IV vascularization according to the Ni classification is needed. Methods The study was performed between June 2015 and December 2022. All recruited patients (n = 434) were symptomatic, with macroscopic laryngeal lesions (n = 674). Patients were investigated using the enhanced endoscopic methods of narrow band imaging (NBI) and the Storz Professional Image Enhancement System (IMAGE1 S). The microvascular patterns in the lesions were categorized according to Ni classification from 2011 and all lesions were examined histologically. Results A total of 674 lesions (434 patients) were investigated using flexible NBI endoscopy and IMAGE1 S endoscopy. Type IV vascularization was recognized in 293/674 (43.5%) lesions. Among these 293 lesions, 178 (60.7%) were benign (chronic laryngitis, hyperplasia, hyperkeratosis, polyps, cysts, granulomas, Reinkeho oedema and recurrent respiratory papillomatosis); 9 (3.1%) were squamous cell carcinoma; 61 (20.8%) were mildly dysplastic, 29 (9.9%) were moderately dysplastic, 14 (4.8%) were severe dysplastic and 2 (0.7%) were carcinoma in situ. The ability to recognize histologically benign lesions in group of nonmalignant vascular pattern according to Ni (vascularization type I-IV) and distinguish them from precancers and malignancies was with accuracy 75.5%, sensitivity 54.4%, specificity 94.4%, positive predictive value 89.6% and negative predictive value 69.9%. Conclusion Laryngeal lesions with type IV vascularization as defined by Ni present various histological findings, including precancerous and malignant lesions. Patients with type IV vascularization must be followed carefully and, in case of progression mucosal lesion microlaryngoscopy and excision are indicated.
Collapse
Affiliation(s)
- Lucia Staníková
- Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Ostrava, Ostrava, Czechia
- Department of Craniofacial Surgery, Faculty of Medicine, University of Ostrava, Ostrava, Czechia
| | - Peter Kántor
- Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Ostrava, Ostrava, Czechia
- Department of Craniofacial Surgery, Faculty of Medicine, University of Ostrava, Ostrava, Czechia
| | - Katarína Fedorová
- Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Ostrava, Ostrava, Czechia
| | - Karol Zeleník
- Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Ostrava, Ostrava, Czechia
- Department of Craniofacial Surgery, Faculty of Medicine, University of Ostrava, Ostrava, Czechia
| | - Pavel Komínek
- Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Ostrava, Ostrava, Czechia
- Department of Craniofacial Surgery, Faculty of Medicine, University of Ostrava, Ostrava, Czechia
| |
Collapse
|
20
|
Wang SX, Li Y, Zhu JQ, Wang ML, Zhang W, Tie CW, Wang GQ, Ni XG. The Detection of Nasopharyngeal Carcinomas Using a Neural Network Based on Nasopharyngoscopic Images. Laryngoscope 2024; 134:127-135. [PMID: 37254946 DOI: 10.1002/lary.30781] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2022] [Revised: 05/14/2023] [Accepted: 05/15/2023] [Indexed: 06/01/2023]
Abstract
OBJECTIVE To construct and validate a deep convolutional neural network (DCNN)-based artificial intelligence (AI) system for the detection of nasopharyngeal carcinoma (NPC) using archived nasopharyngoscopic images. METHODS We retrospectively collected 14107 nasopharyngoscopic images (7108 NPCs and 6999 noncancers) to construct a DCNN model and prepared a validation dataset containing 3501 images (1744 NPCs and 1757 noncancers) from a single center between January 2009 and December 2020. The DCNN model was established using the You Only Look Once (YOLOv5) architecture. Four otolaryngologists were asked to review the images of the validation set to benchmark the DCNN model performance. RESULTS The DCNN model analyzed the 3501 images in 69.35 s. For the validation dataset, the precision, recall, accuracy, and F1 score of the DCNN model in the detection of NPCs on white light imaging (WLI) and narrow band imaging (NBI) were 0.845 ± 0.038, 0.942 ± 0.021, 0.920 ± 0.024, and 0.890 ± 0.045, and 0.895 ± 0.045, 0.941 ± 0.018, and 0.975 ± 0.013, 0.918 ± 0.036, respectively. The diagnostic outcome of the DCNN model on WLI and NBI images was significantly higher than that of two junior otolaryngologists (p < 0.05). CONCLUSION The DCNN model showed better diagnostic outcomes for NPCs than those of junior otolaryngologists. Therefore, it could assist them in improving their diagnostic level and reducing missed diagnoses. LEVEL OF EVIDENCE 3 Laryngoscope, 134:127-135, 2024.
Collapse
Affiliation(s)
- Shi-Xu Wang
- Department of Head and Neck Surgery, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Ying Li
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Ji-Qing Zhu
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Mei-Ling Wang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Wei Zhang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Cheng-Wei Tie
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Gui-Qi Wang
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| | - Xiao-Guang Ni
- Department of Endoscopy, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China
| |
Collapse
|
21
|
Kryukov AI, Sudarev PA, Romanenko SG, Kurbanova DI, Lesogorova EV, Krasilnikova EN, Pavlikhin OG, Ivanova AA, Osadchiy AP, Shevyrina NG. [Diagnosis of benign laryngeal tumors using neural network]. Vestn Otorinolaringol 2024; 89:24-28. [PMID: 39104269 DOI: 10.17116/otorino20248903124] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/07/2024]
Abstract
The article describes our experience in developing and training an artificial neural network based on artificial intelligence algorithms for recognizing the characteristic features of benign laryngeal tumors and variants of the norm of the larynx based on the analysis of laryngoscopy pictures obtained during the examination of patients. During the preparation of data for training the neural network, a dataset was collected, labeled and loaded, consisting of 1471 images of the larynx in digital formats (jpg, bmp). Next, the neural network was trained and tested in order to recognize images of the norm and neoplasms of the larynx. The developed and trained artificial neural network demonstrated an accuracy of 86% in recognizing of benign laryngeal tumors and variants of the norm of the larynx. The proposed technology can be further used in practical healthcare to control and improve the quality of diagnosis of laryngeal pathologies.
Collapse
Affiliation(s)
- A I Kryukov
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - P A Sudarev
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - S G Romanenko
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - D I Kurbanova
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - E V Lesogorova
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - E N Krasilnikova
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | - O G Pavlikhin
- Sverzhevsky Research Clinical Institute of Otorhinolaryngology, Moscow, Russia
| | | | | | | |
Collapse
|
22
|
You Z, Han B, Shi Z, Zhao M, Du S, Yan J, Liu H, Hei X, Ren X, Yan Y. Vocal cord leukoplakia classification using deep learning models in white light and narrow band imaging endoscopy images. Head Neck 2023; 45:3129-3145. [PMID: 37837264 DOI: 10.1002/hed.27543] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2023] [Revised: 09/15/2023] [Accepted: 09/29/2023] [Indexed: 10/15/2023] Open
Abstract
BACKGROUND Accurate vocal cord leukoplakia classification is critical for the individualized treatment and early detection of laryngeal cancer. Numerous deep learning techniques have been proposed, but it is unclear how to select one to apply in the laryngeal tasks. This article introduces and reliably evaluates existing deep learning models for vocal cord leukoplakia classification. METHODS We created white light and narrow band imaging (NBI) image datasets of vocal cord leukoplakia which were classified into six classes: normal tissues (NT), inflammatory keratosis (IK), mild dysplasia (MiD), moderate dysplasia (MoD), severe dysplasia (SD), and squamous cell carcinoma (SCC). Vocal cord leukoplakia classification was performed using six classical deep learning models, AlexNet, VGG, Google Inception, ResNet, DenseNet, and Vision Transformer. RESULTS GoogLeNet (i.e., Google Inception V1), DenseNet-121, and ResNet-152 perform excellent classification. The highest overall accuracy of white light image classification is 0.9583, while the highest overall accuracy of NBI image classification is 0.9478. These three neural networks all provide very high sensitivity, specificity, and precision values. CONCLUSION GoogLeNet, ResNet, and DenseNet can provide accurate pathological classification of vocal cord leukoplakia. It facilitates early diagnosis, providing judgment on conservative treatment or surgical treatment of different degrees, and reducing the burden on endoscopists.
Collapse
Affiliation(s)
- Zhenzhen You
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Botao Han
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Zhenghao Shi
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Minghua Zhao
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Shuangli Du
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Jing Yan
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| | - Haiqin Liu
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| | - Xinhong Hei
- Shaanxi Key Laboratory for Network Computing and Security Technology, School of Computer Science and Engineering, Xi'an University of Technology, Xi'an, China
| | - Xiaoyong Ren
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| | - Yan Yan
- Department of Otorhinolaryngology, Second Affiliated Hospital of Medical College, Xi'an Jiaotong University, Xi'an, China
| |
Collapse
|
23
|
Tsilivigkos C, Athanasopoulos M, Micco RD, Giotakis A, Mastronikolis NS, Mulita F, Verras GI, Maroulis I, Giotakis E. Deep Learning Techniques and Imaging in Otorhinolaryngology-A State-of-the-Art Review. J Clin Med 2023; 12:6973. [PMID: 38002588 PMCID: PMC10672270 DOI: 10.3390/jcm12226973] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2023] [Revised: 11/02/2023] [Accepted: 11/06/2023] [Indexed: 11/26/2023] Open
Abstract
Over the last decades, the field of medicine has witnessed significant progress in artificial intelligence (AI), the Internet of Medical Things (IoMT), and deep learning (DL) systems. Otorhinolaryngology, and imaging in its various subspecialties, has not remained untouched by this transformative trend. As the medical landscape evolves, the integration of these technologies becomes imperative in augmenting patient care, fostering innovation, and actively participating in the ever-evolving synergy between computer vision techniques in otorhinolaryngology and AI. To that end, we conducted a thorough search on MEDLINE for papers published until June 2023, utilizing the keywords 'otorhinolaryngology', 'imaging', 'computer vision', 'artificial intelligence', and 'deep learning', and at the same time conducted manual searching in the references section of the articles included in our manuscript. Our search culminated in the retrieval of 121 related articles, which were subsequently subdivided into the following categories: imaging in head and neck, otology, and rhinology. Our objective is to provide a comprehensive introduction to this burgeoning field, tailored for both experienced specialists and aspiring residents in the domain of deep learning algorithms in imaging techniques in otorhinolaryngology.
Collapse
Affiliation(s)
- Christos Tsilivigkos
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Michail Athanasopoulos
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Riccardo di Micco
- Department of Otolaryngology and Head and Neck Surgery, Medical School of Hannover, 30625 Hannover, Germany;
| | - Aris Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| | - Nicholas S. Mastronikolis
- Department of Otolaryngology, University Hospital of Patras, 265 04 Patras, Greece; (M.A.); (N.S.M.)
| | - Francesk Mulita
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Georgios-Ioannis Verras
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Ioannis Maroulis
- Department of Surgery, University Hospital of Patras, 265 04 Patras, Greece; (G.-I.V.); (I.M.)
| | - Evangelos Giotakis
- 1st Department of Otolaryngology, National and Kapodistrian University of Athens, Hippocrateion Hospital, 115 27 Athens, Greece; (A.G.); (E.G.)
| |
Collapse
|
24
|
Wu Q, Wang X, Liang G, Luo X, Zhou M, Deng H, Zhang Y, Huang X, Yang Q. Advances in Image-Based Artificial Intelligence in Otorhinolaryngology-Head and Neck Surgery: A Systematic Review. Otolaryngol Head Neck Surg 2023; 169:1132-1142. [PMID: 37288505 DOI: 10.1002/ohn.391] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2023] [Revised: 04/27/2023] [Accepted: 05/13/2023] [Indexed: 06/09/2023]
Abstract
OBJECTIVE To update the literature and provide a systematic review of image-based artificial intelligence (AI) applications in otolaryngology, highlight its advances, and propose future challenges. DATA SOURCES Web of Science, Embase, PubMed, and Cochrane Library. REVIEW METHODS Studies written in English, published between January 2020 and December 2022. Two independent authors screened the search results, extracted data, and assessed studies. RESULTS Overall, 686 studies were identified. After screening titles and abstracts, 325 full-text studies were assessed for eligibility, and 78 studies were included in this systematic review. The studies originated from 16 countries. Among these countries, the top 3 were China (n = 29), Korea (n = 8), the United States, and Japan (n = 7 each). The most common area was otology (n = 35), followed by rhinology (n = 20), pharyngology (n = 18), and head and neck surgery (n = 5). Most applications of AI in otology, rhinology, pharyngology, and head and neck surgery mainly included chronic otitis media (n = 9), nasal polyps (n = 4), laryngeal cancer (n = 12), and head and neck squamous cell carcinoma (n = 3), respectively. The overall performance of AI in accuracy, the area under the curve, sensitivity, and specificity were 88.39 ± 9.78%, 91.91 ± 6.70%, 86.93 ± 11.59%, and 88.62 ± 14.03%, respectively. CONCLUSION This state-of-the-art review aimed to highlight the increasing applications of image-based AI in otorhinolaryngology head and neck surgery. The following steps will entail multicentre collaboration to ensure data reliability, ongoing optimization of AI algorithms, and integration into real-world clinical practice. Future studies should consider 3-dimensional (3D)-based AI, such as 3D surgical AI.
Collapse
Affiliation(s)
- Qingwu Wu
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xinyue Wang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Guixian Liang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xin Luo
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Min Zhou
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Huiyi Deng
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Yana Zhang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Xuekun Huang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| | - Qintai Yang
- Department of Otorhinolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
- Department of Allergy, The Third Affiliated Hospital of Sun Yat-sen University, Guangzhou, China
| |
Collapse
|
25
|
He Z, Zhang K, Zhao N, Wang Y, Hou W, Meng Q, Li C, Chen J, Li J. Deep learning for real-time detection of nasopharyngeal carcinoma during nasopharyngeal endoscopy. iScience 2023; 26:107463. [PMID: 37720094 PMCID: PMC10502364 DOI: 10.1016/j.isci.2023.107463] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2023] [Revised: 07/07/2023] [Accepted: 07/20/2023] [Indexed: 09/19/2023] Open
Abstract
Nasopharyngeal carcinoma (NPC) is known for high curability during early stage of the disease, and early diagnosis relies on nasopharyngeal endoscopy and subsequent pathological biopsy. To enhance the early diagnosis rate by aiding physicians in the real-time identification of NPC and directing biopsy site selection during endoscopy, we assembled a dataset comprising 2,429 nasopharyngeal endoscopy video frames from 690 patients across three medical centers. With these data, we developed a deep learning-based NPC detection model using the you only look once (YOLO) network. Our model demonstrated high performance, with precision, recall, mean average precision, and F1-score values of 0.977, 0.943, 0.977, and 0.960, respectively, for internal test set and 0.825, 0.743, 0.814, and 0.780 for external test set at 0.5 intersection over union. Remarkably, our model demonstrated a high inference speed (52.9 FPS), surpassing the average frame rate (25.0 FPS) of endoscopy videos, thus making real-time detection in endoscopy feasible.
Collapse
Affiliation(s)
- Zicheng He
- Department of Otolaryngology, the First Affiliated Hospital of Sun Yat-sen University, Guangzhou Key Laboratory of Otorhinolaryngology, Otorhinolaryngology Institute of Sun Yat-sen University, Guangzhou 510080, P.R.China
- Guangxi Hospital Division of The First Affiliated Hospital, Sun Yat-sen University, Nanning 530000, P.R.China
| | - Kai Zhang
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, P.R.China
| | - Nan Zhao
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, P.R.China
| | - Yongquan Wang
- Department of Otolaryngology, the First Affiliated Hospital of Sun Yat-sen University, Guangzhou Key Laboratory of Otorhinolaryngology, Otorhinolaryngology Institute of Sun Yat-sen University, Guangzhou 510080, P.R.China
| | | | - Qinxiang Meng
- Guangzhou First People’s Hospital, Guangzhou 510180, P.R.China
| | - Chunwei Li
- Department of Otolaryngology, the First Affiliated Hospital of Sun Yat-sen University, Guangzhou Key Laboratory of Otorhinolaryngology, Otorhinolaryngology Institute of Sun Yat-sen University, Guangzhou 510080, P.R.China
| | - Junzhou Chen
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, P.R.China
| | - Jian Li
- Department of Otolaryngology, the First Affiliated Hospital of Sun Yat-sen University, Guangzhou Key Laboratory of Otorhinolaryngology, Otorhinolaryngology Institute of Sun Yat-sen University, Guangzhou 510080, P.R.China
- Guangxi Hospital Division of The First Affiliated Hospital, Sun Yat-sen University, Nanning 530000, P.R.China
| |
Collapse
|
26
|
Li Y, Gu W, Yue H, Lei G, Guo W, Wen Y, Tang H, Luo X, Tu W, Ye J, Hong R, Cai Q, Gu Q, Liu T, Miao B, Wang R, Ren J, Lei W. Real-time detection of laryngopharyngeal cancer using an artificial intelligence-assisted system with multimodal data. J Transl Med 2023; 21:698. [PMID: 37805551 PMCID: PMC10559609 DOI: 10.1186/s12967-023-04572-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2023] [Accepted: 09/23/2023] [Indexed: 10/09/2023] Open
Abstract
BACKGROUND Laryngopharyngeal cancer (LPC) includes laryngeal and hypopharyngeal cancer, whose early diagnosis can significantly improve the prognosis and quality of life of patients. Pathological biopsy of suspicious cancerous tissue under the guidance of laryngoscopy is the gold standard for diagnosing LPC. However, this subjective examination largely depends on the skills and experience of laryngologists, which increases the possibility of missed diagnoses and repeated unnecessary biopsies. We aimed to develop and validate a deep convolutional neural network-based Laryngopharyngeal Artificial Intelligence Diagnostic System (LPAIDS) for real-time automatically identifying LPC in both laryngoscopy white-light imaging (WLI) and narrow-band imaging (NBI) images to improve the diagnostic accuracy of LPC by reducing diagnostic variation among on-expert laryngologists. METHODS All 31,543 laryngoscopic images from 2382 patients were categorised into training, verification, and test sets to develop, validate, and internal test LPAIDS. Another 25,063 images from five other hospitals were used as external tests. Overall, 551 videos were used to evaluate the real-time performance of the system, and 200 randomly selected videos were used to compare the diagnostic performance of the LPAIDS with that of laryngologists. Two deep-learning models using either WLI (model W) or NBI (model N) images were constructed to compare with LPAIDS. RESULTS LPAIDS had a higher diagnostic performance than models W and N, with accuracies of 0·956 and 0·949 in the internal image and video tests, respectively. The robustness and stability of LPAIDS were validated in external sets with the area under the receiver operating characteristic curve values of 0·965-0·987. In the laryngologist-machine competition, LPAIDS achieved an accuracy of 0·940, which was comparable to expert laryngologists and outperformed other laryngologists with varying qualifications. CONCLUSIONS LPAIDS provided high accuracy and stability in detecting LPC in real-time, which showed great potential for using LPAIDS to improve the diagnostic accuracy of LPC by reducing diagnostic variation among on-expert laryngologists.
Collapse
Affiliation(s)
- Yun Li
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China
| | - Wenxin Gu
- School of Computer Science and Engineering, Guangdong Province Key Lab of Computational Science, Sun Yat-Sen University, Guangzhou, 510006, Guangdong, China
| | - Huijun Yue
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China
| | - Guoqing Lei
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China
| | - Wenbin Guo
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China
| | - Yihui Wen
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China
| | - Haocheng Tang
- Department of Otolaryngology-Head and Neck Surgery, Nanfang Hospital, Southern Medical University, Guangzhou, Guangdong, China
| | - Xin Luo
- Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Wenjuan Tu
- Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Jin Ye
- Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Ruomei Hong
- Department of Otolaryngology-Head and Neck, Sun Yat-Sen Memorial Hospital, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Qian Cai
- Department of Otolaryngology-Head and Neck, Sun Yat-Sen Memorial Hospital, Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Qingyu Gu
- Department of Otorhinolaryngology-Head and Neck Surgery, The Sixth Affiliated Hospital of Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Tianrun Liu
- Department of Otorhinolaryngology-Head and Neck Surgery, The Sixth Affiliated Hospital of Sun Yat-Sen University, Guangzhou, Guangdong, China
| | - Beiping Miao
- Department of Otolaryngology-Head and Neck Surgery, Shenzhen Secondary Hospital and First Affiliated Hospital of Shenzhen University, Shenzhen, Guangdong, China
| | - Ruxin Wang
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, Guangdong, China
| | - Jiangtao Ren
- School of Computer Science and Engineering, Guangdong Province Key Lab of Computational Science, Sun Yat-Sen University, Guangzhou, 510006, Guangdong, China.
| | - Wenbin Lei
- Otorhinolaryngology Hospital, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, 510080, Guangdong, China.
| |
Collapse
|
27
|
Sampieri C, Baldini C, Azam MA, Moccia S, Mattos LS, Vilaseca I, Peretti G, Ioppi A. Artificial Intelligence for Upper Aerodigestive Tract Endoscopy and Laryngoscopy: A Guide for Physicians and State-of-the-Art Review. Otolaryngol Head Neck Surg 2023; 169:811-829. [PMID: 37051892 DOI: 10.1002/ohn.343] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Revised: 03/03/2023] [Accepted: 03/23/2023] [Indexed: 04/14/2023]
Abstract
OBJECTIVE The endoscopic and laryngoscopic examination is paramount for laryngeal, oropharyngeal, nasopharyngeal, nasal, and oral cavity benign lesions and cancer evaluation. Nevertheless, upper aerodigestive tract (UADT) endoscopy is intrinsically operator-dependent and lacks objective quality standards. At present, there has been an increased interest in artificial intelligence (AI) applications in this area to support physicians during the examination, thus enhancing diagnostic performances. The relative novelty of this research field poses a challenge both for the reviewers and readers as clinicians often lack a specific technical background. DATA SOURCES Four bibliographic databases were searched: PubMed, EMBASE, Cochrane, and Google Scholar. REVIEW METHODS A structured review of the current literature (up to September 2022) was performed. Search terms related to topics of AI, machine learning (ML), and deep learning (DL) in UADT endoscopy and laryngoscopy were identified and queried by 3 independent reviewers. Citations of selected studies were also evaluated to ensure comprehensiveness. CONCLUSIONS Forty-one studies were included in the review. AI and computer vision techniques were used to achieve 3 fundamental tasks in this field: classification, detection, and segmentation. All papers were summarized and reviewed. IMPLICATIONS FOR PRACTICE This article comprehensively reviews the latest developments in the application of ML and DL in UADT endoscopy and laryngoscopy, as well as their future clinical implications. The technical basis of AI is also explained, providing guidance for nonexpert readers to allow critical appraisal of the evaluation metrics and the most relevant quality requirements.
Collapse
Affiliation(s)
- Claudio Sampieri
- Department of Experimental Medicine (DIMES), University of Genoa, Genoa, Italy
- Functional Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain
- Otorhinolaryngology Department, Hospital Clínic, Barcelona, Spain
| | - Chiara Baldini
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Dipartimento di Informatica, Bioingegneria, Robotica e Ingegneria dei Sistemi (DIBRIS), University of Genoa, Genoa, Italy
| | - Muhammad Adeel Azam
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
- Dipartimento di Informatica, Bioingegneria, Robotica e Ingegneria dei Sistemi (DIBRIS), University of Genoa, Genoa, Italy
| | - Sara Moccia
- Department of Excellence in Robotics and AI, The BioRobotics Institute, Pisa, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Isabel Vilaseca
- Functional Unit of Head and Neck Tumors, Hospital Clínic, Barcelona, Spain
- Otorhinolaryngology Department, Hospital Clínic, Barcelona, Spain
- Head Neck Clínic, Agència de Gestió d'Ajuts Universitaris i de Recerca, Barcelona, Catalunya, Spain
- Surgery and Medical-Surgical Specialties Department, Faculty of Medicine and Health Sciences, Universitat de Barcelona, Barcelona, Spain
- Translational Genomics and Target Therapies in Solid Tumors Group, Faculty of Medicine, Institut d́Investigacions Biomèdiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
- University of Barcelona, Barcelona, Spain
| | - Giorgio Peretti
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Alessandro Ioppi
- Unit of Otorhinolaryngology-Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
- Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| |
Collapse
|
28
|
Zhong NN, Wang HQ, Huang XY, Li ZZ, Cao LM, Huo FY, Liu B, Bu LL. Enhancing head and neck tumor management with artificial intelligence: Integration and perspectives. Semin Cancer Biol 2023; 95:52-74. [PMID: 37473825 DOI: 10.1016/j.semcancer.2023.07.002] [Citation(s) in RCA: 20] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2023] [Revised: 07/11/2023] [Accepted: 07/15/2023] [Indexed: 07/22/2023]
Abstract
Head and neck tumors (HNTs) constitute a multifaceted ensemble of pathologies that primarily involve regions such as the oral cavity, pharynx, and nasal cavity. The intricate anatomical structure of these regions poses considerable challenges to efficacious treatment strategies. Despite the availability of myriad treatment modalities, the overall therapeutic efficacy for HNTs continues to remain subdued. In recent years, the deployment of artificial intelligence (AI) in healthcare practices has garnered noteworthy attention. AI modalities, inclusive of machine learning (ML), neural networks (NNs), and deep learning (DL), when amalgamated into the holistic management of HNTs, promise to augment the precision, safety, and efficacy of treatment regimens. The integration of AI within HNT management is intricately intertwined with domains such as medical imaging, bioinformatics, and medical robotics. This article intends to scrutinize the cutting-edge advancements and prospective applications of AI in the realm of HNTs, elucidating AI's indispensable role in prevention, diagnosis, treatment, prognostication, research, and inter-sectoral integration. The overarching objective is to stimulate scholarly discourse and invigorate insights among medical practitioners and researchers to propel further exploration, thereby facilitating superior therapeutic alternatives for patients.
Collapse
Affiliation(s)
- Nian-Nian Zhong
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Han-Qi Wang
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Xin-Yue Huang
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Zi-Zhan Li
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Lei-Ming Cao
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Fang-Yi Huo
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China
| | - Bing Liu
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China; Department of Oral & Maxillofacial - Head Neck Oncology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China.
| | - Lin-Lin Bu
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China; Department of Oral & Maxillofacial - Head Neck Oncology, School & Hospital of Stomatology, Wuhan University, Wuhan 430079, China.
| |
Collapse
|
29
|
Liu Z, Zhang J, Wang N, Feng Y, Tang F, Li T, Lv L, Li H, Wang W, Liu Y. Enhanced YOLOv5 network-based object detection (BALFilter Reader) promotes PERFECT filter-enabled liquid biopsy of lung cancer from bronchoalveolar lavage fluid (BALF). MICROSYSTEMS & NANOENGINEERING 2023; 9:121. [PMID: 37786899 PMCID: PMC10541878 DOI: 10.1038/s41378-023-00580-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/22/2023] [Revised: 06/24/2023] [Accepted: 06/24/2023] [Indexed: 10/04/2023]
Abstract
Liquid biopsy of cancers, detecting tumor-related information from liquid samples, has attracted wide attentions as an emerging technology. Our previously reported large-area PERFECT (Precise-Efficient-Robust-Flexible-Easy-Controllable-Thin) filter has demonstrated competitive sensitivity in recovering rare tumor cells from clinical samples. However, it is time-consuming and easily biased to manually inspect rare target cells among numerous background cells distributed in a large area (Φ ≥ 13 mm). This puts forward an urgent demand for rapid and bias-free inspection. Hereby, this paper implemented deep learning-based object detection for the inspection of rare tumor cells from large-field images of PERFECT filters with hematoxylin-eosin (HE)-stained cells recovered from bronchoalveolar lavage fluid (BALF). CenterNet, EfficientDet, and YOLOv5 were trained and validated with 240 and 60 image blocks containing tumor and/or background cells, respectively. YOLOv5 was selected as the basic network given the highest mAP@0.5 of 92.1%, compared to those of CenterNet and EfficientDet at 85.2% and 91.6%, respectively. Then, tricks including CIoU loss, image flip, mosaic, HSV augmentation and TTA were applied to enhance the performance of the YOLOv5 network, improving mAP@0.5 to 96.2%. This enhanced YOLOv5 network-based object detection, named as BALFilter Reader, was tested and cross-validated on 24 clinical cases. The overall diagnosis performance (~2 min) with sensitivity@66.7% ± 16.7%, specificity@100.0% ± 0.0% and accuracy@75.0% ± 12.5% was superior to that from two experienced pathologists (10-30 min) with sensitivity@61.1%, specificity@16.7% and accuracy@50.0%, with the histopathological result as the gold standard. The AUC of the BALFilter Reader is 0.84 ± 0.08. Moreover, a customized Web was developed for a user-friendly interface and the promotion of wide applications. The current results revealed that the developed BALFilter Reader is a rapid, bias-free and easily accessible AI-enabled tool to promote the transplantation of the BALFilter technique. This work can easily expand to other cytopathological diagnoses and improve the application value of micro/nanotechnology-based liquid biopsy in the era of intelligent pathology.
Collapse
Affiliation(s)
- Zheng Liu
- School of Software and Microelectronics, Peking University, Beijing, 100871 China
| | - Jixin Zhang
- Department of Pathology, Peking University First Hospital, Beijing, 100034 China
| | - Ningyu Wang
- School of Integrated Circuits, Peking University, Beijing, 100871 China
| | - Yun’ai Feng
- Department of Respirology and Critical Care Medicine, Peking University First Hospital, Beijing, 100034 China
| | - Fei Tang
- Department of Interventional Lung Disease and Center of Endoscopic Diagnosis and Treatment, Anhui Chest Hospital, Hefei, Anhui 230022 China
| | - Tingyu Li
- School of Integrated Circuits, Peking University, Beijing, 100871 China
| | - Liping Lv
- Department of Interventional Lung Disease and Center of Endoscopic Diagnosis and Treatment, Anhui Chest Hospital, Hefei, Anhui 230022 China
| | - Haichao Li
- Department of Respirology and Critical Care Medicine, Peking University First Hospital, Beijing, 100034 China
| | - Wei Wang
- School of Integrated Circuits, Peking University, Beijing, 100871 China
- National Key Laboratory of Advanced Micro and Nano Manufacture Technology, Beijing, 100871 China
- Frontiers Science Center for Nano-optoelectronics, Peking University, Beijing, 100871 China
| | - Yaoping Liu
- School of Integrated Circuits, Peking University, Beijing, 100871 China
- AntiMicrobial Resistance (AMR) and Critical Analytics for Manufacturing Personalized-Medicine (CAMP) IRGs, Singapore-MIT Alliance for Research and Technology (SMART) Center, Singapore, 138602 Singapore
| |
Collapse
|
30
|
Filipovsky T, Kalfert D, Lukavcova E, Zavazalova S, Hlozek J, Kovar D, Astl J, Holy R. The importance of preoperative and perioperative Narrow Band Imaging endoscopy in the diagnosis of pre-tumor and tumor lesions of the larynx. J Appl Biomed 2023; 21:107-112. [PMID: 37747310 DOI: 10.32725/jab.2023.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Accepted: 09/13/2023] [Indexed: 09/26/2023] Open
Abstract
INTRODUCTION Narrow band imaging (NBI) is an endoscopic imaging method intended for the diagnosis of mucosal lesions of the larynx that are not visible in white-light endoscopy, but are typical of pre-tumor and tumor lesions of the larynx. THE PURPOSE OF THE STUDY To compare preoperative/perioperative white light endoscopy and NBI endoscopy with the results of histopathological examinations in pre-tumor and tumor lesions of the larynx. METHODS A prospective study, over a period of five years (5/2018-5/2023), included 87 patients with laryngeal lesions aged 24-80 years. We evaluated preoperative/ perioperative white light and NBI endoscopy, established a working prehistological diagnosis, and compared this with the definitive histopathological results of laryngeal biopsies. RESULTS In relation to the definitive histology score, a statistically significant correlation was found between the evaluation of the finding and the definitive histology for preoperative and perioperative white light endoscopy and NBI endoscopy (p < 0.001). Both methods showed higher precision when used perioperatively. CONCLUSION NBI endoscopy is an optical method that allows us to improve the diagnosis of laryngeal lesions, perform a controlled perioperative biopsy, and refine the surgical scope. The NBI endoscopy is a suitable method for the diagnosis of early cancerous lesions of the larynx. The use of preoperative/perioperative NBI endoscopy allowed us to achieve a high level of agreement correlation (p < 0.001) between the prehistological working diagnosis and the final histopathological result. The NBI method proves its application in the diagnosis of pre-tumor and tumor lesions of the larynx.
Collapse
Affiliation(s)
- Tomas Filipovsky
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| | - David Kalfert
- Motol University Hospital, Department of Otorhinolaryngology and Head and Neck Surgery, Prague, Czech Republic
- Charles University, First Faculty of Medicine, Prague, Czech Republic
| | - Eva Lukavcova
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
| | - Sarka Zavazalova
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| | - Jiri Hlozek
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| | - Daniel Kovar
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| | - Jaromir Astl
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| | - Richard Holy
- Military University Hospital Prague, Department of Otorhinolaryngology and Maxillofacial Surgery, Prague, Czech Republic
- Charles University, Third Faculty of Medicine, Prague, Czech Republic
| |
Collapse
|
31
|
Wellenstein DJ, Woodburn J, Marres HAM, van den Broek GB. Detection of laryngeal carcinoma during endoscopy using artificial intelligence. Head Neck 2023; 45:2217-2226. [PMID: 37377069 DOI: 10.1002/hed.27441] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/21/2022] [Revised: 04/25/2023] [Accepted: 06/18/2023] [Indexed: 06/29/2023] Open
Abstract
BACKGROUND The objective of this study was to assess the performance and application of a self-developed deep learning (DL) algorithm for the real-time localization and classification of both vocal cord carcinoma and benign vocal cord lesions. METHODS The algorithm was trained and validated upon a dataset of videos and photos collected from our own department, as well as an open-access dataset named "Laryngoscope8". RESULTS The algorithm correctly localizes and classifies vocal cord carcinoma on still images with a sensitivity between 71% and 78% and benign vocal cord lesions with a sensitivity between 70% and 82%. Furthermore, the best algorithm had an average frame per second rate of 63, thus making it suitable to use in an outpatient clinic setting for real-time detection of laryngeal pathology. CONCLUSION We have demonstrated that our developed DL algorithm is able to localize and classify benign and malignant laryngeal pathology during endoscopy.
Collapse
Affiliation(s)
- David J Wellenstein
- Department of Otorhinolaryngology and Head and Neck Surgery, Radboud University Medical Center, Nijmegen, The Netherlands
| | | | - Henri A M Marres
- Department of Otorhinolaryngology and Head and Neck Surgery, Radboud University Medical Center, Nijmegen, The Netherlands
| | - Guido B van den Broek
- Department of Otorhinolaryngology and Head and Neck Surgery, Radboud University Medical Center, Nijmegen, The Netherlands
- Department of Information Management, Radboud University Medical Center, Nijmegen, The Netherlands
| |
Collapse
|
32
|
Kim GH, Hwang YJ, Lee H, Sung ES, Nam KW. Convolutional neural network-based vocal cord tumor classification technique for home-based self-prescreening purpose. Biomed Eng Online 2023; 22:81. [PMID: 37596652 PMCID: PMC10439563 DOI: 10.1186/s12938-023-01139-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Accepted: 07/20/2023] [Indexed: 08/20/2023] Open
Abstract
BACKGROUND In this study, we proposed a deep learning technique that can simultaneously detect suspicious positions of benign vocal cord tumors in laparoscopic images and classify the types of tumors into cysts, granulomas, leukoplakia, nodules and polyps. This technique is useful for simplified home-based self-prescreening purposes to detect the generation of tumors around the vocal cord early in the benign stage. RESULTS We implemented four convolutional neural network (CNN) models (two Mask R-CNNs, Yolo V4, and a single-shot detector) that were trained, validated and tested using 2183 laryngoscopic images. The experimental results demonstrated that among the four applied models, Yolo V4 showed the highest F1-score for all tumor types (0.7664, cyst; 0.9875, granuloma; 0.8214, leukoplakia; 0.8119, nodule; and 0.8271, polyp). The model with the lowest false-negative rate was different for each tumor type (Yolo V4 for cysts/granulomas and Mask R-CNN for leukoplakia/nodules/polyps). In addition, the embedded-operated Yolo V4 model showed an approximately equivalent F1-score (0.8529) to that of the computer-operated Yolo-4 model (0.8683). CONCLUSIONS Based on these results, we conclude that the proposed deep-learning-based home screening techniques have the potential to aid in the early detection of tumors around the vocal cord and can improve the long-term survival of patients with vocal cord tumors.
Collapse
Affiliation(s)
- Gun Ho Kim
- Medical Research Institute, Pusan National University, Yangsan, Korea
- Department of Biomedical Engineering, Pusan National University Yangsan Hospital, Yangsan, Korea
| | - Young Jun Hwang
- Department of Biomedical Engineering, School of Medicine, Pusan National University, 49, Busandaehak-Ro, Mulgeum-Eup, Yangsan, 50629, Korea
| | - Hongje Lee
- Department of Nuclear Medicine, Dongnam Institute of Radiological & Medical Sciences, Busan, Korea
| | - Eui-Suk Sung
- Department of Otolaryngology-Head and Neck Surgery, Pusan National University Yangsan Hospital, Yangsan, Korea.
- Department of Otolaryngology-Head and Neck Surgery, School of Medicine, Pusan National University, Yangsan, Korea.
- Research Institute for Convergence of Biomedical Science and Technology, Pusan National University Yangsan Hospital, Yangsan, Korea.
| | - Kyoung Won Nam
- Department of Biomedical Engineering, Pusan National University Yangsan Hospital, Yangsan, Korea.
- Department of Biomedical Engineering, School of Medicine, Pusan National University, 49, Busandaehak-Ro, Mulgeum-Eup, Yangsan, 50629, Korea.
- Research Institute for Convergence of Biomedical Science and Technology, Pusan National University Yangsan Hospital, Yangsan, Korea.
| |
Collapse
|
33
|
Paderno A, Villani FP, Fior M, Berretti G, Gennarini F, Zigliani G, Ulaj E, Montenegro C, Sordi A, Sampieri C, Peretti G, Moccia S, Piazza C. Instance segmentation of upper aerodigestive tract cancer: site-specific outcomes. ACTA OTORHINOLARYNGOLOGICA ITALICA : ORGANO UFFICIALE DELLA SOCIETA ITALIANA DI OTORINOLARINGOLOGIA E CHIRURGIA CERVICO-FACCIALE 2023; 43:283-290. [PMID: 37488992 PMCID: PMC10366566 DOI: 10.14639/0392-100x-n2336] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/14/2022] [Accepted: 01/08/2023] [Indexed: 07/26/2023]
Abstract
Objective To achieve instance segmentation of upper aerodigestive tract (UADT) neoplasms using a deep learning (DL) algorithm, and to identify differences in its diagnostic performance in three different sites: larynx/hypopharynx, oral cavity and oropharynx. Methods A total of 1034 endoscopic images from 323 patients were examined under narrow band imaging (NBI). The Mask R-CNN algorithm was used for the analysis. The dataset split was: 935 training, 48 validation and 51 testing images. Dice Similarity Coefficient (Dsc) was the main outcome measure. Results Instance segmentation was effective in 76.5% of images. The mean Dsc was 0.90 ± 0.05. The algorithm correctly predicted 77.8%, 86.7% and 55.5% of lesions in the larynx/hypopharynx, oral cavity, and oropharynx, respectively. The mean Dsc was 0.90 ± 0.05 for the larynx/hypopharynx, 0.60 ± 0.26 for the oral cavity, and 0.81 ± 0.30 for the oropharynx. The analysis showed inferior diagnostic results in the oral cavity compared with the larynx/hypopharynx (p < 0.001). Conclusions The study confirms the feasibility of instance segmentation of UADT using DL algorithms and shows inferior diagnostic results in the oral cavity compared with other anatomic areas.
Collapse
Affiliation(s)
- Alberto Paderno
- Unit of Otorhinolaryngology, Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | | | - Milena Fior
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Giulia Berretti
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Francesca Gennarini
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Gabriele Zigliani
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Emanuela Ulaj
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Claudia Montenegro
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Alessandra Sordi
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| | - Claudio Sampieri
- Unit of Otorhinolaryngology, Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
| | - Giorgio Peretti
- Unit of Otorhinolaryngology, Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
| | - Sara Moccia
- The BioRobotics Institute, Scuola Superiore Sant'Anna, Pisa, Italy
- Department of Excellence in Robotics and AI, Scuola Superiore Sant'Anna, Pisa, Italy
| | - Cesare Piazza
- Unit of Otorhinolaryngology, Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, School of Medicine, Brescia, Italy
| |
Collapse
|
34
|
Ma T, Wu Q, Jiang L, Zeng X, Wang Y, Yuan Y, Wang B, Zhang T. Artificial Intelligence and Machine (Deep) Learning in Otorhinolaryngology: A Bibliometric Analysis Based on VOSviewer and CiteSpace. EAR, NOSE & THROAT JOURNAL 2023:1455613231185074. [PMID: 37515527 DOI: 10.1177/01455613231185074] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/31/2023] Open
Abstract
BACKGROUND Otorhinolaryngology diseases are well suited for artificial intelligence (AI)-based interpretation. The use of AI, particularly AI based on deep learning (DL), in the treatment of human diseases is becoming more and more popular. However, there are few bibliometric analyses that have systematically studied this field. OBJECTIVE The objective of this study was to visualize the research hot spots and trends of AI and DL in ENT diseases through bibliometric analysis to help researchers understand the future development of basic and clinical research. METHODS In all, 232 articles and reviews were retrieved from The Web of Science Core Collection. Using CiteSpace and VOSviewer software, countries, institutions, authors, references, and keywords in the field were visualized and examined. RESULTS The majority of these papers came from 44 nations and 498 institutions, with China and the United States leading the way. Common diseases used by AI in ENT include otosclerosis, otitis media, nasal polyps, sinusitis, and so on. In the early years, research focused on the analysis of hearing and articulation disorders, and in recent years mainly on the diagnosis, localization, and grading of diseases. CONCLUSIONS The analysis shows the periodical hot spots and development direction of AI and DL application in ENT diseases from the time dimension. The diagnosis and prognosis of otolaryngology diseases and the analysis of otolaryngology endoscopic images have been the focus of current research and the development trend of future.
Collapse
Affiliation(s)
- Tianyu Ma
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Qilong Wu
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Li Jiang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Xiaoyun Zeng
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Yuyao Wang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Yi Yuan
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Bingxuan Wang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| | - Tianhong Zhang
- Department of Otorhinolaryngology Head and Neck Surgery, The First Affiliated Hospital of Harbin Medical University, Harbin, China
| |
Collapse
|
35
|
Kaphle A, Jayarathna S, Moktan H, Aliru M, Raghuram S, Krishnan S, Cho SH. Deep Learning-Based TEM Image Analysis for Fully Automated Detection of Gold Nanoparticles Internalized Within Tumor Cell. MICROSCOPY AND MICROANALYSIS : THE OFFICIAL JOURNAL OF MICROSCOPY SOCIETY OF AMERICA, MICROBEAM ANALYSIS SOCIETY, MICROSCOPICAL SOCIETY OF CANADA 2023; 29:1474-1487. [PMID: 37488822 PMCID: PMC10433944 DOI: 10.1093/micmic/ozad066] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/07/2023] [Revised: 04/28/2023] [Accepted: 05/22/2023] [Indexed: 07/26/2023]
Abstract
Transmission electron microscopy (TEM) imaging can be used for detection/localization of gold nanoparticles (GNPs) within tumor cells. However, quantitative analysis of GNP-containing cellular TEM images typically relies on conventional/thresholding-based methods, which are manual, time-consuming, and prone to human errors. In this study, therefore, deep learning (DL)-based methods were developed for fully automated detection of GNPs from cellular TEM images. Several models of "you only look once (YOLO)" v5 were implemented, with a few adjustments to enhance the model's performance by applying the transfer learning approach, adjusting the size of the input image, and choosing the best optimization algorithm. Seventy-eight original (12,040 augmented) TEM images of GNP-laden tumor cells were used for model implementation and validation. A maximum F1 score (harmonic mean of the precision and recall) of 0.982 was achieved by the best-trained models, while mean average precision was 0.989 and 0.843 at 0.50 and 0.50-0.95 intersection over union threshold, respectively. These results suggested the developed DL-based approach was capable of precisely estimating the number/position of internalized GNPs from cellular TEM images. A novel DL-based TEM image analysis tool from this study will benefit research/development efforts on GNP-based cancer therapeutics, for example, by enabling the modeling of GNP-laden tumor cells using nanometer-resolution TEM images.
Collapse
Affiliation(s)
- Amrit Kaphle
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Sandun Jayarathna
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Hem Moktan
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Maureen Aliru
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Subhiksha Raghuram
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| | - Sunil Krishnan
- Vivian L. Smith Department of Neurosurgery, The University of Texas Health Science Center, Houston, TX 77030, USA
| | - Sang Hyun Cho
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| |
Collapse
|
36
|
Zhang T, Bur AM, Kraft S, Kavookjian H, Renslo B, Chen X, Luo B, Wang G. Gender, Smoking History, and Age Prediction from Laryngeal Images. J Imaging 2023; 9:109. [PMID: 37367457 DOI: 10.3390/jimaging9060109] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2023] [Revised: 05/22/2023] [Accepted: 05/25/2023] [Indexed: 06/28/2023] Open
Abstract
Flexible laryngoscopy is commonly performed by otolaryngologists to detect laryngeal diseases and to recognize potentially malignant lesions. Recently, researchers have introduced machine learning techniques to facilitate automated diagnosis using laryngeal images and achieved promising results. The diagnostic performance can be improved when patients' demographic information is incorporated into models. However, the manual entry of patient data is time-consuming for clinicians. In this study, we made the first endeavor to employ deep learning models to predict patient demographic information to improve the detector model's performance. The overall accuracy for gender, smoking history, and age was 85.5%, 65.2%, and 75.9%, respectively. We also created a new laryngoscopic image set for the machine learning study and benchmarked the performance of eight classical deep learning models based on CNNs and Transformers. The results can be integrated into current learning models to improve their performance by incorporating the patient's demographic information.
Collapse
Affiliation(s)
- Tianxiao Zhang
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Andrés M Bur
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Shannon Kraft
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Hannah Kavookjian
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Bryan Renslo
- Department of Otolaryngology-Head and Neck Surgery, University of Kansas Medical Center, Kansas City, KS 66160, USA
| | - Xiangyu Chen
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Bo Luo
- Department of Electrical Engineering and Computer Science, University of Kansas, Lawrence, KS 66045, USA
| | - Guanghui Wang
- Department of Computer Science, Toronto Metropolitan University, Toronto, ON M5B 2K3, Canada
| |
Collapse
|
37
|
Ghourabi M, Mourad-Chehade F, Chkeir A. Eye Recognition by YOLO for Inner Canthus Temperature Detection in the Elderly Using a Transfer Learning Approach. SENSORS (BASEL, SWITZERLAND) 2023; 23:1851. [PMID: 36850447 PMCID: PMC9964838 DOI: 10.3390/s23041851] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/08/2022] [Revised: 01/23/2023] [Accepted: 02/02/2023] [Indexed: 06/18/2023]
Abstract
Early detection of physical frailty and infectious diseases in seniors is important to avoid any fatal drawback and promptly provide them with the necessary healthcare. One of the major symptoms of viral infections is elevated body temperature. In this work, preparation and implementation of multi-age thermal faces dataset is done to train different "You Only Look Once" (YOLO) object detection models (YOLOv5,6 and 7) for eye detection. Eye detection allows scanning for the most accurate temperature in the face, which is the inner canthus temperature. An approach using an elderly thermal dataset is performed in order to produce an eye detection model specifically for elderly people. An application of transfer learning is applied from a multi-age YOLOv7 model to an elderly YOLOv7 model. The comparison of speed, accuracy, and size between the trained models shows that the YOLOv7 model performed the best (Mean average precision at Intersection over Union of 0.5 (mAP@.5) = 0.996 and Frames per Seconds (FPS) = 150). The bounding box of eyes is scanned for the highest temperature, resulting in a normalized error distance of 0.03. This work presents a fast and reliable temperature detection model generated using non-contact infrared camera and a deep learning approach.
Collapse
|
38
|
Yan P, Li S, Zhou Z, Liu Q, Wu J, Ren Q, Chen Q, Chen Z, Chen Z, Chen S, Scholp A, Jiang JJ, Kang J, Ge P. Automated detection of glottic laryngeal carcinoma in laryngoscopic images from a multicentre database using a convolutional neural network. Clin Otolaryngol 2023; 48:436-441. [PMID: 36624555 DOI: 10.1111/coa.14029] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2021] [Revised: 11/22/2022] [Accepted: 12/31/2022] [Indexed: 01/11/2023]
Abstract
OBJECTIVE Little is known about the efficacy of using artificial intelligence (AI) to identify laryngeal carcinoma from images of vocal lesions taken in different hospitals with multiple laryngoscope systems. This multicentre study aimed to establish an AI system and provide a reliable auxiliary tool to screen for laryngeal carcinoma. STUDY DESIGN Multicentre case-control study. SETTING Six tertiary care centres. PARTICIPANTS Laryngoscopy images were collected from 2179 patients with vocal fold lesions. OUTCOME MEASURES An automatic detection system of laryngeal carcinoma was established and used to distinguish malignant and benign vocal lesions in 2179 laryngoscopy images acquired from 6 hospitals with 5 types of laryngoscopy systems. Pathological examination was the gold standard for identifying malignant and benign vocal lesions. RESULTS Out of 89 cases in the malignant group, the classifier was able to correctly identify laryngeal carcinoma in 66 patients (74.16%, sensitivity). Out of 640 cases in the benign group, the classifier was able to accurately assess the laryngeal lesion in 503 cases (78.59%, specificity). Furthermore, the region-based convolutional neural network (R-CNN) classifier achieved an overall accuracy of 78.05%, with a 95.63% negative predictive value and a 32.51% positive predictive value for the testing data set. CONCLUSION This automatic diagnostic system has the potential to assist clinical laryngeal carcinoma diagnosis which may improve and standardise the diagnostic capacity of laryngologists using different laryngoscopes.
Collapse
Affiliation(s)
- Peikai Yan
- Department of Otolaryngology & Head Neck Surgery, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, China.,School of Medicine, South China University of Technology, Guangzhou, China
| | - Shaohua Li
- Department of Otorhinolaryngology Head and Neck Surgery, Zhongshan Hospital of Traditional Chinese Medicine, Affiliated to Guangzhou University of Chinese Medicine, Guangdong, Zhongshan, Guangdong, China
| | - Zhou Zhou
- Department of Otolaryngology, Shenzhen People's Hospital (The Second Clinical Medical College, Jinan University; The First Affiliated Hospital, Southern University of Science and Technology), Shenzhen, China
| | - Qian Liu
- Department of Otolaryngology, Shenzhen People's Hospital (The Second Clinical Medical College, Jinan University; The First Affiliated Hospital, Southern University of Science and Technology), Shenzhen, China
| | - Jiahui Wu
- Department of Otolaryngology & Head Neck Surgery, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, China
| | - Qingyi Ren
- Department of Otolaryngology & Head Neck Surgery, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, China
| | - Qiuhuan Chen
- Department of Otolaryngology, Zhaoqing Gaoyao People's Hospital, Zhaoqing, China
| | - Zhipeng Chen
- Department of Otolaryngology, The Second People's Hospital of Longgang District, Shenzhen, China
| | - Ze Chen
- Department of Otolaryngology, Gaozhou People's Hospital, Gaozhou, China
| | - Shaohua Chen
- Department of Otolaryngology & Head Neck Surgery, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, China
| | - Austin Scholp
- Department of Biomedical Engineering, University of Iowa, Iowa City, Iowa, USA.,Division of Otolaryngology-Head and Neck Surgery, Department of Surgery, School of Medicine and Public Health (A.S.), University of Wisconsin-Madison, Madison, Wisconsin, USA
| | - Jack J Jiang
- Division of Otolaryngology-Head and Neck Surgery, Department of Surgery, School of Medicine and Public Health (A.S.), University of Wisconsin-Madison, Madison, Wisconsin, USA
| | - Jing Kang
- Department of Otolaryngology & Head Neck Surgery, Guangdong Provincial People's Hospital (Guangdong Academy of Medical Sciences), Southern Medical University, Guangzhou, China.,School of Medicine, South China University of Technology, Guangzhou, China
| | - Pingjiang Ge
- School of Medicine, South China University of Technology, Guangzhou, China
| |
Collapse
|
39
|
Choi SJ, Kim DK, Kim BS, Cho M, Jeong J, Jo YH, Song KJ, Kim YJ, Kim S. Mask R-CNN based multiclass segmentation model for endotracheal intubation using video laryngoscope. Digit Health 2023; 9:20552076231211547. [PMID: 38025115 PMCID: PMC10631336 DOI: 10.1177/20552076231211547] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2023] [Accepted: 10/17/2023] [Indexed: 12/01/2023] Open
Abstract
Objective Endotracheal intubation (ETI) is critical to secure the airway in emergent situations. Although artificial intelligence algorithms are frequently used to analyze medical images, their application to evaluating intraoral structures based on images captured during emergent ETI remains limited. The aim of this study is to develop an artificial intelligence model for segmenting structures in the oral cavity using video laryngoscope (VL) images. Methods From 54 VL videos, clinicians manually labeled images that include motion blur, foggy vision, blood, mucus, and vomitus. Anatomical structures of interest included the tongue, epiglottis, vocal cord, and corniculate cartilage. EfficientNet-B5 with DeepLabv3+, EffecientNet-B5 with U-Net, and Configured Mask R-Convolution Neural Network (CNN) were used; EffecientNet-B5 was pretrained on ImageNet. Dice similarity coefficient (DSC) was used to measure the segmentation performance of the model. Accuracy, recall, specificity, and F1 score were used to evaluate the model's performance in targeting the structure from the value of the intersection over union between the ground truth and prediction mask. Results The DSC of tongue, epiglottis, vocal cord, and corniculate cartilage obtained from the EfficientNet-B5 with DeepLabv3+, EfficientNet-B5 with U-Net, and Configured Mask R-CNN model were 0.3351/0.7675/0.766/0.6539, 0.0/0.7581/0.7395/0.6906, and 0.1167/0.7677/0.7207/0.57, respectively. Furthermore, the processing speeds (frames per second) of the three models stood at 3, 24, and 32, respectively. Conclusions The algorithm developed in this study can assist medical providers performing ETI in emergent situations.
Collapse
Affiliation(s)
- Seung Jae Choi
- Transdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Seoul, Republic of Korea
| | - Dae Kon Kim
- Department of Emergency Medicine, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Emergency Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
- Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Byeong Soo Kim
- Interdisciplinary Program in Bioengineering, Graduate School, Seoul National University, Seoul, Republic of Korea
| | - Minwoo Cho
- Transdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Seoul, Republic of Korea
| | - Joo Jeong
- Department of Emergency Medicine, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Emergency Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - You Hwan Jo
- Department of Emergency Medicine, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Emergency Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Kyoung Jun Song
- Department of Emergency Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
- Department of Emergency Medicine, Seoul Metropolitan Government-Seoul National University Boramae Medical Center, Seoul, Republic of Korea
| | - Yu Jin Kim
- Department of Emergency Medicine, Seoul National University Bundang Hospital, Seongnam, Republic of Korea
- Department of Emergency Medicine, Seoul National University College of Medicine, Seoul, Republic of Korea
| | - Sungwan Kim
- Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, Republic of Korea
- Institute of Bioengineering, Seoul National University, Seoul, Republic of Korea
| |
Collapse
|
40
|
Bassani S, Lee YK, Campagnari V, Eccher A, Monzani D, Nocini R, Sacchetto L, Molteni G. From Hype To Reality: A Narrative Review on the Promising Role of Artificial Intelligence in Larynx Cancer Detection and Transoral Microsurgery. Crit Rev Oncog 2023; 28:21-24. [PMID: 37968989 DOI: 10.1615/critrevoncog.2023049134] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2023]
Abstract
Early larynx cancer detection plays a crucial role in improving treatment outcomes and recent studies have shown promising results in using artificial intelligence for larynx cancer detection. Artificial intelligence also has the potential to enhance transoral larynx microsurgery. This narrative review summarizes the current evidence regarding its use in larynx cancer detection and potential applications in transoral larynx microsurgery. The utilization of artificial intelligence in larynx cancer detection with white light endoscopy and narrow-band imaging helps improve diagnostic accuracy and efficiency. It can also potentially enhance transoral larynx microsurgery by aiding surgeons in real-time decision-making and minimizing the risk of complications. However, further prospective studies are warranted to validate the findings, and additional research is necessary to optimize the integration of artificial intelligence in our clinical practice.
Collapse
Affiliation(s)
- Sara Bassani
- Otolaryngology-Head and Neck Surgery Department, University of Verona, Verona, Italy
| | - Ying Ki Lee
- Head and Neck Surgery Department, Guy's and St. Thomas' NHS Foundation Trust, London, United Kingdom
| | - Valentina Campagnari
- Unit of Otorhinolaryngology- Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy
| | - Albino Eccher
- Department of Pathology and Diagnostics, University and Hospital Trust of Verona, Verona, Italy
| | - Daniele Monzani
- Otolaryngology-Head and Neck Surgery Department, University of Verona, Verona, Italy
| | - Riccardo Nocini
- Otolaryngology-Head and Neck Surgery Department, University of Verona, Verona, Italy
| | - Luca Sacchetto
- Otolaryngology-Head and Neck Surgery Department, University of Verona, Verona, Italy
| | - Gabriele Molteni
- Department of Otorhinolaryngology, Head and Neck Surgery, IRCCS Azienda Ospedaliero-Universitaria di Bologna, Policlinico S. Orsola-Malpighi, Bologna, Italy
| |
Collapse
|
41
|
Gu Y, Li Y, Ge S, Lu W, Mao Y, Chen M, Qian Y. A SERS Biosensor Based on Functionalized Au-SiNCA Integrated with a Dual Signal Amplification Strategy for Sensitive Detection of Telomerase Activity During EMT in Laryngeal Carcinoma. Int J Nanomedicine 2023; 18:2553-2565. [PMID: 37213349 PMCID: PMC10198182 DOI: 10.2147/ijn.s409864] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Accepted: 05/09/2023] [Indexed: 05/23/2023] Open
Abstract
Purpose This paper aims to construct a surface-enhanced Raman spectroscopy (SERS) biosensor based on functionalized Au-Si nanocone arrays (Au-SiNCA) using a dual signal amplification strategy (SDA-CHA) to evaluate telomerase activity during epithelial-mesenchymal transition (EMT) in laryngeal carcinoma (LC). Methods A SERS biosensor based on functionalized Au-SiNCA was designed with an integrated dual-signal amplification strategy to achieve ultrasensitive detection of telomerase activity during EMT in LC patients. Results Labeled probes (Au-AgNRs@4-MBA@H1) and capture substrates (Au-SiNCA@H2) were prepared by modifying hairpin DNA and Raman signal molecules. Using this scheme, telomerase activity in peripheral mononuclear cells (PMNC) could be successfully detected with a limit of detection (LOD) as low as 10-6 IU/mL. In addition, biological experiments using BLM treatment of TU686 effectively mimicked the EMT process. The results of this scheme were highly consistent with the ELISA scheme, confirming its accuracy. Conclusion This scheme provides a reproducible, selective, and ultrasensitive assay for telomerase activity, which is expected to be a potential tool for the early screening of LC in future clinical applications.
Collapse
Affiliation(s)
- Yuexing Gu
- Institute of Translational Medicine, Medical College, Yangzhou University, Yangzhou, People’s Republic of China
- Jiangsu Key Laboratory of Integrated Traditional Chinese and Western Medicine for Prevention and Treatment of Senile Diseases, Yangzhou University, Yangzhou, People’s Republic of China
| | - Yan Li
- Department of Obstetrics and Gynecology, The Second People’s Hospital of Taizhou City, Taizhou, People’s Republic of China
| | - Shengjie Ge
- Institute of Translational Medicine, Medical College, Yangzhou University, Yangzhou, People’s Republic of China
- Jiangsu Key Laboratory of Integrated Traditional Chinese and Western Medicine for Prevention and Treatment of Senile Diseases, Yangzhou University, Yangzhou, People’s Republic of China
| | - Wenbo Lu
- Shanxi Normal University, College of Chemistry and Material Science, Linfen, People’s Republic of China
| | - Yu Mao
- Institute of Translational Medicine, Medical College, Yangzhou University, Yangzhou, People’s Republic of China
- Jiangsu Key Laboratory of Integrated Traditional Chinese and Western Medicine for Prevention and Treatment of Senile Diseases, Yangzhou University, Yangzhou, People’s Republic of China
| | - Miao Chen
- Institute of Translational Medicine, Medical College, Yangzhou University, Yangzhou, People’s Republic of China
- Jiangsu Key Laboratory of Integrated Traditional Chinese and Western Medicine for Prevention and Treatment of Senile Diseases, Yangzhou University, Yangzhou, People’s Republic of China
| | - Yayun Qian
- Institute of Translational Medicine, Medical College, Yangzhou University, Yangzhou, People’s Republic of China
- Jiangsu Key Laboratory of Integrated Traditional Chinese and Western Medicine for Prevention and Treatment of Senile Diseases, Yangzhou University, Yangzhou, People’s Republic of China
- Correspondence: Yayun Qian, Email
| |
Collapse
|
42
|
Sahoo PK, Mishra S, Panigrahi R, Bhoi AK, Barsocchi P. An Improvised Deep-Learning-Based Mask R-CNN Model for Laryngeal Cancer Detection Using CT Images. SENSORS (BASEL, SWITZERLAND) 2022; 22:8834. [PMID: 36433430 PMCID: PMC9697116 DOI: 10.3390/s22228834] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/05/2022] [Revised: 11/08/2022] [Accepted: 11/10/2022] [Indexed: 05/30/2023]
Abstract
Recently, laryngeal cancer cases have increased drastically across the globe. Accurate treatment for laryngeal cancer is intricate, especially in the later stages. This type of cancer is an intricate malignancy inside the head and neck area of patients. In recent years, diverse diagnosis approaches and tools have been developed by researchers for helping clinical experts to identify laryngeal cancer effectively. However, these existing tools and approaches have diverse issues related to performance constraints such as lower accuracy in the identification of laryngeal cancer in the initial stage, more computational complexity, and large time consumption in patient screening. In this paper, the authors present a novel and enhanced deep-learning-based Mask R-CNN model for the identification of laryngeal cancer and its related symptoms by utilizing diverse image datasets and CT images in real time. Furthermore, our suggested model is capable of capturing and detecting minor malignancies of the larynx portion in a significant and faster manner in the real-time screening of patients, and it saves time for the clinicians, allowing for more patient screening every day. The outcome of the suggested model is enhanced and pragmatic and obtained an accuracy of 98.99%, precision of 98.99%, F1 score of 97.99%, and recall of 96.79% on the ImageNet dataset. Several studies have been performed in recent years on laryngeal cancer detection by using diverse approaches from researchers. For the future, there are vigorous opportunities for further research to investigate new approaches for laryngeal cancer detection by utilizing diverse and large dataset images.
Collapse
Affiliation(s)
- Pravat Kumar Sahoo
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar 751024, India
| | - Sushruta Mishra
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar 751024, India
| | - Ranjit Panigrahi
- Department of Computer Applications, Sikkim Manipal Institute of Technology, Sikkim Manipal University, Majitar, Rangpo 737136, India
| | - Akash Kumar Bhoi
- KIET Group of Institutions, Delhi-NCR, Ghaziabad 201206, India
- Directorate of Research, Sikkim Manipal University, Gangtok 737102, India
- Institute of Information Science and Technologies, National Research Council, 56124 Pisa, Italy
| | - Paolo Barsocchi
- Institute of Information Science and Technologies, National Research Council, 56124 Pisa, Italy
| |
Collapse
|
43
|
Comparison of convolutional neural networks for classification of vocal fold nodules from high-speed video images. Eur Arch Otorhinolaryngol 2022; 280:2365-2371. [PMID: 36357609 DOI: 10.1007/s00405-022-07736-6] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2022] [Accepted: 10/29/2022] [Indexed: 11/12/2022]
Abstract
OBJECTIVES Deep learning is in this study used through convolutional neural networks (CNN) to the determination of vocal fold nodules. Through high-speed video (HSV) images and computer-assisted tools, a comparison of convolutional neural network models and their accuracy will be presented. METHODS The data have been collected by an Ear Nose Throat (ENT) specialist with a 90° rigid scope in the years from 2007 to 2019, where 15.732 high-speed videos have been collected from 7909 patients. A total of 4000 images have been carefully selected, 2000 images were of normal vocal folds and 2000 images were of vocal folds with varying degrees of vocal fold nodules. These images were then split into training-, validation-, and testing-data set, for use with a CNN model with 5 layers (CNN5) and compared to other models: VGG19, MobileNetV2, and Inception-ResNetV2. To compare the neural network models, the following evaluation metrics have been calculated: accuracy, sensitivity, specificity, precision, and negative predictive values. RESULTS All the trained CNN models have shown high accuracy when applied to the test set. The accuracy is 97.75%, 83.5%, 91.5%, and 89.75%, for CNN5, VGG19, MobileNetV2, and InceptionResNetV2, respectively. CONCLUSIONS Precision was identified as the most relevant performance metric for a study that focuses on the classification of vocal fold nodules. The highest performing model was MobilNetV2 with a precision of 97.7%. The average accuracy across all 4 neural networks was 90.63% showing that neural networks can be used for classifying vocal fold nodules in a clinical setting.
Collapse
|
44
|
Deng Y, Chen Y, Xie L, Wang L, Zhan J. The investigation of construction and clinical application of image recognition technology assisted bronchoscopy diagnostic model of lung cancer. Front Oncol 2022; 12:1001840. [PMID: 36387178 PMCID: PMC9647035 DOI: 10.3389/fonc.2022.1001840] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2022] [Accepted: 10/07/2022] [Indexed: 12/02/2022] Open
Abstract
Background The incidence and mortality of lung cancer ranks first in China. Bronchoscopy is one of the most common diagnostic methods for lung cancer. In recent years, image recognition technology(IRT) has been more and more widely studied and applied in the medical field. We developed a diagnostic model of lung cancer under bronchoscopy based on deep learning method and tried to classify pathological types. Methods A total of 2238 lesion images were collected retrospectively from 666 cases of lung cancer diagnosed by pathology in the bronchoscopy center of the Third Xiangya Hospital from Oct.01 2017 to Dec.31 2020 and 152 benign cases from Jun.01 2015 to Dec.31 2020. The benign and malignant images were divided into training, verification and test set according to 7:1:2 respectively. The model was trained and tested based on deep learning method. We also tried to classify different pathological types of lung cancer using the model. Furthermore, 9 clinicians with different experience were invited to diagnose the same test images and the results were compared with the model. Results The diagnostic model took a total of 30s to diagnose 467 test images. The overall accuracy, sensitivity, specificity and area under curve (AUC) of the model to differentiate benign and malignant lesions were 0.951, 0.978, 0.833 and 0.940, which were equivalent to the judgment results of 2 doctors in the senior group and higher than those of other doctors. In the classification of squamous cell carcinoma (SCC) and adenocarcinoma (AC), the overall accuracy was 0.745, including 0.790 for SCC, 0.667 for AC and AUC was 0.728. Conclusion The performance of our diagnostic model to distinguish benign and malignant lesions in bronchoscopy is roughly the same as that of experienced clinicians and the efficiency is much higher than manually. Our study verifies the possibility of applying IRT in diagnosis of lung cancer during white light bronchoscopy.
Collapse
Affiliation(s)
- Yihong Deng
- Department of Pulmonary and Critical Care Medicine, the Third Xiangya Hospital, Central South University, Changsha, Hunan, China
| | - Yuan Chen
- Department of Computer Science, School of Informatics, Xiamen University, Xiamen, Fujian, China
| | - Lihua Xie
- Department of Pulmonary and Critical Care Medicine, the Third Xiangya Hospital, Central South University, Changsha, Hunan, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| | - Liansheng Wang
- Department of Computer Science, School of Informatics, Xiamen University, Xiamen, Fujian, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| | - Juan Zhan
- Department of Oncology, Zhongshan Hospital affiliated to Xiamen University, Xiamen, Fujian, China
- *Correspondence: Lihua Xie, ; Liansheng Wang, ; Juan Zhan,
| |
Collapse
|
45
|
Paderno A, Gennarini F, Sordi A, Montenegro C, Lancini D, Villani FP, Moccia S, Piazza C. Artificial intelligence in clinical endoscopy: Insights in the field of videomics. Front Surg 2022; 9:933297. [PMID: 36171813 PMCID: PMC9510389 DOI: 10.3389/fsurg.2022.933297] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/30/2022] [Accepted: 08/22/2022] [Indexed: 11/13/2022] Open
Abstract
Artificial intelligence is being increasingly seen as a useful tool in medicine. Specifically, these technologies have the objective to extract insights from complex datasets that cannot easily be analyzed by conventional statistical methods. While promising results have been obtained for various -omics datasets, radiological images, and histopathologic slides, analysis of videoendoscopic frames still represents a major challenge. In this context, videomics represents a burgeoning field wherein several methods of computer vision are systematically used to organize unstructured data from frames obtained during diagnostic videoendoscopy. Recent studies have focused on five broad tasks with increasing complexity: quality assessment of endoscopic images, classification of pathologic and nonpathologic frames, detection of lesions inside frames, segmentation of pathologic lesions, and in-depth characterization of neoplastic lesions. Herein, we present a broad overview of the field, with a focus on conceptual key points and future perspectives.
Collapse
Affiliation(s)
- Alberto Paderno
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, School of Medicine, University of Brescia, Brescia, Italy
- Correspondence: Alberto Paderno
| | - Francesca Gennarini
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, School of Medicine, University of Brescia, Brescia, Italy
| | - Alessandra Sordi
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, School of Medicine, University of Brescia, Brescia, Italy
| | - Claudia Montenegro
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, School of Medicine, University of Brescia, Brescia, Italy
| | - Davide Lancini
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
| | - Francesca Pia Villani
- The BioRobotics Institute, Scuola Superiore Sant’Anna, Pisa, Italy
- Department of Excellence in Robotics and AI, Scuola Superiore Sant’Anna, Pisa, Italy
| | - Sara Moccia
- The BioRobotics Institute, Scuola Superiore Sant’Anna, Pisa, Italy
- Department of Excellence in Robotics and AI, Scuola Superiore Sant’Anna, Pisa, Italy
| | - Cesare Piazza
- Unit of Otorhinolaryngology—Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy
- Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, School of Medicine, University of Brescia, Brescia, Italy
| |
Collapse
|
46
|
Azam MA, Sampieri C, Ioppi A, Benzi P, Giordano GG, De Vecchi M, Campagnari V, Li S, Guastini L, Paderno A, Moccia S, Piazza C, Mattos LS, Peretti G. Videomics of the Upper Aero-Digestive Tract Cancer: Deep Learning Applied to White Light and Narrow Band Imaging for Automatic Segmentation of Endoscopic Images. Front Oncol 2022; 12:900451. [PMID: 35719939 PMCID: PMC9198427 DOI: 10.3389/fonc.2022.900451] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2022] [Accepted: 04/26/2022] [Indexed: 12/13/2022] Open
Abstract
Introduction Narrow Band Imaging (NBI) is an endoscopic visualization technique useful for upper aero-digestive tract (UADT) cancer detection and margins evaluation. However, NBI analysis is strongly operator-dependent and requires high expertise, thus limiting its wider implementation. Recently, artificial intelligence (AI) has demonstrated potential for applications in UADT videoendoscopy. Among AI methods, deep learning algorithms, and especially convolutional neural networks (CNNs), are particularly suitable for delineating cancers on videoendoscopy. This study is aimed to develop a CNN for automatic semantic segmentation of UADT cancer on endoscopic images. Materials and Methods A dataset of white light and NBI videoframes of laryngeal squamous cell carcinoma (LSCC) was collected and manually annotated. A novel DL segmentation model (SegMENT) was designed. SegMENT relies on DeepLabV3+ CNN architecture, modified using Xception as a backbone and incorporating ensemble features from other CNNs. The performance of SegMENT was compared to state-of-the-art CNNs (UNet, ResUNet, and DeepLabv3). SegMENT was then validated on two external datasets of NBI images of oropharyngeal (OPSCC) and oral cavity SCC (OSCC) obtained from a previously published study. The impact of in-domain transfer learning through an ensemble technique was evaluated on the external datasets. Results 219 LSCC patients were retrospectively included in the study. A total of 683 videoframes composed the LSCC dataset, while the external validation cohorts of OPSCC and OCSCC contained 116 and 102 images. On the LSCC dataset, SegMENT outperformed the other DL models, obtaining the following median values: 0.68 intersection over union (IoU), 0.81 dice similarity coefficient (DSC), 0.95 recall, 0.78 precision, 0.97 accuracy. For the OCSCC and OPSCC datasets, results were superior compared to previously published data: the median performance metrics were, respectively, improved as follows: DSC=10.3% and 11.9%, recall=15.0% and 5.1%, precision=17.0% and 14.7%, accuracy=4.1% and 10.3%. Conclusion SegMENT achieved promising performances, showing that automatic tumor segmentation in endoscopic images is feasible even within the highly heterogeneous and complex UADT environment. SegMENT outperformed the previously published results on the external validation cohorts. The model demonstrated potential for improved detection of early tumors, more precise biopsies, and better selection of resection margins.
Collapse
Affiliation(s)
- Muhammad Adeel Azam
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Claudio Sampieri
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Alessandro Ioppi
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Pietro Benzi
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Giorgio Gregory Giordano
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Marta De Vecchi
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Valentina Campagnari
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Shunlei Li
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Luca Guastini
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| | - Alberto Paderno
- Unit of Otorhinolaryngology - Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy.,Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, Brescia, Italy
| | - Sara Moccia
- The BioRobotics Institute and Department of Excellence in Robotics and AI, Scuola Superiore Sant'Anna, Pisa, Italy
| | - Cesare Piazza
- Unit of Otorhinolaryngology - Head and Neck Surgery, ASST Spedali Civili of Brescia, Brescia, Italy.,Department of Medical and Surgical Specialties, Radiological Sciences, and Public Health, University of Brescia, Brescia, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| | - Giorgio Peretti
- Unit of Otorhinolaryngology - Head and Neck Surgery, IRCCS Ospedale Policlinico San Martino, Genoa, Italy.,Department of Surgical Sciences and Integrated Diagnostics (DISC), University of Genoa, Genoa, Italy
| |
Collapse
|
47
|
Smoke removal and image enhancement of laparoscopic images by an artificial multi-exposure image fusion method. Soft comput 2022. [DOI: 10.1007/s00500-022-06990-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
48
|
Xue J, Cheng F, Li Y, Song Y, Mao T. Detection of Farmland Obstacles Based on an Improved YOLOv5s Algorithm by Using CIoU and Anchor Box Scale Clustering. SENSORS 2022; 22:s22051790. [PMID: 35270935 PMCID: PMC8915062 DOI: 10.3390/s22051790] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/29/2022] [Revised: 02/18/2022] [Accepted: 02/21/2022] [Indexed: 02/05/2023]
Abstract
It is necessary to detect multi-type farmland obstacles in real time and accurately for unmanned agricultural vehicles. An improved YOLOv5s algorithm based on the K-Means clustering algorithm and CIoU Loss function was proposed to improve detection precision and speed up real-time detection. The K-Means clustering algorithm was used in order to generate anchor box scales to accelerate the convergence speed of model training. The CIoU Loss function, combining the three geometric measures of overlap area, center distance and aspect ratio, was adopted to reduce the occurrence of missed and false detection and improve detection precision. The experimental results showed that the inference time of a single image was reduced by 75% with the improved YOLOv5s algorithm; compared with that of the Faster R-CNN algorithm, real-time performance was effectively improved. Furthermore, the mAP value of the improved algorithm was increased by 5.80% compared with that of the original YOLOv5s, which indicates that using the CIoU Loss function had an obvious effect on reducing the missed detection and false detection of the original YOLOv5s. Moreover, the detection of small target obstacles of the improved algorithm was better than that of the Faster R-CNN.
Collapse
|