1
|
Fang Y, Wu Y, Gao L. Machine learning-based myocardial infarction bibliometric analysis. Front Med (Lausanne) 2025; 12:1477351. [PMID: 39981082 PMCID: PMC11839716 DOI: 10.3389/fmed.2025.1477351] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2024] [Accepted: 01/17/2025] [Indexed: 02/22/2025] Open
Abstract
Purpose This study analyzed the research trends in machine learning (ML) pertaining to myocardial infarction (MI) from 2008 to 2024, aiming to identify emerging trends and hotspots in the field, providing insights into the future directions of research and development in ML for MI. Additionally, it compared the contributions of various countries, authors, and agencies to the field of ML research focused on MI. Method A total of 1,036 publications were collected from the Web of Science Core Collection database. CiteSpace 6.3.R1, Bibliometrix, and VOSviewer were utilized to analyze bibliometric characteristics, determining the number of publications, countries, institutions, authors, keywords, and cited authors, documents, and journals in popular scientific fields. CiteSpace was used for temporal trend analysis, Bibliometrix for quantitative country and institutional analysis, and VOSviewer for visualization of collaboration networks. Results Since the emergence of research literature on medical imaging and machine learning (ML) in 2008, interest in this field has grown rapidly, particularly since the pivotal moment in 2016. The ML and MI domains, represented by China and the United States, have experienced swift development in research after 2015, albeit with the United States significantly outperforming China in research quality (as evidenced by the higher impact factors of journals and citation counts of publications from the United States). Institutional collaborations have formed, notably between Harvard Medical School in the United States and Capital Medical University in China, highlighting the need for enhanced cooperation among domestic and international institutions. In the realm of MI and ML research, cooperative teams led by figures such as Dey, Damini, and Berman, Daniel S. in the United States have emerged, indicating that Chinese scholars should strengthen their collaborations and focus on both qualitative and quantitative development. The overall direction of MI and ML research trends toward Medicine, Medical Sciences, Molecular Biology, and Genetics. In particular, publications in "Circulation" and "Computers in Biology and Medicine" from the United States hold prominent positions in this study. Conclusion This paper presents a comprehensive exploration of the research hotspots, trends, and future directions in the field of MI and ML over the past two decades. The analysis reveals that deep learning is an emerging research direction in MI, with neural networks playing a crucial role in early diagnosis, risk assessment, and rehabilitation therapy.
Collapse
Affiliation(s)
- Ying Fang
- Xiaoshan District Hospital of Traditional Chinese Medicine, Hangzhou, Zhejiang Province, China
| | | | | |
Collapse
|
2
|
Li H, Yuan Q, Wang Y, Qu P, Jiang C, Kuang H. An algorithm for cardiac disease detection based on the magnetic resonance imaging. Sci Rep 2025; 15:4053. [PMID: 39901039 PMCID: PMC11790828 DOI: 10.1038/s41598-025-88567-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2024] [Accepted: 01/29/2025] [Indexed: 02/05/2025] Open
Abstract
In experiments to detect heart disease on cardiac magnetic resonance imaging (MRI) medical images, existing object detection models face several challenges including low accuracy and unreliable detection results. To tackle these issues, this article proposes an innovative method for Object Detection in cardiac MRI medical images called SA-YOLO. This method is based on the YOLOv8 model but introduces several key modifications. Firstly, the standard Spatial Pyramid Pooling Fast module is replaced with a Multi-Channel Spatial Pyramid Pooling module. Secondly, an attention mechanism combining the ideas of Squeeze-Excitation and Coordinate Attention designed, and integrated into the Neck part of the baseline model. Subsequently, the bounding box regression loss function CIoU loss of the model was replaced with the iSD-IoU loss that combines shape loss and distance loss. Finally, comparative experiments were conducted on the Automated Cardiac Diagnosis Challenge cardiac MRI image dataset where it was found that SA-YOLOv8 achieved better results in detecting cardiac pathologies, and improvement of 7.4% in mAP0.5 value and 5.1% in mAP0.5-0.95 value compared to the baseline model.
Collapse
Affiliation(s)
- Heng Li
- Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang, 550025, China
| | - Qingni Yuan
- Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang, 550025, China.
| | - Yi Wang
- The First People's Hospital of Guiyang, Guiyang, 550002, China
| | - Pengju Qu
- Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang, 550025, China
| | - Chunhui Jiang
- Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang, 550025, China
| | - Hu Kuang
- Key Laboratory of Advanced Manufacturing Technology of the Ministry of Education, Guizhou University, Guiyang, 550025, China
| |
Collapse
|
3
|
Arega TW, Bricq S, Meriaudeau F. Post-hoc out-of-distribution detection for cardiac MRI segmentation. Comput Med Imaging Graph 2025; 119:102476. [PMID: 39700904 DOI: 10.1016/j.compmedimag.2024.102476] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2024] [Revised: 10/29/2024] [Accepted: 12/04/2024] [Indexed: 12/21/2024]
Abstract
In real-world scenarios, medical image segmentation models encounter input images that may deviate from the training images in various ways. These differences can arise from changes in image scanners and acquisition protocols, or even the images can come from a different modality or domain. When the model encounters these out-of-distribution (OOD) images, it can behave unpredictably. Therefore, it is important to develop a system that handles such out-of-distribution images to ensure the safe usage of the models in clinical practice. In this paper, we propose a post-hoc out-of-distribution (OOD) detection method that can be used with any pre-trained segmentation model. Our method utilizes multi-scale representations extracted from the encoder blocks of the segmentation model and employs Mahalanobis distance as a metric to measure the similarity between the input image and the in-distribution images. The segmentation model is pre-trained on a publicly available cardiac short-axis cine MRI dataset. The detection performance of the proposed method is evaluated on 13 different OOD datasets, which can be categorized as near, mild, and far OOD datasets based on their similarity to the in-distribution dataset. The results show that our method outperforms state-of-the-art feature space-based and uncertainty-based OOD detection methods across the various OOD datasets. Our method successfully detects near, mild, and far OOD images with high detection accuracy, showcasing the advantage of using the multi-scale and semantically rich representations of the encoder. In addition to the feature-based approach, we also propose a Dice coefficient-based OOD detection method, which demonstrates superior performance for adversarial OOD detection and shows a high correlation with segmentation quality. For the uncertainty-based method, despite having a strong correlation with the quality of the segmentation results in the near OOD datasets, they failed to detect mild and far OOD images, indicating the weakness of these methods when the images are more dissimilar. Future work will explore combining Mahalanobis distance and uncertainty scores for improved detection of challenging OOD images that are difficult to segment.
Collapse
|
4
|
Zheng F, Yin P, Liang K, Liu T, Wang Y, Hao W, Hao Q, Hong N. Comparison of Different Fusion Radiomics for Predicting Benign and Malignant Sacral Tumors: A Pilot Study. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2415-2427. [PMID: 38717515 PMCID: PMC11522258 DOI: 10.1007/s10278-024-01134-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 04/27/2024] [Accepted: 04/29/2024] [Indexed: 10/30/2024]
Abstract
Differentiating between benign and malignant sacral tumors is crucial for determining appropriate treatment options. This study aims to develop two benchmark fusion models and a deep learning radiomic nomogram (DLRN) capable of distinguishing between benign and malignant sacral tumors using multiple imaging modalities. We reviewed axial T2-weighted imaging (T2WI) and non-contrast computed tomography (NCCT) of 134 patients pathologically confirmed as sacral tumors. The two benchmark fusion models were developed using fusion deep learning (DL) features and fusion classical machine learning (CML) features from multiple imaging modalities, employing logistic regression, K-nearest neighbor classification, and extremely randomized trees. The two benchmark models exhibiting the most robust predictive performance were merged with clinical data to formulate the DLRN. Performance assessment involved computing the area under the receiver operating characteristic curve (AUC), sensitivity, specificity, accuracy, negative predictive value (NPV), and positive predictive value (PPV). The DL benchmark fusion model demonstrated superior performance compared to the CML fusion model. The DLRN, identified as the optimal model, exhibited the highest predictive performance, achieving an accuracy of 0.889 and an AUC of 0.961 in the test sets. Calibration curves were utilized to evaluate the predictive capability of the models, and decision curve analysis (DCA) was conducted to assess the clinical net benefit of the DLR model. The DLRN could serve as a practical predictive tool, capable of distinguishing between benign and malignant sacral tumors, offering valuable information for risk counseling, and aiding in clinical treatment decisions.
Collapse
Affiliation(s)
- Fei Zheng
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Ping Yin
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Kewei Liang
- Intelligent Manufacturing Research Institute, Visual 3D Medical Science and Technology Development, Fengtai District, No. 186 South Fourth Ring Road West, Beijing, 100071, People's Republic of China
| | - Tao Liu
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Yujian Wang
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Wenhan Hao
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Qi Hao
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China
| | - Nan Hong
- Department of Radiology, Peking University People's Hospital, No. 11 Xizhimen South Street, Xicheng District, Beijing, 100044, People's Republic of China.
| |
Collapse
|
5
|
Huang L, Ruan S, Xing Y, Feng M. A review of uncertainty quantification in medical image analysis: Probabilistic and non-probabilistic methods. Med Image Anal 2024; 97:103223. [PMID: 38861770 DOI: 10.1016/j.media.2024.103223] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Revised: 03/16/2024] [Accepted: 05/27/2024] [Indexed: 06/13/2024]
Abstract
The comprehensive integration of machine learning healthcare models within clinical practice remains suboptimal, notwithstanding the proliferation of high-performing solutions reported in the literature. A predominant factor hindering widespread adoption pertains to an insufficiency of evidence affirming the reliability of the aforementioned models. Recently, uncertainty quantification methods have been proposed as a potential solution to quantify the reliability of machine learning models and thus increase the interpretability and acceptability of the results. In this review, we offer a comprehensive overview of the prevailing methods proposed to quantify the uncertainty inherent in machine learning models developed for various medical image tasks. Contrary to earlier reviews that exclusively focused on probabilistic methods, this review also explores non-probabilistic approaches, thereby furnishing a more holistic survey of research pertaining to uncertainty quantification for machine learning models. Analysis of medical images with the summary and discussion on medical applications and the corresponding uncertainty evaluation protocols are presented, which focus on the specific challenges of uncertainty in medical image analysis. We also highlight some potential future research work at the end. Generally, this review aims to allow researchers from both clinical and technical backgrounds to gain a quick and yet in-depth understanding of the research in uncertainty quantification for medical image analysis machine learning models.
Collapse
Affiliation(s)
- Ling Huang
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore
| | - Su Ruan
- Quantif, LITIS, University of Rouen Normandy, France.
| | - Yucheng Xing
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore
| | - Mengling Feng
- Saw Swee Hock School of Public Health, National University of Singapore, Singapore; Institute of Data Science, National University of Singapore, Singapore
| |
Collapse
|
6
|
Manzke M, Iseke S, Böttcher B, Klemenz AC, Weber MA, Meinel FG. Development and performance evaluation of fully automated deep learning-based models for myocardial segmentation on T1 mapping MRI data. Sci Rep 2024; 14:18895. [PMID: 39143126 PMCID: PMC11324648 DOI: 10.1038/s41598-024-69529-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2024] [Accepted: 08/06/2024] [Indexed: 08/16/2024] Open
Abstract
To develop a deep learning-based model capable of segmenting the left ventricular (LV) myocardium on native T1 maps from cardiac MRI in both long-axis and short-axis orientations. Models were trained on native myocardial T1 maps from 50 healthy volunteers and 75 patients using manual segmentation as the reference standard. Based on a U-Net architecture, we systematically optimized the model design using two different training metrics (Sørensen-Dice coefficient = DSC and Intersection-over-Union = IOU), two different activation functions (ReLU and LeakyReLU) and various numbers of training epochs. Training with DSC metric and a ReLU activation function over 35 epochs achieved the highest overall performance (mean error in T1 10.6 ± 17.9 ms, mean DSC 0.88 ± 0.07). Limits of agreement between model results and ground truth were from -35.5 to + 36.1 ms. This was superior to the agreement between two human raters (-34.7 to + 59.1 ms). Segmentation was as accurate for long-axis views (mean error T1: 6.77 ± 8.3 ms, mean DSC: 0.89 ± 0.03) as for short-axis images (mean error ΔT1: 11.6 ± 19.7 ms, mean DSC: 0.88 ± 0.08). Fully automated segmentation and quantitative analysis of native myocardial T1 maps is possible in both long-axis and short-axis orientations with very high accuracy.
Collapse
Affiliation(s)
- Mathias Manzke
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany
| | - Simon Iseke
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany
| | - Benjamin Böttcher
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany
| | - Ann-Christin Klemenz
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany
| | - Marc-André Weber
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany
| | - Felix G Meinel
- Institute of Diagnostic and Interventional Radiology, Pediatric Radiology and Neuroradiology, University Medical Centre Rostock, Ernst-Heydemann-Str. 6, 18057, Rostock, Germany.
| |
Collapse
|
7
|
Lambert B, Forbes F, Doyle S, Dehaene H, Dojat M. Trustworthy clinical AI solutions: A unified review of uncertainty quantification in Deep Learning models for medical image analysis. Artif Intell Med 2024; 150:102830. [PMID: 38553168 DOI: 10.1016/j.artmed.2024.102830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2023] [Revised: 02/28/2024] [Accepted: 03/01/2024] [Indexed: 04/02/2024]
Abstract
The full acceptance of Deep Learning (DL) models in the clinical field is rather low with respect to the quantity of high-performing solutions reported in the literature. End users are particularly reluctant to rely on the opaque predictions of DL models. Uncertainty quantification methods have been proposed in the literature as a potential solution, to reduce the black-box effect of DL models and increase the interpretability and the acceptability of the result by the final user. In this review, we propose an overview of the existing methods to quantify uncertainty associated with DL predictions. We focus on applications to medical image analysis, which present specific challenges due to the high dimensionality of images and their variable quality, as well as constraints associated with real-world clinical routine. Moreover, we discuss the concept of structural uncertainty, a corpus of methods to facilitate the alignment of segmentation uncertainty estimates with clinical attention. We then discuss the evaluation protocols to validate the relevance of uncertainty estimates. Finally, we highlight the open challenges for uncertainty quantification in the medical field.
Collapse
Affiliation(s)
- Benjamin Lambert
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut des Neurosciences, Grenoble, 38000, France; Pixyl Research and Development Laboratory, Grenoble, 38000, France
| | - Florence Forbes
- Univ. Grenoble Alpes, Inria, CNRS, Grenoble INP, LJK, Grenoble, 38000, France
| | - Senan Doyle
- Pixyl Research and Development Laboratory, Grenoble, 38000, France
| | - Harmonie Dehaene
- Pixyl Research and Development Laboratory, Grenoble, 38000, France
| | - Michel Dojat
- Univ. Grenoble Alpes, Inserm, U1216, Grenoble Institut des Neurosciences, Grenoble, 38000, France.
| |
Collapse
|
8
|
Mariscal-Harana J, Asher C, Vergani V, Rizvi M, Keehn L, Kim RJ, Judd RM, Petersen SE, Razavi R, King AP, Ruijsink B, Puyol-Antón E. An artificial intelligence tool for automated analysis of large-scale unstructured clinical cine cardiac magnetic resonance databases. EUROPEAN HEART JOURNAL. DIGITAL HEALTH 2023; 4:370-383. [PMID: 37794871 PMCID: PMC10545512 DOI: 10.1093/ehjdh/ztad044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/13/2023] [Revised: 06/05/2023] [Accepted: 07/12/2023] [Indexed: 10/06/2023]
Abstract
Aims Artificial intelligence (AI) techniques have been proposed for automating analysis of short-axis (SAX) cine cardiac magnetic resonance (CMR), but no CMR analysis tool exists to automatically analyse large (unstructured) clinical CMR datasets. We develop and validate a robust AI tool for start-to-end automatic quantification of cardiac function from SAX cine CMR in large clinical databases. Methods and results Our pipeline for processing and analysing CMR databases includes automated steps to identify the correct data, robust image pre-processing, an AI algorithm for biventricular segmentation of SAX CMR and estimation of functional biomarkers, and automated post-analysis quality control to detect and correct errors. The segmentation algorithm was trained on 2793 CMR scans from two NHS hospitals and validated on additional cases from this dataset (n = 414) and five external datasets (n = 6888), including scans of patients with a range of diseases acquired at 12 different centres using CMR scanners from all major vendors. Median absolute errors in cardiac biomarkers were within the range of inter-observer variability: <8.4 mL (left ventricle volume), <9.2 mL (right ventricle volume), <13.3 g (left ventricular mass), and <5.9% (ejection fraction) across all datasets. Stratification of cases according to phenotypes of cardiac disease and scanner vendors showed good performance across all groups. Conclusion We show that our proposed tool, which combines image pre-processing steps, a domain-generalizable AI algorithm trained on a large-scale multi-domain CMR dataset and quality control steps, allows robust analysis of (clinical or research) databases from multiple centres, vendors, and cardiac diseases. This enables translation of our tool for use in fully automated processing of large multi-centre databases.
Collapse
Affiliation(s)
- Jorge Mariscal-Harana
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
| | - Clint Asher
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
- Department of Adult and Paediatric Cardiology, Guy’s and St Thomas’ NHS Foundation Trust, Westminster Bridge Road, London SE1 7EH, London, UK
| | - Vittoria Vergani
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
| | - Maleeha Rizvi
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
- Department of Adult and Paediatric Cardiology, Guy’s and St Thomas’ NHS Foundation Trust, Westminster Bridge Road, London SE1 7EH, London, UK
| | - Louise Keehn
- Department of Clinical Pharmacology, King’s College London British Heart Foundation Centre, St Thomas’ Hospital, London, Westminster Bridge Road, London SE1 7EH, UK
| | - Raymond J Kim
- Division of Cardiology, Department of Medicine, Duke University, 40 Duke Medicine Circle, Durham, NC 27710, USA
| | - Robert M Judd
- Division of Cardiology, Department of Medicine, Duke University, 40 Duke Medicine Circle, Durham, NC 27710, USA
| | - Steffen E Petersen
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, Charterhouse Square, London EC1M 6BQ, UK
- Barts Heart Centre, St Bartholomew’s Hospital, Barts Health NHS Trust, W Smithfield, London EC1A 7BE, UK
- Health Data Research UK, Gibbs Building, 215 Euston Rd., London NW1 2BE, UK
- Alan Turing Institute, 96 Euston Rd., London NW1 2DB, UK
| | - Reza Razavi
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
- Department of Adult and Paediatric Cardiology, Guy’s and St Thomas’ NHS Foundation Trust, Westminster Bridge Road, London SE1 7EH, London, UK
| | - Andrew P King
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
| | - Bram Ruijsink
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
- Department of Adult and Paediatric Cardiology, Guy’s and St Thomas’ NHS Foundation Trust, Westminster Bridge Road, London SE1 7EH, London, UK
- Department of Cardiology, Division Heart and Lungs, University Medical Center Utrecht, Utrecht University, 3584 CX Utrecht, the Netherlands
| | - Esther Puyol-Antón
- School of Biomedical Engineering & Imaging Sciences Rayne Institute, 4th Floor, Lambeth Wing St. Thomas' Hospital Westminster Bridge Road London SE1 7EH
| |
Collapse
|
9
|
Jin S, Han H, Huang Z, Xiang Y, Du M, Hua F, Guan X, Liu J, Chen F, He H. Automatic three-dimensional nasal and pharyngeal airway subregions identification via Vision Transformer. J Dent 2023; 136:104595. [PMID: 37343616 DOI: 10.1016/j.jdent.2023.104595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2023] [Revised: 06/06/2023] [Accepted: 06/19/2023] [Indexed: 06/23/2023] Open
Abstract
OBJECTIVES Upper airway assessment requires a fully-automated segmentation system for complete or sub-regional identification. This study aimed to develop a novel Deep Learning (DL) model for accurate segmentation of the upper airway and achieve entire and subregional identification. METHODS Fifty cone-beam computed tomography (CBCT) scans, including 24,502 slices, were labelled as the ground truth by one orthodontist and two otorhinolaryngologists. A novel model, a lightweight multitask network based on the Swin Transformer and U-Net, was built for automatic segmentation of the entire upper airway and subregions. Segmentation performance was evaluated using Precision, Recall, Dice similarity coefficient (DSC) and Intersection over union (IoU). The clinical implications of the precision errors were quantitatively analysed, and comparisons between the AI model and Dolphin software were conducted. RESULTS Our model achieved good performance with a precision of 85.88-94.25%, recall of 93.74-98.44%, DSC of 90.95-96.29%, IoU of 83.68-92.85% in the overall and subregions of three-dimensional (3D) upper airway, and a precision of 91.22-97.51%, recall of 90.70-97.62%, DSC of 90.92-97.55%, and IoU of 83.41-95.29% in the subregions of two-dimensional (2D) crosssections. Discrepancies in volume and area caused by precision errors did not affect clinical outcomes. Both our AI model and the Dolphin software provided clinically acceptable consistency for pharyngeal airway assessments. CONCLUSION The novel DL model not only achieved segmentation of the entire upper airway, including the nasal cavity and subregion identification, but also performed exceptionally well, making it well suited for 3D upper airway assessment from the nasal cavity to the hypopharynx, especially for intricate structures. CLINICAL SIGNIFICANCE This system provides insights into the aetiology, risk, severity, treatment effect, and prognosis of dentoskeletal deformities and obstructive sleep apnea. It achieves rapid assessment of the entire upper airway and its subregions, making airway management-an integral part of orthodontic treatment, orthognathic surgery, and ENT surgery-easier.
Collapse
Affiliation(s)
- Suhan Jin
- Department of Orthodontics, Hubei-MOST KLOS & KLOBM, School & Hospital of Stomatology, Wuhan University,Wuhan, China; Department of Orthodontics, Affiliated Stomatological Hospital of Zunyi Medical University, Zunyi, China
| | - Haojie Han
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, Nanjing, China
| | - Zhiqun Huang
- Department of Otolaryngology-Head and Neck Surgery, Renmin Hospital of Wuhan University, Wuhan, China
| | - Yuandi Xiang
- Department of Otolaryngology-Head and Neck Surgery, Renmin Hospital of Wuhan University, Wuhan, China
| | - Mingyuan Du
- Department of Orthodontics, Hubei-MOST KLOS & KLOBM, School & Hospital of Stomatology, Wuhan University,Wuhan, China
| | - Fang Hua
- Department of Orthodontics, Hubei-MOST KLOS & KLOBM, School & Hospital of Stomatology, Wuhan University,Wuhan, China
| | - Xiaoyan Guan
- Department of Orthodontics, Affiliated Stomatological Hospital of Zunyi Medical University, Zunyi, China
| | - Jianguo Liu
- School of Stomatology, Zunyi Medical University, Zunyi, China; Special Key Laboratory of Oral Diseases Research, Higher Education Institution, Zunyi, China
| | - Fang Chen
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, Nanjing, China.
| | - Hong He
- Department of Orthodontics, Hubei-MOST KLOS & KLOBM, School & Hospital of Stomatology, Wuhan University,Wuhan, China.
| |
Collapse
|