1
|
Zhu Q, Bi Y, Chen J, Chu X, Wang D, Wang Y. Central loss guides coordinated Transformer for reliable anatomical landmark detection. Neural Netw 2025; 187:107391. [PMID: 40138918 DOI: 10.1016/j.neunet.2025.107391] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2024] [Revised: 02/06/2025] [Accepted: 03/10/2025] [Indexed: 03/29/2025]
Abstract
Heatmap-based anatomical landmark detection is still facing two unresolved challenges: (1) inability to accurately evaluate the distribution of heatmap; (2) inability to effectively exploit global spatial structure information. To address the computational inability challenge, we propose a novel position-aware and sample-aware central loss. Specifically, our central loss can absorb position information, enabling accurate evaluation of the heatmap distribution. More advanced is that our central loss is sample-aware, which can adaptively distinguish easy and hard samples and make the model more focused on hard samples while solving the challenge of extreme imbalance between landmarks and non-landmarks. To address the challenge of ignoring structure information, a Coordinated Transformer, called CoorTransformer, is proposed, which establishes long-range dependencies under the guidance of landmark coordinate information, making the attention more focused on the sparse landmarks while taking advantage of global spatial structure. Furthermore, CoorTransformer can speed up convergence, effectively avoiding the defect that Transformers have difficulty converging in sparse representation learning. Using the advanced CoorTransformer and central loss, we propose a generalized detection model that can handle various scenarios, inherently exploiting the underlying relationship between landmarks and incorporating rich structural knowledge around the target landmarks. We analyzed and evaluated CoorTransformer and central loss on three challenging landmark detection tasks. The experimental results show that our CoorTransformer outperforms state-of-the-art methods, and the central loss significantly improves the model's performance with p-values <0.05. The source code of this work is available at the GitHub repository.
Collapse
Affiliation(s)
- Qikui Zhu
- Department of Biomedical Engineering, Case Western Reserve University, OH, USA.
| | - Yihui Bi
- Department of Orthopaedics, The Second Affiliated Hospital of Anhui Medical University, Hefei 230601, China; Institute of Orthopaedics, Research Center for Translational Medicine, The Second Affiliated Hospital of Anhui Medical University, Hefei 230601, China
| | - Jie Chen
- Department of Pathology and Institute of Clinical Pathology, West China Hospital, Chengdu, China
| | - Xiangpeng Chu
- Guangzhou Twelfth People's Hospital, Guangzhou Occupational Disease Prevention and Treatment Hospital, Guangzhou Otolaryngology-head and Neck Surgery Hospital, Guangzhou, China
| | - Danxin Wang
- College of Computer Science and Technology, China University of Petroleum, Qingdao 266580, China
| | - Yanqing Wang
- Department of Gynecology, Renmin Hospital of Wuhan University, Wuhan, China
| |
Collapse
|
2
|
Dai W, Lu R, Zhu J, Chen P, Yang H. Harnessing unlabeled data: Enhanced rare earth component content prediction based on BiLSTM-Deep autoencoder. ISA TRANSACTIONS 2025; 157:357-367. [PMID: 39757067 DOI: 10.1016/j.isatra.2024.12.027] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2024] [Revised: 12/13/2024] [Accepted: 12/13/2024] [Indexed: 01/07/2025]
Abstract
Traditional data-driven models for predicting rare earth component content are primarily developed by relying on supervised learning methods, which suffer from limitations such as a lack of labeled data, lagging, and poor usage of a major amount of unlabeled data. This paper proposes a novel prediction approach based on the BiLSTM-Deep autoencoder enhanced traditional LSSVM algorithm, termed BiLSTM-DeepAE-LSSVM. This approach thoroughly exploits the implicit information contained in copious amounts of unlabeled data in the rare earth production process, thereby improving the traditional supervised prediction method and increasing the accuracy of component content predictions. Initially, a BiLSTM autoencoder is established for unsupervised training on the rare earth production process data, enabling the extraction of inherent time series characteristics. Subsequently, boolean vectors are introduced in the Deep autoencoder training process to perform masking operations on the input data, simulating scenarios with noise and missing data. This is facilitated by their adherence to Bernoulli distributions, which allow for the random setting of certain input vector dimensions to zero. Additionally, the Deep autoencoder is capable of extracting high-dimensional implicit features from the data. After that, the conventional supervised prediction technique, least squares support vector machine (LSSVM), is fused with the implicit characteristics derived from the well-constructed BiLSTM-Deep autoencoder, culminating in the creation of a prediction model for rare earth component content. Ultimately, the simulation verification using LaCe/PrNd extraction field data demonstrates the effectiveness of the proposed approach in harnessing substantial quantities of unlabeled data from the rare earth extraction production process, thereby bolstering the accuracy of model predictions.
Collapse
Affiliation(s)
- Wenhao Dai
- School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang, 330013, Jiangxi, China; Key Laboratory of Advanced Control & Optimization of Jiangxi Province, Nanchang, 330013, Jiangxi, China.
| | - Rongxiu Lu
- School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang, 330013, Jiangxi, China; Key Laboratory of Advanced Control & Optimization of Jiangxi Province, Nanchang, 330013, Jiangxi, China.
| | - Jianyong Zhu
- School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang, 330013, Jiangxi, China; Key Laboratory of Advanced Control & Optimization of Jiangxi Province, Nanchang, 330013, Jiangxi, China.
| | - Pengzhan Chen
- School of Intelligent Manufacture, Taizhou University, Taizhou, 318000, Zhejiang, China.
| | - Hui Yang
- School of Electrical and Automation Engineering, East China Jiaotong University, Nanchang, 330013, Jiangxi, China; Key Laboratory of Advanced Control & Optimization of Jiangxi Province, Nanchang, 330013, Jiangxi, China.
| |
Collapse
|
3
|
Ramesh M, Reibman AR. SURABHI: Self-Training Using Rectified Annotations-Based Hard Instances for Eidetic Cattle Recognition. SENSORS (BASEL, SWITZERLAND) 2024; 24:7680. [PMID: 39686218 DOI: 10.3390/s24237680] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/07/2024] [Revised: 11/28/2024] [Accepted: 11/28/2024] [Indexed: 12/18/2024]
Abstract
We propose a self-training scheme, SURABHI, that trains deep-learning keypoint detection models on machine-annotated instances, together with the methodology to generate those instances. SURABHI aims to improve the keypoint detection accuracy not by altering the structure of a deep-learning-based keypoint detector model but by generating highly effective training instances. The machine-annotated instances used in SURABHI are hard instances-instances that require a rectifier to correct the keypoints misplaced by the keypoint detection model. We engineer this scheme for the task of predicting keypoints of cattle from the top, in conjunction with our Eidetic Cattle Recognition System, which is dependent on accurate prediction of keypoints for predicting the correct cow ID. We show that the final cow ID prediction accuracy on previously unseen cows also improves significantly after applying SURABHI to a deep-learning detection model with high capacity, especially when available training data are minimal. SURABHI helps us achieve a top-6 cow recognition accuracy of 91.89% on a dataset of cow videos. Using SURABHI on this dataset also improves the number of cow instances with correct identification by 22% over the baseline result from fully supervised training.
Collapse
Affiliation(s)
- Manu Ramesh
- School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN 47907, USA
| | - Amy R Reibman
- School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN 47907, USA
| |
Collapse
|
4
|
Song C, Jeong Y, Huh H, Park JW, Paeng JY, Ahn J, Son J, Jung E. Multi-Scale 3D Cephalometric Landmark Detection Based on Direct Regression with 3D CNN Architectures. Diagnostics (Basel) 2024; 14:2605. [PMID: 39594271 PMCID: PMC11592740 DOI: 10.3390/diagnostics14222605] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2024] [Revised: 11/13/2024] [Accepted: 11/19/2024] [Indexed: 11/28/2024] Open
Abstract
BACKGROUND Cephalometric analysis is important in diagnosing and planning treatments for patients, traditionally relying on 2D cephalometric radiographs. With advancements in 3D imaging, automated landmark detection using deep learning has gained prominence. However, 3D imaging introduces challenges due to increased network complexity and computational demands. This study proposes a multi-scale 3D CNN-based approach utilizing direct regression to improve the accuracy of maxillofacial landmark detection. METHODS The method employs a coarse-to-fine framework, first identifying landmarks in a global context and then refining their positions using localized 3D patches. A clinical dataset of 150 CT scans from maxillofacial surgery patients, annotated with 30 anatomical landmarks, was used for training and evaluation. RESULTS The proposed method achieved an average RMSE of 2.238 mm, outperforming conventional 3D CNN architectures. The approach demonstrated consistent detection without failure cases. CONCLUSIONS Our multi-scale-based 3D CNN framework provides a reliable method for automated landmark detection in maxillofacial CT images, showing potential for other clinical applications.
Collapse
Affiliation(s)
- Chanho Song
- Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea; (C.S.); (H.H.); (J.-W.P.)
| | - Yoosoo Jeong
- Daegu-Gyeongbuk Research Division, Electronics and Telecommunications Research Institute (ETRI), Daegu 42994, Republic of Korea;
| | - Hyungkyu Huh
- Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea; (C.S.); (H.H.); (J.-W.P.)
| | - Jee-Woong Park
- Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea; (C.S.); (H.H.); (J.-W.P.)
| | - Jun-Young Paeng
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.-Y.P.); (J.A.)
| | - Jaemyung Ahn
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul 06351, Republic of Korea; (J.-Y.P.); (J.A.)
| | - Jaebum Son
- Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea; (C.S.); (H.H.); (J.-W.P.)
| | - Euisung Jung
- Medical Device Development Center, Daegu-Gyeongbuk Medical Innovation Foundation (K-MEDI Hub), Daegu 41061, Republic of Korea; (C.S.); (H.H.); (J.-W.P.)
| |
Collapse
|
5
|
Sahlsten J, Järnstedt J, Jaskari J, Naukkarinen H, Mahasantipiya P, Charuakkra A, Vasankari K, Hietanen A, Sundqvist O, Lehtinen A, Kaski K. Deep learning for 3D cephalometric landmarking with heterogeneous multi-center CBCT dataset. PLoS One 2024; 19:e0305947. [PMID: 38917161 PMCID: PMC11198780 DOI: 10.1371/journal.pone.0305947] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2024] [Accepted: 06/07/2024] [Indexed: 06/27/2024] Open
Abstract
Cephalometric analysis is critically important and common procedure prior to orthodontic treatment and orthognathic surgery. Recently, deep learning approaches have been proposed for automatic 3D cephalometric analysis based on landmarking from CBCT scans. However, these approaches have relied on uniform datasets from a single center or imaging device but without considering patient ethnicity. In addition, previous works have considered a limited number of clinically relevant cephalometric landmarks and the approaches were computationally infeasible, both impairing integration into clinical workflow. Here our aim is to analyze the clinical applicability of a light-weight deep learning neural network for fast localization of 46 clinically significant cephalometric landmarks with multi-center, multi-ethnic, and multi-device data consisting of 309 CBCT scans from Finnish and Thai patients. The localization performance of our approach resulted in the mean distance of 1.99 ± 1.55 mm for the Finnish cohort and 1.96 ± 1.25 mm for the Thai cohort. This performance turned out to be clinically significant i.e., ≤ 2 mm with 61.7% and 64.3% of the landmarks with Finnish and Thai cohorts, respectively. Furthermore, the estimated landmarks were used to measure cephalometric characteristics successfully i.e., with ≤ 2 mm or ≤ 2° error, on 85.9% of the Finnish and 74.4% of the Thai cases. Between the two patient cohorts, 33 of the landmarks and all cephalometric characteristics had no statistically significant difference (p < 0.05) measured by the Mann-Whitney U test with Benjamini-Hochberg correction. Moreover, our method is found to be computationally light, i.e., providing the predictions with the mean duration of 0.77 s and 2.27 s with single machine GPU and CPU computing, respectively. Our findings advocate for the inclusion of this method into clinical settings based on its technical feasibility and robustness across varied clinical datasets.
Collapse
Affiliation(s)
- Jaakko Sahlsten
- Department of Computer Science, Aalto University School of Science, Espoo, Finland
| | - Jorma Järnstedt
- Department of Radiology, Tampere University Hospital, Wellbeing Services County of Pirkanmaa, Tampere, Finland
- Faculty of Medicine and Health Technology, University of Tampere, Tampere, Finland
| | - Joel Jaskari
- Department of Computer Science, Aalto University School of Science, Espoo, Finland
| | | | - Phattaranant Mahasantipiya
- Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Chiang Mai, Thailand
- Division of Oral and Maxillofacial Radiology, Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Chiang Mai, Thailand
| | - Arnon Charuakkra
- Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Chiang Mai, Thailand
- Division of Oral and Maxillofacial Radiology, Department of Oral Biology and Diagnostic Sciences, Faculty of Dentistry, Chiang Mai University, Chiang Mai, Thailand
| | - Krista Vasankari
- Department of Oral Diseases, Tampere University Hospital, Tampere, Finland
| | | | | | - Antti Lehtinen
- Department of Radiology, Tampere University Hospital, Wellbeing Services County of Pirkanmaa, Tampere, Finland
- Faculty of Medicine and Health Technology, University of Tampere, Tampere, Finland
| | - Kimmo Kaski
- Department of Computer Science, Aalto University School of Science, Espoo, Finland
- The Alan Turing Institute, British Library, London, United Kingdom
| |
Collapse
|
6
|
Hector I, Panjanathan R. Predictive maintenance in Industry 4.0: a survey of planning models and machine learning techniques. PeerJ Comput Sci 2024; 10:e2016. [PMID: 38855197 PMCID: PMC11157603 DOI: 10.7717/peerj-cs.2016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 04/02/2024] [Indexed: 06/11/2024]
Abstract
Equipment downtime resulting from maintenance in various sectors around the globe has become a major concern. The effectiveness of conventional reactive maintenance methods in addressing interruptions and enhancing operational efficiency has become inadequate. Therefore, acknowledging the constraints associated with reactive maintenance and the growing need for proactive approaches to proactively detect possible breakdowns is necessary. The need for optimisation of asset management and reduction of costly downtime emerges from the demand for industries. The work highlights the use of Internet of Things (IoT)-enabled Predictive Maintenance (PdM) as a revolutionary strategy across many sectors. This article presents a picture of a future in which the use of IoT technology and sophisticated analytics will enable the prediction and proactive mitigation of probable equipment failures. This literature study has great importance as it thoroughly explores the complex steps and techniques necessary for the development and implementation of efficient PdM solutions. The study offers useful insights into the optimisation of maintenance methods and the enhancement of operational efficiency by analysing current information and approaches. The article outlines essential stages in the application of PdM, encompassing underlying design factors, data preparation, feature selection, and decision modelling. Additionally, the study discusses a range of ML models and methodologies for monitoring conditions. In order to enhance maintenance plans, it is necessary to prioritise ongoing study and improvement in the field of PdM. The potential for boosting PdM skills and guaranteeing the competitiveness of companies in the global economy is significant through the incorporation of IoT, Artificial Intelligence (AI), and advanced analytics.
Collapse
Affiliation(s)
- Ida Hector
- School of Computer Science and Engineering, Vellore Institute of Technology Chennai, Chennai, Tamil Nadu, India
| | - Rukmani Panjanathan
- School of Computer Science and Engineering, Vellore Institute of Technology Chennai, Chennai, Tamil Nadu, India
| |
Collapse
|
7
|
Creux C, Zehraoui F, Hanczar B, Tahi F. A3SOM, abstained explainable semi-supervised neural network based on self-organizing map. PLoS One 2023; 18:e0286137. [PMID: 37228138 DOI: 10.1371/journal.pone.0286137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Accepted: 05/09/2023] [Indexed: 05/27/2023] Open
Abstract
In the sea of data generated daily, unlabeled samples greatly outnumber labeled ones. This is due to the fact that, in many application areas, labels are scarce or hard to obtain. In addition, unlabeled samples might belong to new classes that are not available in the label set associated with data. In this context, we propose A3SOM, an abstained explainable semi-supervised neural network that associates a self-organizing map to dense layers in order to classify samples. Abstained classification enables the detection of new classes and class overlaps. The use of a self-organizing map in A3SOM allows integrated visualization and makes the model explainable. Along with describing our approach, this paper shows that the method is competitive with other classifiers and demonstrates the benefits of including abstention rules. A use case is presented on breast cancer subtype classification and discovery to show the relevance of our method in real-world medical problems.
Collapse
Affiliation(s)
- Constance Creux
- Univ Evry, IBISC, Université Paris-Saclay, Evry-Courcouronnes, France
| | - Farida Zehraoui
- Univ Evry, IBISC, Université Paris-Saclay, Evry-Courcouronnes, France
| | - Blaise Hanczar
- Univ Evry, IBISC, Université Paris-Saclay, Evry-Courcouronnes, France
| | - Fariza Tahi
- Univ Evry, IBISC, Université Paris-Saclay, Evry-Courcouronnes, France
| |
Collapse
|
8
|
Serafin M, Baldini B, Cabitza F, Carrafiello G, Baselli G, Del Fabbro M, Sforza C, Caprioglio A, Tartaglia GM. Accuracy of automated 3D cephalometric landmarks by deep learning algorithms: systematic review and meta-analysis. LA RADIOLOGIA MEDICA 2023; 128:544-555. [PMID: 37093337 PMCID: PMC10181977 DOI: 10.1007/s11547-023-01629-2] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Accepted: 03/28/2023] [Indexed: 04/25/2023]
Abstract
OBJECTIVES The aim of the present systematic review and meta-analysis is to assess the accuracy of automated landmarking using deep learning in comparison with manual tracing for cephalometric analysis of 3D medical images. METHODS PubMed/Medline, IEEE Xplore, Scopus and ArXiv electronic databases were searched. Selection criteria were: ex vivo and in vivo volumetric data images suitable for 3D landmarking (Problem), a minimum of five automated landmarking performed by deep learning method (Intervention), manual landmarking (Comparison), and mean accuracy, in mm, between manual and automated landmarking (Outcome). QUADAS-2 was adapted for quality analysis. Meta-analysis was performed on studies that reported as outcome mean values and standard deviation of the difference (error) between manual and automated landmarking. Linear regression plots were used to analyze correlations between mean accuracy and year of publication. RESULTS The initial electronic screening yielded 252 papers published between 2020 and 2022. A total of 15 studies were included for the qualitative synthesis, whereas 11 studies were used for the meta-analysis. Overall random effect model revealed a mean value of 2.44 mm, with a high heterogeneity (I2 = 98.13%, τ2 = 1.018, p-value < 0.001); risk of bias was high due to the presence of issues for several domains per study. Meta-regression indicated a significant relation between mean error and year of publication (p value = 0.012). CONCLUSION Deep learning algorithms showed an excellent accuracy for automated 3D cephalometric landmarking. In the last two years promising algorithms have been developed and improvements in landmarks annotation accuracy have been done.
Collapse
Affiliation(s)
- Marco Serafin
- Department of Biomedical Sciences for Health, University of Milan, Via Mangiagalli 31, 20133, Milan, Italy
| | - Benedetta Baldini
- Department of Electronics, Information and Bioengineering, Politecnico Di Milano, Via Ponzio 34/5, 20133, Milan, Italy.
| | - Federico Cabitza
- Department of Informatics, System and Communication, University of Milano-Bicocca, Viale Sarca 336, 20126, Milan, Italy
- IRCCS Istituto Ortopedico Galeazzi, Via Belgioioso 173, 20157, Milan, Italy
| | - Gianpaolo Carrafiello
- Department of Oncology and Hematology-Oncology, University of Milan, Via Sforza 35, 20122, Milan, Italy
- Fondazione IRCCS Cà Granda, Ospedale Maggiore Policlinico, Via Sforza 35, 20122, Milan, Italy
| | - Giuseppe Baselli
- Department of Electronics, Information and Bioengineering, Politecnico Di Milano, Via Ponzio 34/5, 20133, Milan, Italy
| | - Massimo Del Fabbro
- Department of Biomedical, Surgical and Dental Sciences, University of Milan, Via della Commenda 10, 20122, Milan, Italy
- Fondazione IRCCS Cà Granda, Ospedale Maggiore Policlinico, Via Sforza 35, 20122, Milan, Italy
| | - Chiarella Sforza
- Department of Biomedical Sciences for Health, University of Milan, Via Mangiagalli 31, 20133, Milan, Italy
| | - Alberto Caprioglio
- Department of Biomedical, Surgical and Dental Sciences, University of Milan, Via della Commenda 10, 20122, Milan, Italy
- Fondazione IRCCS Cà Granda, Ospedale Maggiore Policlinico, Via Sforza 35, 20122, Milan, Italy
| | - Gianluca M Tartaglia
- Department of Biomedical, Surgical and Dental Sciences, University of Milan, Via della Commenda 10, 20122, Milan, Italy
- Fondazione IRCCS Cà Granda, Ospedale Maggiore Policlinico, Via Sforza 35, 20122, Milan, Italy
| |
Collapse
|
9
|
Zhang C, Yang G, Li F, Wen Y, Yao Y, Shu H, Simon A, Dillenseger JL, Coatrieux JL. CTANet: Confidence-based Threshold Adaption Network for Semi-supervised Segmentation of Uterine Regions from MR Images for HIFU Treatment. Ing Rech Biomed 2023. [DOI: 10.1016/j.irbm.2022.100747] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
|
10
|
Xu J, Zeng B, Egger J, Wang C, Smedby Ö, Jiang X, Chen X. A review on AI-based medical image computing in head and neck surgery. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac840f] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Accepted: 07/25/2022] [Indexed: 11/11/2022]
Abstract
Abstract
Head and neck surgery is a fine surgical procedure with a complex anatomical space, difficult operation and high risk. Medical image computing (MIC) that enables accurate and reliable preoperative planning is often needed to reduce the operational difficulty of surgery and to improve patient survival. At present, artificial intelligence, especially deep learning, has become an intense focus of research in MIC. In this study, the application of deep learning-based MIC in head and neck surgery is reviewed. Relevant literature was retrieved on the Web of Science database from January 2015 to May 2022, and some papers were selected for review from mainstream journals and conferences, such as IEEE Transactions on Medical Imaging, Medical Image Analysis, Physics in Medicine and Biology, Medical Physics, MICCAI, etc. Among them, 65 references are on automatic segmentation, 15 references on automatic landmark detection, and eight references on automatic registration. In the elaboration of the review, first, an overview of deep learning in MIC is presented. Then, the application of deep learning methods is systematically summarized according to the clinical needs, and generalized into segmentation, landmark detection and registration of head and neck medical images. In segmentation, it is mainly focused on the automatic segmentation of high-risk organs, head and neck tumors, skull structure and teeth, including the analysis of their advantages, differences and shortcomings. In landmark detection, the focus is mainly on the introduction of landmark detection in cephalometric and craniomaxillofacial images, and the analysis of their advantages and disadvantages. In registration, deep learning networks for multimodal image registration of the head and neck are presented. Finally, their shortcomings and future development directions are systematically discussed. The study aims to serve as a reference and guidance for researchers, engineers or doctors engaged in medical image analysis of head and neck surgery.
Collapse
|
11
|
Dot G, Schouman T, Chang S, Rafflenbeul F, Kerbrat A, Rouch P, Gajny L. Automatic 3-Dimensional Cephalometric Landmarking via Deep Learning. J Dent Res 2022; 101:1380-1387. [PMID: 35982646 DOI: 10.1177/00220345221112333] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
The increasing use of 3-dimensional (3D) imaging by orthodontists and maxillofacial surgeons to assess complex dentofacial deformities and plan orthognathic surgeries implies a critical need for 3D cephalometric analysis. Although promising methods were suggested to localize 3D landmarks automatically, concerns about robustness and generalizability restrain their clinical use. Consequently, highly trained operators remain needed to perform manual landmarking. In this retrospective diagnostic study, we aimed to train and evaluate a deep learning (DL) pipeline based on SpatialConfiguration-Net for automatic localization of 3D cephalometric landmarks on computed tomography (CT) scans. A retrospective sample of consecutive presurgical CT scans was randomly distributed between a training/validation set (n = 160) and a test set (n = 38). The reference data consisted of 33 landmarks, manually localized once by 1 operator(n = 178) or twice by 3 operators (n = 20, test set only). After inference on the test set, 1 CT scan showed "very low" confidence level predictions; we excluded it from the overall analysis but still assessed and discussed the corresponding results. The model performance was evaluated by comparing the predictions with the reference data; the outcome set included localization accuracy, cephalometric measurements, and comparison to manual landmarking reproducibility. On the hold-out test set, the mean localization error was 1.0 ± 1.3 mm, while success detection rates for 2.0, 2.5, and 3.0 mm were 90.4%, 93.6%, and 95.4%, respectively. Mean errors were -0.3 ± 1.3° and -0.1 ± 0.7 mm for angular and linear measurements, respectively. When compared to manual reproducibility, the measurements were within the Bland-Altman 95% limits of agreement for 91.9% and 71.8% of skeletal and dentoalveolar variables, respectively. To conclude, while our DL method still requires improvement, it provided highly accurate 3D landmark localization on a challenging test set, with a reliability for skeletal evaluation on par with what clinicians obtain.
Collapse
Affiliation(s)
- G Dot
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France.,Universite Paris Cite, AP-HP, Hopital Pitie Salpetriere, Service de Medecine Bucco-Dentaire, Paris, France
| | - T Schouman
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France.,Medecine Sorbonne Universite, AP-HP, Hopital Pitie-Salpetriere, Service de Chirurgie Maxillo-Faciale, Paris, France
| | - S Chang
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France
| | - F Rafflenbeul
- Department of Dentofacial Orthopedics, Faculty of Dental Surgery, Strasbourg University, Strasbourg, France
| | - A Kerbrat
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France
| | - P Rouch
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France
| | - L Gajny
- Institut de Biomecanique Humaine Georges Charpak, Arts et Metiers Institute of Technology, Paris, France
| |
Collapse
|