1
|
Zhu Q, Bi Y, Chen J, Chu X, Wang D, Wang Y. Central loss guides coordinated Transformer for reliable anatomical landmark detection. Neural Netw 2025; 187:107391. [PMID: 40138918 DOI: 10.1016/j.neunet.2025.107391] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2024] [Revised: 02/06/2025] [Accepted: 03/10/2025] [Indexed: 03/29/2025]
Abstract
Heatmap-based anatomical landmark detection is still facing two unresolved challenges: (1) inability to accurately evaluate the distribution of heatmap; (2) inability to effectively exploit global spatial structure information. To address the computational inability challenge, we propose a novel position-aware and sample-aware central loss. Specifically, our central loss can absorb position information, enabling accurate evaluation of the heatmap distribution. More advanced is that our central loss is sample-aware, which can adaptively distinguish easy and hard samples and make the model more focused on hard samples while solving the challenge of extreme imbalance between landmarks and non-landmarks. To address the challenge of ignoring structure information, a Coordinated Transformer, called CoorTransformer, is proposed, which establishes long-range dependencies under the guidance of landmark coordinate information, making the attention more focused on the sparse landmarks while taking advantage of global spatial structure. Furthermore, CoorTransformer can speed up convergence, effectively avoiding the defect that Transformers have difficulty converging in sparse representation learning. Using the advanced CoorTransformer and central loss, we propose a generalized detection model that can handle various scenarios, inherently exploiting the underlying relationship between landmarks and incorporating rich structural knowledge around the target landmarks. We analyzed and evaluated CoorTransformer and central loss on three challenging landmark detection tasks. The experimental results show that our CoorTransformer outperforms state-of-the-art methods, and the central loss significantly improves the model's performance with p-values <0.05. The source code of this work is available at the GitHub repository.
Collapse
Affiliation(s)
- Qikui Zhu
- Department of Biomedical Engineering, Case Western Reserve University, OH, USA.
| | - Yihui Bi
- Department of Orthopaedics, The Second Affiliated Hospital of Anhui Medical University, Hefei 230601, China; Institute of Orthopaedics, Research Center for Translational Medicine, The Second Affiliated Hospital of Anhui Medical University, Hefei 230601, China
| | - Jie Chen
- Department of Pathology and Institute of Clinical Pathology, West China Hospital, Chengdu, China
| | - Xiangpeng Chu
- Guangzhou Twelfth People's Hospital, Guangzhou Occupational Disease Prevention and Treatment Hospital, Guangzhou Otolaryngology-head and Neck Surgery Hospital, Guangzhou, China
| | - Danxin Wang
- College of Computer Science and Technology, China University of Petroleum, Qingdao 266580, China
| | - Yanqing Wang
- Department of Gynecology, Renmin Hospital of Wuhan University, Wuhan, China
| |
Collapse
|
2
|
Naskar S, Sharma S, Kuotsu K, Halder S, Pal G, Saha S, Mondal S, Biswas UK, Jana M, Bhattacharjee S. The biomedical applications of artificial intelligence: an overview of decades of research. J Drug Target 2025; 33:717-748. [PMID: 39744873 DOI: 10.1080/1061186x.2024.2448711] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2024] [Revised: 12/13/2024] [Accepted: 12/26/2024] [Indexed: 01/11/2025]
Abstract
A significant area of computer science called artificial intelligence (AI) is successfully applied to the analysis of intricate biological data and the extraction of substantial associations from datasets for a variety of biomedical uses. AI has attracted significant interest in biomedical research due to its features: (i) better patient care through early diagnosis and detection; (ii) enhanced workflow; (iii) lowering medical errors; (v) lowering medical costs; (vi) reducing morbidity and mortality; (vii) enhancing performance; (viii) enhancing precision; and (ix) time efficiency. Quantitative metrics are crucial for evaluating AI implementations, providing insights, enabling informed decisions, and measuring the impact of AI-driven initiatives, thereby enhancing transparency, accountability, and overall impact. The implementation of AI in biomedical fields faces challenges such as ethical and privacy concerns, lack of awareness, technology unreliability, and professional liability. A brief discussion is given of the AI techniques, which include Virtual screening (VS), DL, ML, Hidden Markov models (HMMs), Neural networks (NNs), Generative models (GMs), Molecular dynamics (MD), and Structure-activity relationship (SAR) models. The study explores the application of AI in biomedical fields, highlighting its enhanced predictive accuracy, treatment efficacy, diagnostic efficiency, faster decision-making, personalised treatment strategies, and precise medical interventions.
Collapse
Affiliation(s)
- Sweet Naskar
- Department of Pharmaceutics, Institute of Pharmacy, Kalyani, West Bengal, India
| | - Suraj Sharma
- Department of Pharmaceutics, Sikkim Professional College of Pharmaceutical Sciences, Sikkim, India
| | - Ketousetuo Kuotsu
- Department of Pharmaceutical Technology, Jadavpur University, Kolkata, West Bengal, India
| | - Suman Halder
- Medical Department, Department of Indian Railway, Kharagpur Division, Kharagpur, West Bengal, India
| | - Goutam Pal
- Service Dispensary, ESI Hospital, Hoogly, West Bengal, India
| | - Subhankar Saha
- Department of Pharmaceutical Technology, Jadavpur University, Kolkata, West Bengal, India
| | - Shubhadeep Mondal
- Department of Pharmacology, Momtaz Begum Pharmacy College, Rajarhat, West Bengal, India
| | - Ujjwal Kumar Biswas
- School of Pharmaceutical Science (SPS), Siksha O Anusandhan (SOA) University, Bhubaneswar, Odisha, India
| | - Mayukh Jana
- School of Pharmacy, Centurion University of Technology and Management, Centurion University, Bhubaneswar, Odisha, India
| | - Sunirmal Bhattacharjee
- Department of Pharmaceutics, Bharat Pharmaceutical Technology, Amtali, Agartala, Tripura, India
| |
Collapse
|
3
|
Van Santvliet L, Zappon E, Gsell MAF, Thaler F, Blondeel M, Dymarkowski S, Claessen G, Willems R, Urschler M, Vandenberk B, Plank G, De Vos M. Integrating anatomy and electrophysiology in the healthy human heart: Insights from biventricular statistical shape analysis using universal coordinates. Comput Biol Med 2025; 192:110230. [PMID: 40324309 DOI: 10.1016/j.compbiomed.2025.110230] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2024] [Revised: 03/13/2025] [Accepted: 04/16/2025] [Indexed: 05/07/2025]
Abstract
A cardiac digital twin is a virtual replica of a patient-specific heart, mimicking its anatomy and physiology. A crucial step of building a cardiac digital twin is anatomical twinning, where the computational mesh of the digital twin is tailored to the patient-specific cardiac anatomy. In a number of studies, the effect of anatomical variation on clinically relevant functional measurements like electrocardiograms (ECGs) is investigated, using computational simulations. While such a simulation environment provides researchers with a carefully controlled ground truth, the impact of anatomical differences on functional measurements in real-world patients remains understudied. In this study, we develop a biventricular statistical shape model and use it to quantify the effect of biventricular anatomy on ECG-derived and demographic features, providing novel insights for the development of digital twins of cardiac electrophysiology. To this end, a dataset comprising high-resolution cardiac CT scans from 271 healthy individuals, including athletes, is utilized. Furthermore, a novel, universal, ventricular coordinate-based method is developed to establish lightweight shape correspondence. The performance of the shape model is rigorously established, focusing on its dimensionality reduction capabilities and the training data requirements. The most important variability in healthy ventricles captured by the model is their size, followed by their elongation. These anatomical factors are found to significantly correlate with ECG-derived and demographic features. Additionally, a comprehensive synthetic cohort is made available, featuring ready-to-use biventricular meshes with fiber structures and anatomical region annotations. These meshes are well-suited for electrophysiological simulations.
Collapse
Affiliation(s)
- Lore Van Santvliet
- STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Department of Electrical Engineering (ESAT), KU Leuven, Kasteelpark Arenberg 10, Leuven, 3001, Belgium.
| | - Elena Zappon
- Division of Medical Physics and Biophysics, Gottfried Schatz Research Center, Medical University of Graz, Graz, Austria; BioTechMed-Graz, Graz, Austria
| | - Matthias A F Gsell
- Division of Medical Physics and Biophysics, Gottfried Schatz Research Center, Medical University of Graz, Graz, Austria
| | - Franz Thaler
- Division of Medical Physics and Biophysics, Gottfried Schatz Research Center, Medical University of Graz, Graz, Austria; Institute of Computer Graphics and Vision, Graz University of Technology, Graz, Austria; Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, Graz, Austria
| | - Maarten Blondeel
- Department of Cardiology, University Hospitals Leuven, Herestraat 49, Leuven, 3000, Belgium; Department of Cardiovascular Sciences, KU Leuven, Herestraat 49, Leuven, 3000, Belgium
| | - Steven Dymarkowski
- Division of Radiology, University Hospitals Leuven, Herestraat 49, Leuven, 3000, Belgium
| | - Guido Claessen
- Division of Cardiology, Hartcentrum, Jessa Ziekenhuis, Stadsomvaart 11, Hasselt, 3500, Belgium; Department of Medicine and Life Sciences, University of Hasselt, Stadsomvaart 11, Hasselt, 3500, Belgium
| | - Rik Willems
- Department of Cardiology, University Hospitals Leuven, Herestraat 49, Leuven, 3000, Belgium; Department of Cardiovascular Sciences, KU Leuven, Herestraat 49, Leuven, 3000, Belgium
| | - Martin Urschler
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, Graz, Austria; BioTechMed-Graz, Graz, Austria
| | - Bert Vandenberk
- Department of Cardiology, University Hospitals Leuven, Herestraat 49, Leuven, 3000, Belgium; Department of Cardiovascular Sciences, KU Leuven, Herestraat 49, Leuven, 3000, Belgium
| | - Gernot Plank
- Division of Medical Physics and Biophysics, Gottfried Schatz Research Center, Medical University of Graz, Graz, Austria; BioTechMed-Graz, Graz, Austria
| | - Maarten De Vos
- STADIUS Center for Dynamical Systems, Signal Processing and Data Analytics, Department of Electrical Engineering (ESAT), KU Leuven, Kasteelpark Arenberg 10, Leuven, 3001, Belgium
| |
Collapse
|
4
|
Li Y, Liu S, Zhang Y, Zhang M, Jiang C, Ni M, Jin D, Qian Z, Wang J, Pan X, Yuan H. Deep Learning-enhanced Opportunistic Osteoporosis Screening in Ultralow-Voltage (80 kV) Chest CT: A Preliminary Study. Acad Radiol 2025:S1076-6332(24)00937-1. [PMID: 40318972 DOI: 10.1016/j.acra.2024.11.062] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2024] [Revised: 11/23/2024] [Accepted: 11/24/2024] [Indexed: 05/07/2025]
Abstract
RATIONALE AND OBJECTIVES To explore the feasibility of deep learning (DL)-enhanced, fully automated bone mineral density (BMD) measurement using the ultralow-voltage 80 kV chest CT scans performed for lung cancer screening. MATERIALS AND METHODS This study involved 987 patients who underwent 80 kV chest and 120 kV lumbar CT from January to July 2024. Patients were collected from six CT scanners and divided into the training, validation, and test sets 1 and 2 (561: 177: 112: 137). Four convolutional neural networks (CNNs) were employed for automated segmentation (3D VB-Net and SCN), region of interest extraction (3D VB-Net), and BMD calculation (DenseNet and ResNet) of the target vertebrae (T12-L2). The BMD values of T12-L2 were obtained using 80 and 120 kV quantitative CT (QCT), the latter serving as the standard reference. Linear regression and Bland-Altman analyses were used to compare BMD values between 120 kV QCT and 80 kV CNNs, and between 120 kV QCT and 80 kV QCT. Receiver operating characteristic curve analysis was used to assess the diagnostic performance of the 80 kV CNNs and 80 kV QCT for osteoporosis and low BMD from normal BMD. RESULTS Linear regression and Bland-ltman analyses revealed a stronger correlation (R2=0.991-0.998 and 0.990-0.991, P<0.001) and better agreement (mean error, -1.36 to 1.62 and 1.72 to 2.27 mg/cm3; 95% limits of agreement, -9.73 to 7.01 and -5.71 to 10.19mg/cm3) for BMD between 120 kV QCT and 80 kV CNNs than between 120 kV QCT and 80 kV QCT. The areas under the curve of the 80 kV CNNs and 80 kV QCT in detecting osteoporosis and low BMD were 0.997-1.000 and 0.997-0.998, and 0.998-1.000 and 0.997, respectively. CONCLUSION The DL method could achieve fully automated BMD calculation for opportunistic osteoporosis screening with high accuracy using ultralow-voltage 80 kV chest CT performed for lung cancer screening.
Collapse
Affiliation(s)
- Yali Li
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Suwei Liu
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Yan Zhang
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Mengze Zhang
- The Institute of Intelligent Diagnostics, Beijing United-Imaging Research Institute of Intelligent Imaging, Building 3-4F, 9 Yongteng N. Road, Beijing, China (M.Z., Z.Q.)
| | - Chenyu Jiang
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Ming Ni
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Dan Jin
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Zhen Qian
- The Institute of Intelligent Diagnostics, Beijing United-Imaging Research Institute of Intelligent Imaging, Building 3-4F, 9 Yongteng N. Road, Beijing, China (M.Z., Z.Q.)
| | - Jiangxuan Wang
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Xuemin Pan
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.)
| | - Huishu Yuan
- Department of Radiology, Peking University Third Hospital, 49 Huayuan N Rd, Haidian District, Beijing, China (Y.L., S.L., Y.Z., CC.J., M.N., D.J., J.W., X.P., H.Y.).
| |
Collapse
|
5
|
Groun N, Villalba-Orero M, Casado-Martín L, Lara-Pezzi E, Valero E, Le Clainche S, Garicano-Mena J. Eigenhearts: Cardiac diseases classification using eigenfaces approach. Comput Biol Med 2025; 192:110167. [PMID: 40288290 DOI: 10.1016/j.compbiomed.2025.110167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2024] [Revised: 04/04/2025] [Accepted: 04/04/2025] [Indexed: 04/29/2025]
Abstract
In the realm of cardiovascular medicine, medical imaging plays a crucial role in accurately classifying cardiac diseases and making precise diagnoses. However, the integration of data science techniques in this field presents significant challenges, as it requires a large volume of images, while ethical constraints, high costs, and variability in imaging protocols limit data acquisition. As a consequence, it is necessary to investigate different avenues to overcome this challenge. In this contribution, we offer an innovative tool to conquer this limitation. In particular, we delve into the application of a well recognized method known as the eigenfaces approach to classify cardiac diseases. This approach was originally motivated for efficiently representing pictures of faces using principal component analysis, which provides a set of eigenvectors (aka eigenfaces), explaining the variation between face images. Given its effectiveness in face recognition, we sought to evaluate its applicability to more complex medical imaging datasets. In particular, we integrate this approach with convolutional neural networks to classify echocardiography images taken from mice in five distinct cardiac conditions (healthy, diabetic cardiomyopathy, myocardial infarction, obesity and TAC hypertension). The results show a substantial and noteworthy enhancement when employing the singular value decomposition for pre-processing, with classification accuracy increasing by approximately 50%.
Collapse
Affiliation(s)
- Nourelhouda Groun
- ETSI Aeronáutica y del Espacio - Universidad Politécnica de Madrid, Pl. del Cardenal Cisneros, 3, 28040, Madrid, Spain; Université Mohamed Khider Biskra, BP 145 RP, 07000, Biskra, Algeria.
| | - María Villalba-Orero
- Departamento de Medicina y Cirugía Animal, Facultad de Veterinaria - Universidad Complutense de Madrid, Av. Puerta de Hierro, 28040, Madrid, Spain; Centro Nacional de Investigaciones Cardiovasculares (CNIC), C. de Melchor Fernández Almagro, 3, 28029, Madrid, Spain
| | - Lucía Casado-Martín
- Departamento de Medicina y Cirugía Animal, Facultad de Veterinaria - Universidad Complutense de Madrid, Av. Puerta de Hierro, 28040, Madrid, Spain
| | - Enrique Lara-Pezzi
- Centro Nacional de Investigaciones Cardiovasculares (CNIC), C. de Melchor Fernández Almagro, 3, 28029, Madrid, Spain
| | - Eusebio Valero
- ETSI Aeronáutica y del Espacio - Universidad Politécnica de Madrid, Pl. del Cardenal Cisneros, 3, 28040, Madrid, Spain; Center for Computational Simulation (CCS), Boadilla del Monte, 28660, Madrid, Spain
| | - Soledad Le Clainche
- ETSI Aeronáutica y del Espacio - Universidad Politécnica de Madrid, Pl. del Cardenal Cisneros, 3, 28040, Madrid, Spain; Center for Computational Simulation (CCS), Boadilla del Monte, 28660, Madrid, Spain
| | - Jesús Garicano-Mena
- ETSI Aeronáutica y del Espacio - Universidad Politécnica de Madrid, Pl. del Cardenal Cisneros, 3, 28040, Madrid, Spain; Center for Computational Simulation (CCS), Boadilla del Monte, 28660, Madrid, Spain
| |
Collapse
|
6
|
Wu Y, Zhang Z, Liang J, Wu W, Wu W. Torg-Pavlov ratio qualification to diagnose developmental cervical spinal stenosis based on HRViT neural network. BMC Musculoskelet Disord 2025; 26:405. [PMID: 40269821 PMCID: PMC12016334 DOI: 10.1186/s12891-025-08667-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/03/2024] [Accepted: 04/16/2025] [Indexed: 04/25/2025] Open
Abstract
BACKGROUND Developing computer-assisted methods to measure the Torg-Pavlov ratio (TPR), defined as the ratio of the sagittal diameter of the cervical spinal canal to the sagittal diameter of the corresponding vertebral body on lateral radiographs, can reduce subjective influence and speed up processing. The TPR is a critical diagnostic parameter for developmental cervical spinal stenosis (DCSS), as it normalizes variations in radiographic magnification and provides a cost-effective alternative to CT/MRI in resource-limited settings. No study focusing on automatic measurement was reported. The aim was to develop a deep learning-based model for automatically measuring the TPR, and then to establish the distribution of asymptomatic Chinese TPR. METHODS A total of 1623 lateral cervical X-ray images from normal individuals were collected. 1466 and 157 images were used as the training dataset and testing dataset, respectively. We adopted a neural network called High-Resolution Vision Transformer (HRViT), which was trained on the annotated X-ray image dataset to automatically locate the landmarks and calculate the TPR. The accuracy of the TPR measurement was evaluated using mean absolute error (MAE), intra-class correlation coefficient (ICC), r value and Bland-Altman plot. RESULTS The TPR at C2-C7 was 1.26, 0.92, 0.90, 0.93, 0.92, and 0.89, respectively. The MAE between HRViT and surgeon R1 was 0.01, between surgeon R1 and surgeon R2 was 0.17, between surgeon R1 and surgeon R3 was 0.17. The accuracy of HRViT for DCSS diagnosis was 84.1%, which was greatly higher than those of both surgeon R2 (57.3%) and surgeon R3 (56.7%). The consistency of TPR measurements was 0.77-0.9 (ICC) and 0.78-0.9 (r value) between HRViT and surgeon R1. CONCLUSIONS We have explored a deep-learning algorithm for automated measurement of the TPR on cervical lateral radiographs to diagnose DCSS, which had outstanding performance comparable to clinical senior doctors.
Collapse
Affiliation(s)
- Yao Wu
- Department of Orthopedics, the First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China
- Yichang Central People's Hospital, Yichang, 443000, China
| | - Zhenxi Zhang
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, 518107, China
| | - Jie Liang
- Department of Orthopedics, the First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China
- Yichang Central People's Hospital, Yichang, 443000, China
| | - Weiwen Wu
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, 518107, China
| | - Weifei Wu
- Department of Orthopedics, the First College of Clinical Medical Science, China Three Gorges University, Yichang, 443000, China.
- Yichang Central People's Hospital, Yichang, 443000, China.
- Third-grade Pharmacological Laboratory on Traditional Chinese Medicine, State Administration of Traditional Chinese Medicine, China Three Gorges University, Yichang, 443002, China.
| |
Collapse
|
7
|
Jiang Y, Jiang C, Shi B, Wu Y, Xing S, Liang H, Huang J, Huang X, Huang L, Lin L. Automatic identification of hard and soft tissue landmarks in cone-beam computed tomography via deep learning with diversity datasets: a methodological study. BMC Oral Health 2025; 25:505. [PMID: 40200295 PMCID: PMC11980328 DOI: 10.1186/s12903-025-05831-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/28/2024] [Accepted: 03/17/2025] [Indexed: 04/10/2025] Open
Abstract
BACKGROUND Manual landmark detection in cone beam computed tomography (CBCT) for evaluating craniofacial structures relies on medical expertise and is time-consuming. This study aimed to apply a new deep learning method to predict and locate soft and hard tissue craniofacial landmarks on CBCT in patients with various types of malocclusion. METHODS A total of 498 CBCT images were collected. Following the calibration procedure, two experienced clinicians identified 43 landmarks in the x-, y-, and z-coordinate planes on the CBCT images using Checkpoint Software, creating the ground truth by averaging the landmark coordinates. To evaluate the accuracy of our algorithm, we determined the mean absolute error along the x-, y-, and z-axes and calculated the mean radial error (MRE) between the reference landmark and predicted landmark, as well as the successful detection rate (SDR). RESULTS Each landmark prediction took approximately 4.2 s on a conventional graphics processing unit. The mean absolute error across all coordinates was 0.74 mm. The overall MRE for the 43 landmarks was 1.76 ± 1.13 mm, and the SDR was 60.16%, 91.05%, and 97.58% within 2-, 3-, and 4-mm error ranges of manual marking, respectively. The average MRE of the hard tissue landmarks (32/43) was 1.73 mm, while that for soft tissue landmarks (11/43) was 1.84 mm. CONCLUSIONS Our proposed algorithm demonstrates a clinically acceptable level of accuracy and robustness for automatic detection of CBCT soft- and hard-tissue landmarks across all types of malformations. The potential for artificial intelligence to assist in identifying three dimensional-CT landmarks in routine clinical practice and analysing large datasets for future research is promising.
Collapse
Affiliation(s)
- Yan Jiang
- Department of Stomatology, The First Affiliated Hospital of Fujian Medical University, Tai-Jiang District, No.20 Cha-Ting-Zhong Road, Fuzhou, 350005, China
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China
| | - Canyang Jiang
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China
- Department of Oral and Maxillofacial Surgery, The First Affiliated Hospital of Fujian Medical University, Fuzhou, 350005, China
| | - Bin Shi
- Department of Stomatology, The First Affiliated Hospital of Fujian Medical University, Tai-Jiang District, No.20 Cha-Ting-Zhong Road, Fuzhou, 350005, China
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China
- Department of Oral and Maxillofacial Surgery, The First Affiliated Hospital of Fujian Medical University, Fuzhou, 350005, China
| | - You Wu
- School of Stomatology, Fujian Medical University, Fuzhou, 350122, China
| | - Shuli Xing
- College of Computer Science and Mathematics, Fujian University of Technology, Fujian, 350118, China
| | - Hao Liang
- College of Computer Science and Mathematics, Fujian University of Technology, Fujian, 350118, China
| | - Jianping Huang
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China
- Department of Oral and Maxillofacial Surgery, The First Affiliated Hospital of Fujian Medical University, Fuzhou, 350005, China
| | - Xiaohong Huang
- Department of Stomatology, The First Affiliated Hospital of Fujian Medical University, Tai-Jiang District, No.20 Cha-Ting-Zhong Road, Fuzhou, 350005, China.
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China.
| | - Li Huang
- Department of Stomatology, The First Affiliated Hospital of Fujian Medical University, Tai-Jiang District, No.20 Cha-Ting-Zhong Road, Fuzhou, 350005, China.
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China.
- Department of Oral and Maxillofacial Surgery, The First Affiliated Hospital of Fujian Medical University, Fuzhou, 350005, China.
| | - Lisong Lin
- Department of Stomatology, The First Affiliated Hospital of Fujian Medical University, Tai-Jiang District, No.20 Cha-Ting-Zhong Road, Fuzhou, 350005, China.
- Department of Stomatology, National Regional Medical Center, Binhai Campus of the First Affiliated Hospital, Fujian Medical University, Fuzhou, 350212, China.
- Department of Oral and Maxillofacial Surgery, The First Affiliated Hospital of Fujian Medical University, Fuzhou, 350005, China.
| |
Collapse
|
8
|
Tang H, Liu S, Shi Y, Wei J, Peng J, Feng H. Automatic segmentation and landmark detection of 3D CBCT images using semi supervised learning for assisting orthognathic surgery planning. Sci Rep 2025; 15:8814. [PMID: 40087502 PMCID: PMC11909187 DOI: 10.1038/s41598-025-93317-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2024] [Accepted: 03/06/2025] [Indexed: 03/17/2025] Open
Abstract
Patients with abnormal relative position of the upper and lower jaws (the main part of the facial bones) require orthognathic surgery to improve the occlusal relationship and facial appearance. However, in addition to the retraction and protrusion of the maxillomandibular advancement, these patients may also develop asymmetry. This study aims to use a semi-supervised learning method to demonstrate the maxillary and mandible retraction, protrudation and asymmetry of patients before orthognathic surgery through automatic segmentation of 3D cone beam computed tomography (CBCT) images and landmark detection, so as to provide help for the preoperative planning of orthognathic surgery. Among them, the dice of the semi-supervised algorithm adopted in this study reached 93.41 and 96.89% in maxillary and mandibular segmentation tasks, and the average error of landmark detection tasks reached 1.908 ± 1.166 mm, both of which were superior to the full-supervised algorithm with the same data volume annotation. Therefore, we propose that the method can be applied in a clinical setting to assist surgeons in preoperative planning for orthognathic surgery.
Collapse
Affiliation(s)
- Haomin Tang
- College of Medicine, Guizhou University, Guiyang, 550025, China
| | - Shu Liu
- Department of Orthodontics, Guiyang Hospital of Stomatology, Guiyang, 550002, China
| | - Yongxin Shi
- School of Stomatology, Zunyi Medical University, Guiyang, 563006, China
| | - Jin Wei
- Department of Oral and Maxillofacial Surgery, Guiyang Hospital of Stomatology, Guiyang, 550002, China
| | - Juxiang Peng
- Department of Orthodontics, Guiyang Hospital of Stomatology, Guiyang, 550002, China
| | - Hongchao Feng
- Department of Oral and Maxillofacial Surgery, Guiyang Hospital of Stomatology, Guiyang, 550002, China.
| |
Collapse
|
9
|
Baltus SC, Geitenbeek RTJ, Frieben M, Thibeau-Sutre E, Wolterink JM, Tan CO, Vermeulen MC, Consten ECJ, Broeders IAMJ. Deep learning-based pelvimetry in pelvic MRI volumes for pre-operative difficulty assessment of total mesorectal excision. Surg Endosc 2025; 39:1536-1543. [PMID: 39753930 PMCID: PMC11870868 DOI: 10.1007/s00464-024-11485-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Accepted: 12/14/2024] [Indexed: 03/03/2025]
Abstract
BACKGROUND Specific pelvic bone dimensions have been identified as predictors of total mesorectal excision (TME) difficulty and outcomes. However, manual measurement of these dimensions (pelvimetry) is labor intensive and thus, anatomic criteria are not included in the pre-operative difficulty assessment. In this work, we propose an automated workflow for pelvimetry based on pre-operative magnetic resonance imaging (MRI) volumes. METHODS We implement a deep learning-based framework to measure the predictive pelvic dimensions automatically. A 3D U-Net takes a sagittal T2-weighted MRI volume as input and determines five anatomic landmark locations: promontorium, S3-vertebrae, coccyx, dorsal, and cranial part of the os pubis. The landmarks are used to quantify the lengths of the pelvic inlet, outlet, depth, and the angle of the sacrum. For the development of the network, we used MRI volumes from 1707 patients acquired in eight TME centers. The automated landmark localization and pelvic dimensions measurements are assessed by comparison with manual annotation. RESULTS A center-stratified fivefold cross-validation showed a mean landmark localization error of 5.6 mm. The inter-observer variation for manual annotation was 3.7 ± 8.4 mm. The automated dimension measurements had a Spearman correlation coefficient ranging between 0.7 and 0.87. CONCLUSION To our knowledge, this is the first study to automate pelvimetry in MRI volumes using deep learning. Our framework can measure the pelvic dimensions with high accuracy, enabling the extraction of metrics that facilitate a pre-operative difficulty assessment of the TME.
Collapse
Affiliation(s)
- Simon C Baltus
- Surgery Department, Meander Medical Centre, Maatweg, Amersfoort, 3818 TZ, Utrecht, The Netherlands.
- Robotics and Mechatronics, University of Twente, Drienerlolaan, Enschede, 5722 NB, Overijssel, The Netherlands.
| | - Ritch T J Geitenbeek
- Surgery Department, Meander Medical Centre, Maatweg, Amersfoort, 3818 TZ, Utrecht, The Netherlands
- Surgery Department, University Medical Center Groningen, Hanzeplein, Groningen, 9713 GZ, Groningen, The Netherlands
| | - Maike Frieben
- Surgery Department, University Medical Center Groningen, Hanzeplein, Groningen, 9713 GZ, Groningen, The Netherlands
| | - Elina Thibeau-Sutre
- Department of Applied Mathematics, Technical Medicine Center, University of Twente, Drienerlolaan, Enschede, 5722 NB, Overijssel, The Netherlands
| | - Jelmer M Wolterink
- Department of Applied Mathematics, Technical Medicine Center, University of Twente, Drienerlolaan, Enschede, 5722 NB, Overijssel, The Netherlands
| | - Can O Tan
- Robotics and Mechatronics, University of Twente, Drienerlolaan, Enschede, 5722 NB, Overijssel, The Netherlands
| | - Matthijs C Vermeulen
- Surgery Department, Meander Medical Centre, Maatweg, Amersfoort, 3818 TZ, Utrecht, The Netherlands
| | - Esther C J Consten
- Surgery Department, Meander Medical Centre, Maatweg, Amersfoort, 3818 TZ, Utrecht, The Netherlands
- Surgery Department, University Medical Center Groningen, Hanzeplein, Groningen, 9713 GZ, Groningen, The Netherlands
| | - Ivo A M J Broeders
- Surgery Department, Meander Medical Centre, Maatweg, Amersfoort, 3818 TZ, Utrecht, The Netherlands
- Robotics and Mechatronics, University of Twente, Drienerlolaan, Enschede, 5722 NB, Overijssel, The Netherlands
| |
Collapse
|
10
|
Ye K, Sun W, Tao R, Zheng G. A Projective-Geometry-Aware Network for 3D Vertebra Localization in Calibrated Biplanar X-Ray Images. SENSORS (BASEL, SWITZERLAND) 2025; 25:1123. [PMID: 40006352 PMCID: PMC11858964 DOI: 10.3390/s25041123] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/14/2025] [Revised: 02/07/2025] [Accepted: 02/12/2025] [Indexed: 02/27/2025]
Abstract
Current Deep Learning (DL)-based methods for vertebra localization in biplanar X-ray images mainly focus on two-dimensional (2D) information and neglect the projective geometry, limiting the accuracy of 3D navigation in X-ray-guided spine surgery. A 3D vertebra localization method from calibrated biplanar X-ray images is highly desired to address the problem. In this study, a projective-geometry-aware network for localizing 3D vertebrae in calibrated biplanar X-ray images, referred to as ProVLNet, is proposed. The network design of ProVLNet features three components: a Siamese 2D feature extractor to extract local appearance features from the biplanar X-ray images, a spatial alignment fusion module to incorporate the projective geometry in fusing the extracted 2D features in 3D space, and a 3D landmark regression module to regress the 3D coordinates of the vertebrae from the 3D fused features. Evaluated on two typical and challenging datasets acquired from the lumbar and the thoracic spine, ProVLNet achieved an identification rate of 99.53% and 98.98% and a point-to-point error of 0.64 mm and 1.38 mm, demonstrating superior performance of our proposed approach over the state-of-the-art (SOTA) methods.
Collapse
Affiliation(s)
| | | | | | - Guoyan Zheng
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China; (K.Y.); (W.S.); (R.T.)
| |
Collapse
|
11
|
Mao Y, Feng Q, Zhang Y, Ning Z. Semantics and instance interactive learning for labeling and segmentation of vertebrae in CT images. Med Image Anal 2025; 99:103380. [PMID: 39515182 DOI: 10.1016/j.media.2024.103380] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Revised: 10/17/2024] [Accepted: 10/22/2024] [Indexed: 11/16/2024]
Abstract
Automatically labeling and segmenting vertebrae in 3D CT images compose a complex multi-task problem. Current methods progressively conduct vertebra labeling and semantic segmentation, which typically include two separate models and may ignore feature interaction among different tasks. Although instance segmentation approaches with multi-channel prediction have been proposed to alleviate such issues, their utilization of semantic information remains insufficient. Additionally, another challenge for an accurate model is how to effectively distinguish similar adjacent vertebrae and model their sequential attribute. In this paper, we propose a Semantics and Instance Interactive Learning (SIIL) paradigm for synchronous labeling and segmentation of vertebrae in CT images. SIIL models semantic feature learning and instance feature learning, in which the former extracts spinal semantics and the latter distinguishes vertebral instances. Interactive learning involves semantic features to improve the separability of vertebral instances and instance features to help learn position and contour information, during which a Morphological Instance Localization Learning (MILL) module is introduced to align semantic and instance features and facilitate their interaction. Furthermore, an Ordinal Contrastive Prototype Learning (OCPL) module is devised to differentiate adjacent vertebrae with high similarity (via cross-image contrastive learning), and simultaneously model their sequential attribute (via a temporal unit). Extensive experiments on several datasets demonstrate that our method significantly outperforms other approaches in labeling and segmenting vertebrae. Our code is available at https://github.com/YuZhang-SMU/Vertebrae-Labeling-Segmentation.
Collapse
Affiliation(s)
- Yixiao Mao
- School of Biomedical Engineering, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China
| | - Yu Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| | - Zhenyuan Ning
- School of Biomedical Engineering, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, Guangdong, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou 510515, China.
| |
Collapse
|
12
|
Sun W, Zou X, Zheng G. A fully automatic fiducial detection and correspondence establishing method for online C-arm calibration. Int J Comput Assist Radiol Surg 2025; 20:43-55. [PMID: 38730187 DOI: 10.1007/s11548-024-03162-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2024] [Accepted: 04/22/2024] [Indexed: 05/12/2024]
Abstract
PURPOSE Online C-arm calibration with a mobile fiducial cage plays an essential role in various image-guided interventions. However, it is challenging to develop a fully automatic approach, which requires not only an accurate detection of fiducial projections but also a robust 2D-3D correspondence establishment. METHODS We propose a novel approach for online C-arm calibration with a mobile fiducial cage. Specifically, a novel mobile calibration cage embedded with 16 fiducials is designed, where the fiducials are arranged to form 4 line patterns with different cross-ratios. Then, an auto-context-based detection network (ADNet) is proposed to perform an accurate and robust detection of 2D projections of those fiducials in acquired C-arm images. Subsequently, we present a cross-ratio consistency-based 2D-3D correspondence establishing method to automatically match the detected 2D fiducial projections with those 3D fiducials, allowing for an accurate online C-arm calibration. RESULTS We designed and conducted comprehensive experiments to evaluate the proposed approach. For automatic detection of 2D fiducial projections, the proposed ADNet achieved a mean point-to-point distance of 0.65 ± 1.33 pixels. Additionally, the proposed C-arm calibration approach achieved a mean re-projection error of 1.01 ± 0.63 pixels and a mean point-to-line distance of 0.22 ± 0.12 mm. When the proposed C-arm calibration approach was applied to downstream tasks involving landmark and surface model reconstruction, sub-millimeter accuracy was achieved. CONCLUSION In summary, we developed a novel approach for online C-arm calibration. Both qualitative and quantitative results of comprehensive experiments demonstrated the accuracy and robustness of the proposed approach. Our approach holds potentials for various image-guided interventions.
Collapse
Affiliation(s)
- Wenyuan Sun
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, 800, Dongchuan Road, Shanghai, 200240, China
| | - Xiaoyang Zou
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, 800, Dongchuan Road, Shanghai, 200240, China
| | - Guoyan Zheng
- Institute of Medical Robotics, School of Biomedical Engineering, Shanghai Jiao Tong University, 800, Dongchuan Road, Shanghai, 200240, China.
| |
Collapse
|
13
|
Yang Y, Wang Y, Liu T, Wang M, Sun M, Song S, Fan W, Huang G. Anatomical prior-based vertebral landmark detection for spinal disorder diagnosis. Artif Intell Med 2025; 159:103011. [PMID: 39612522 DOI: 10.1016/j.artmed.2024.103011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 07/25/2024] [Accepted: 11/02/2024] [Indexed: 12/01/2024]
Abstract
As one of fundamental ways to interpret spine images, detection of vertebral landmarks is an informative prerequisite for further diagnosis and management of spine disorders such as scoliosis and fractures. Most existing machine learning-based methods for automatic vertebral landmark detection suffer from overlapping landmarks or abnormally long distances between nearby landmarks against anatomical priors, and thus lack sufficient reliability and interpretability. To tackle the problem, this paper systematically utilizes anatomical prior knowledge in vertebral landmark detection. We explicitly formulate anatomical priors of the spine, related to distances among vertebrae and spatial order within the spine, and integrate these geometrical constraints within training loss, inference procedure, and evaluation metrics. First, we introduce an anatomy-constraint loss to regularize the training process with the aforementioned contextual priors explicitly. Second, we propose a simple-yet-effective anatomy-aided inference procedure by employing sequential prediction rather than a parallel counterpart. Third, we provide novel anatomy-related metrics to quantitatively evaluate to which extent landmark predictions follow the anatomical priors, as is not reflected within the widely-used landmark localization error metric. We employ the localization framework on 1410 anterior-posterior radiographic images. Compared with competitive baseline models, we achieve superior landmark localization accuracy and comparable Cobb angle estimation for scoliosis assessment. Ablation studies demonstrate the effectiveness of designed components on the decrease of localization error and improvement of anatomical plausibility. Additionally, we exhibit effective generalization performance by transferring our detection method onto sagittal 2-D slices of CT scans and boost the performance of downstream compression fracture classification at vertebra-level.
Collapse
Affiliation(s)
- Yukang Yang
- Department of Automation, BNRist, Tsinghua University, Beijing, 100084, China.
| | - Yu Wang
- Department of Orthopaedics, Peking University First Hospital, Beijing, 100034, China.
| | - Tianyu Liu
- Department of Automation, BNRist, Tsinghua University, Beijing, 100084, China.
| | - Miao Wang
- Department of Orthopaedics, Aarhus University Hospital, Aarhus, 8200, Denmark.
| | - Ming Sun
- Department of Orthopaedics, Aarhus University Hospital, Aarhus, 8200, Denmark.
| | - Shiji Song
- Department of Automation, BNRist, Tsinghua University, Beijing, 100084, China.
| | - Wenhui Fan
- Department of Automation, BNRist, Tsinghua University, Beijing, 100084, China.
| | - Gao Huang
- Department of Automation, BNRist, Tsinghua University, Beijing, 100084, China; Beijing Academy of Artificial Intelligence, Beijing, 100084, China.
| |
Collapse
|
14
|
Quan Q, Yao Q, Zhu H, Kevin Zhou S. IGU-Aug: Information-Guided Unsupervised Augmentation and Pixel-Wise Contrastive Learning for Medical Image Analysis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:154-164. [PMID: 39088491 DOI: 10.1109/tmi.2024.3436713] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/03/2024]
Abstract
Contrastive learning (CL) is a form of self-supervised learning and has been widely used for various tasks. Different from widely studied instance-level contrastive learning, pixel-wise contrastive learning mainly helps with pixel-wise dense prediction tasks. The counterpart to an instance in instance-level CL is a pixel, along with its neighboring context, in pixel-wise CL. Aiming to build better feature representation, there is a vast literature about designing instance augmentation strategies for instance-level CL; but there is little similar work on pixel augmentation for pixel-wise CL with a pixel granularity. In this paper, we attempt to bridge this gap. We first classify a pixel into three categories, namely low-, medium-, and high-informative, based on the information quantity the pixel contains. We then adaptively design separate augmentation strategies for each category in terms of augmentation intensity and sampling ratio. Extensive experiments validate that our information-guided pixel augmentation strategy succeeds in encoding more discriminative representations and surpassing other competitive approaches in unsupervised local feature matching. Furthermore, our pretrained model improves the performance of both one-shot and fully supervised models. To the best of our knowledge, we are the first to propose a pixel augmentation method with a pixel granularity for enhancing unsupervised pixel-wise contrastive learning. Code is available at https://github.com/Curli-quan/IGU-Aug.
Collapse
|
15
|
Rashmi S, Srinath S, Deshmukh S, Prashanth S, Patil K. Cephalometric landmark annotation using transfer learning: Detectron2 and YOLOv8 baselines on a diverse cephalometric image dataset. Comput Biol Med 2024; 183:109318. [PMID: 39467377 DOI: 10.1016/j.compbiomed.2024.109318] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2024] [Revised: 10/19/2024] [Accepted: 10/21/2024] [Indexed: 10/30/2024]
Abstract
BACKGROUND Cephalometric landmark annotation is a key challenge in radiographic analysis, requiring automation due to its time-consuming process and inherent subjectivity. This study investigates the application of advanced transfer learning techniques to enhance the accuracy of anatomical landmarks in cephalometric images, which is a vital aspect of orthodontic diagnosis and treatment planning. METHODS We assess the suitability of transfer learning methods by employing state-of-the-art pose estimation models. The first framework is Detectron2, with two baselines featuring different ResNet backbone architectures: rcnn_R_50_FPN_3x and rcnn_R_101_FPN_3x. The second framework is YOLOv8, with three variants reflecting different network sizes: YOLOv8s-pose, YOLOv8m-pose, and YOLOv8l-pose. These pose estimation models are adopted for the landmark annotation task. The models are trained and evaluated on the DiverseCEPH19 dataset, comprising 1692 radiographic images with 19 landmarks, and their performance is analyzed across various images categories within the dataset. Additionally, the study is extended to a benchmark dataset of 400 images to investigate how dataset size impacts the performance of these frameworks. RESULTS Despite variations in objectives and evaluation metrics between pose estimation and landmark localization tasks, the results are promising. Detectron2's variant outperforms others with an accuracy of 85.89%, compared to 72.92% achieved by YOLOv8's variant on the DiverseCEPH19 dataset. This superior performance is also observed in the smaller benchmark dataset, where Detectron2 consistently maintains higher accuracy than YOLOv8. CONCLUSION The noted enhancements in annotation precision suggest the suitability of Detectron2 for deployment in applications that require high precision while taking into account factors such as model size, inference time, and resource utilization, the evidence favors YOLOv8 baselines.
Collapse
Affiliation(s)
- S Rashmi
- Dept. of Computer Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru, India.
| | - S Srinath
- Dept. of Computer Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru, India
| | - Seema Deshmukh
- Dept. of Pediatric & Preventive Dentistry, JSS Dental College & Hospital, JSS Academy of Higher Education & Research, Mysuru, India
| | - S Prashanth
- Dept. of Pediatric & Preventive Dentistry, JSS Dental College & Hospital, JSS Academy of Higher Education & Research, Mysuru, India
| | - Karthikeya Patil
- Dept. of Oral Medicine and Radiology, JSS Dental College & Hospital, JSS Academy of Higher Education & Research, Mysuru, India
| |
Collapse
|
16
|
Wu Y, Chen X, Dong F, He L, Cheng G, Zheng Y, Ma C, Yao H, Zhou S. Performance evaluation of a deep learning-based cascaded HRNet model for automatic measurement of X-ray imaging parameters of lumbar sagittal curvature. EUROPEAN SPINE JOURNAL : OFFICIAL PUBLICATION OF THE EUROPEAN SPINE SOCIETY, THE EUROPEAN SPINAL DEFORMITY SOCIETY, AND THE EUROPEAN SECTION OF THE CERVICAL SPINE RESEARCH SOCIETY 2024; 33:4104-4118. [PMID: 37787781 DOI: 10.1007/s00586-023-07937-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/03/2023] [Revised: 04/03/2023] [Accepted: 08/30/2023] [Indexed: 10/04/2023]
Abstract
PURPOSE To develop a deep learning-based cascaded HRNet model, in order to automatically measure X-ray imaging parameters of lumbar sagittal curvature and to evaluate its prediction performance. METHODS A total of 3730 lumbar lateral digital radiography (DR) images were collected from picture archiving and communication system (PACS). Among them, 3150 images were randomly selected as the training dataset and validation dataset, and 580 images as the test dataset. The landmarks of the lumbar curve index (LCI), lumbar lordosis angle (LLA), sacral slope (SS), lumbar lordosis index (LLI), and the posterior edge tangent angle of the vertebral body (PTA) were identified and marked. The measured results of landmarks on the test dataset were compared with the mean values of manual measurement as the reference standard. Percentage of correct key-points (PCK), intra-class correlation coefficient (ICC), Pearson correlation coefficient (r), mean absolute error (MAE), mean square error (MSE), root-mean-square error (RMSE), and Bland-Altman plot were used to evaluate the performance of the cascade HRNet model. RESULTS The PCK of the cascaded HRNet model was 97.9-100% in the 3 mm distance threshold. The mean differences between the reference standard and the predicted values for LCI, LLA, SS, LLI, and PTA were 0.43 mm, 0.99°, 1.11°, 0.01 mm, and 0.23°, respectively. There were strong correlation and consistency of the five parameters between the cascaded HRNet model and manual measurements (ICC = 0.989-0.999, R = 0.991-0.999, MAE = 0.63-1.65, MSE = 0.61-4.06, RMSE = 0.78-2.01). CONCLUSION The cascaded HRNet model based on deep learning algorithm could accurately identify the sagittal curvature-related landmarks on lateral lumbar DR images and automatically measure the relevant parameters, which is of great significance in clinical application.
Collapse
Affiliation(s)
- Yuhua Wu
- The First Clinical Medical College of Gansu University of Chinese Medicine, Lanzhou, 730000, Gansu, China
| | - Xiaofei Chen
- Department of Radiology, Gansu Provincial Hospital of Traditional Chinese Medicine (The first affiliated hospital of Gansu University of Traditional Chinese Medicine), Lanzhou, 730050, Gansu, China
| | - Fuwen Dong
- Department of Radiology, Gansu Provincial Hospital of Traditional Chinese Medicine (The first affiliated hospital of Gansu University of Traditional Chinese Medicine), Lanzhou, 730050, Gansu, China
| | - Linyang He
- Hangzhou Jianpei Technology Company Ltd, Hangzhou, 311200, Zhejiang, China
| | - Guohua Cheng
- Hangzhou Jianpei Technology Company Ltd, Hangzhou, 311200, Zhejiang, China
| | - Yuwen Zheng
- The First Clinical Medical College of Gansu University of Chinese Medicine, Lanzhou, 730000, Gansu, China
| | - Chunyu Ma
- The First Clinical Medical College of Gansu University of Chinese Medicine, Lanzhou, 730000, Gansu, China
| | - Hongyan Yao
- Department of Radiology, Gansu Provincial Hospital, No. 204, Donggang West Road, Lanzhou, 730000, Gansu, China
| | - Sheng Zhou
- Department of Radiology, Gansu Provincial Hospital, No. 204, Donggang West Road, Lanzhou, 730000, Gansu, China.
| |
Collapse
|
17
|
Chai Y, Boudali AM, Maes V, Walter WL. Clinical benchmark dataset for AI accuracy analysis: quantifying radiographic annotation of pelvic tilt. Sci Data 2024; 11:1162. [PMID: 39438488 PMCID: PMC11496730 DOI: 10.1038/s41597-024-04003-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Accepted: 10/14/2024] [Indexed: 10/25/2024] Open
Abstract
Radiographic landmark annotation determines patients' anatomical parameters and influences diagnoses. However, challenges arise from ambiguous region-based definitions, human error, and image quality variations, potentially compromising patient care. Additionally, AI landmark localization often presents its predictions in a probability-based heatmap format, which lacks a corresponding clinical standard for accuracy validation. This Data Descriptor presents a clinical benchmark dataset for pelvic tilt landmarks, gathered through a probabilistic approach to measure annotation accuracy within clinical environments. A retrospective analysis of 115 pelvic sagittal radiographs was conducted for annotating pelvic tilt parameters by five annotators, revealing landmark cloud sizes of 6.04 mm-17.90 mm at a 95% dataset threshold, corresponding to 9.51°-16.55° maximum angular disagreement in clinical settings. The outcome provides a quantified point cloud dataset for each landmark corresponding to different probabilities, which enables assessment of directional annotation distribution and parameter-wise impact, providing clinical benchmarks. The data is readily reusable for AI studies analyzing the same landmarks, and the method can be easily replicated for establishing clinical accuracy benchmarks of other landmarks.
Collapse
Affiliation(s)
- Yuan Chai
- Sydney Musculoskeletal Health and The Kolling Institute, Northern Clinical School, Faculty of Medicine and Health and the Northern Sydney Local Health District, Sydney, NSW, Australia.
| | - A Mounir Boudali
- Sydney Musculoskeletal Health and The Kolling Institute, Northern Clinical School, Faculty of Medicine and Health and the Northern Sydney Local Health District, Sydney, NSW, Australia
| | - Vincent Maes
- University Hospitals Leuven, Department of Orthopedic Surgery, Leuven, Belgium
- Department of Orthopedics and Traumatic Surgery, Royal North Shore Hospital, St Leonard's, NSW, Australia
| | - William L Walter
- Sydney Musculoskeletal Health and The Kolling Institute, Northern Clinical School, Faculty of Medicine and Health and the Northern Sydney Local Health District, Sydney, NSW, Australia
- Department of Orthopedics and Traumatic Surgery, Royal North Shore Hospital, St Leonard's, NSW, Australia
- The Orthopaedic Department, St Vincent's Hospital, Darlinghurst, NSW, Australia
| |
Collapse
|
18
|
Joham SJ, Hadzic A, Urschler M. Implicit Is Not Enough: Explicitly Enforcing Anatomical Priors inside Landmark Localization Models. Bioengineering (Basel) 2024; 11:932. [PMID: 39329674 PMCID: PMC11428392 DOI: 10.3390/bioengineering11090932] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2024] [Revised: 09/12/2024] [Accepted: 09/13/2024] [Indexed: 09/28/2024] Open
Abstract
The task of localizing distinct anatomical structures in medical image data is an essential prerequisite for several medical applications, such as treatment planning in orthodontics, bone-age estimation, or initialization of segmentation methods in automated image analysis tools. Currently, Anatomical Landmark Localization (ALL) is mainly solved by deep-learning methods, which cannot guarantee robust ALL predictions; there may always be outlier predictions that are far from their ground truth locations due to out-of-distribution inputs. However, these localization outliers are detrimental to the performance of subsequent medical applications that rely on ALL results. The current ALL literature relies heavily on implicit anatomical constraints built into the loss function and network architecture to reduce the risk of anatomically infeasible predictions. However, we argue that in medical imaging, where images are generally acquired in a controlled environment, we should use stronger explicit anatomical constraints to reduce the number of outliers as much as possible. Therefore, we propose the end-to-end trainable Global Anatomical Feasibility Filter and Analysis (GAFFA) method, which uses prior anatomical knowledge estimated from data to explicitly enforce anatomical constraints. GAFFA refines the initial localization results of a U-Net by approximately solving a Markov Random Field (MRF) with a single iteration of the sum-product algorithm in a differentiable manner. Our experiments demonstrate that GAFFA outperforms all other landmark refinement methods investigated in our framework. Moreover, we show that GAFFA is more robust to large outliers than state-of-the-art methods on the studied X-ray hand dataset. We further motivate this claim by visualizing the anatomical constraints used in GAFFA as spatial energy heatmaps, which allowed us to find an annotation error in the hand dataset not previously discussed in the literature.
Collapse
Affiliation(s)
- Simon Johannes Joham
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria
- Institute of Computer Graphics and Vision, Graz University of Technology, 8010 Graz, Austria
| | - Arnela Hadzic
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria
| | - Martin Urschler
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria
- BioTechMed-Graz, 8010 Graz, Austria
| |
Collapse
|
19
|
Jin H, Che H, Chen H. Rethinking Self-Training for Semi-Supervised Landmark Detection: A Selection-Free Approach. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:4952-4965. [PMID: 39236118 DOI: 10.1109/tip.2024.3451937] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/07/2024]
Abstract
Self-training is a simple yet effective method for semi-supervised learning, during which pseudo-label selection plays an important role for handling confirmation bias. Despite its popularity, applying self-training to landmark detection faces three problems: 1) The selected confident pseudo-labels often contain data bias, which may hurt model performance; 2) It is not easy to decide a proper threshold for sample selection as the localization task can be sensitive to noisy pseudo-labels; 3) coordinate regression does not output confidence, making selection-based self-training infeasible. To address the above issues, we propose Self-Training for Landmark Detection (STLD), a method that does not require explicit pseudo-label selection. Instead, STLD constructs a task curriculum to deal with confirmation bias, which progressively transitions from more confident to less confident tasks over the rounds of self-training. Pseudo pretraining and shrink regression are two essential components for such a curriculum, where the former is the first task of the curriculum for providing a better model initialization and the latter is further added in the later rounds to directly leverage the pseudo-labels in a coarse-to-fine manner. Experiments on three facial and one medical landmark detection benchmark show that STLD outperforms the existing methods consistently in both semi- and omni-supervised settings. The code is available at https://github.com/jhb86253817/STLD.
Collapse
|
20
|
Sackl M, Tinauer C, Urschler M, Enzinger C, Stollberger R, Ropele S. Fully Automated Hippocampus Segmentation using T2-informed Deep Convolutional Neural Networks. Neuroimage 2024; 298:120767. [PMID: 39103064 DOI: 10.1016/j.neuroimage.2024.120767] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2023] [Revised: 07/26/2024] [Accepted: 07/31/2024] [Indexed: 08/07/2024] Open
Abstract
Hippocampal atrophy (tissue loss) has become a fundamental outcome parameter in clinical trials on Alzheimer's disease. To accurately estimate hippocampus volume and track its volume loss, a robust and reliable segmentation is essential. Manual hippocampus segmentation is considered the gold standard but is extensive, time-consuming, and prone to rater bias. Therefore, it is often replaced by automated programs like FreeSurfer, one of the most commonly used tools in clinical research. Recently, deep learning-based methods have also been successfully applied to hippocampus segmentation. The basis of all approaches are clinically used T1-weighted whole-brain MR images with approximately 1 mm isotropic resolution. However, such T1 images show low contrast-to-noise ratios (CNRs), particularly for many hippocampal substructures, limiting delineation reliability. To overcome these limitations, high-resolution T2-weighted scans are suggested for better visualization and delineation, as they show higher CNRs and usually allow for higher resolutions. Unfortunately, such time-consuming T2-weighted sequences are not feasible in a clinical routine. We propose an automated hippocampus segmentation pipeline leveraging deep learning with T2-weighted MR images for enhanced hippocampus segmentation of clinical T1-weighted images based on a series of 3D convolutional neural networks and a specifically acquired multi-contrast dataset. This dataset consists of corresponding pairs of T1- and high-resolution T2-weighted images, with the T2 images only used to create more accurate manual ground truth annotations and to train the segmentation network. The T2-based ground truth labels were also used to evaluate all experiments by comparing the masks visually and by various quantitative measures. We compared our approach with four established state-of-the-art hippocampus segmentation algorithms (FreeSurfer, ASHS, HippoDeep, HippMapp3r) and demonstrated a superior segmentation performance. Moreover, we found that the automated segmentation of T1-weighted images benefits from the T2-based ground truth data. In conclusion, this work showed the beneficial use of high-resolution, T2-based ground truth data for training an automated, deep learning-based hippocampus segmentation and provides the basis for a reliable estimation of hippocampal atrophy in clinical studies.
Collapse
Affiliation(s)
- Maximilian Sackl
- Department of Neurology, Medical University of Graz, Austria; BioTechMed-Graz, Austria
| | | | - Martin Urschler
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, Austria; BioTechMed-Graz, Austria
| | | | - Rudolf Stollberger
- Institute of Biomedical Imaging, Graz University of Technology, Austria; BioTechMed-Graz, Austria
| | - Stefan Ropele
- Department of Neurology, Medical University of Graz, Austria; BioTechMed-Graz, Austria.
| |
Collapse
|
21
|
Barbosa RM, Serrador L, da Silva MV, Macedo CS, Santos CP. Knee landmarks detection via deep learning for automatic imaging evaluation of trochlear dysplasia and patellar height. Eur Radiol 2024; 34:5736-5747. [PMID: 38337072 PMCID: PMC11364617 DOI: 10.1007/s00330-024-10596-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2023] [Revised: 12/13/2023] [Accepted: 12/19/2023] [Indexed: 02/12/2024]
Abstract
OBJECTIVES To develop and validate a deep learning-based approach to automatically measure the patellofemoral instability (PFI) indices related to patellar height and trochlear dysplasia in knee magnetic resonance imaging (MRI) scans. METHODS A total of 763 knee MRI slices from 95 patients were included in the study, and 3393 anatomical landmarks were annotated for measuring sulcus angle (SA), trochlear facet asymmetry (TFA), trochlear groove depth (TGD) and lateral trochlear inclination (LTI) to assess trochlear dysplasia, and Insall-Salvati index (ISI), modified Insall-Salvati index (MISI), Caton Deschamps index (CDI) and patellotrochlear index (PTI) to assess patellar height. A U-Net based network was implemented to predict the landmarks' locations. The successful detection rate (SDR) and the mean absolute error (MAE) evaluation metrics were used to evaluate the performance of the network. The intraclass correlation coefficient (ICC) was also used to evaluate the reliability of the proposed framework to measure the mentioned PFI indices. RESULTS The developed models achieved good accuracy in predicting the landmarks' locations, with a maximum value for the MAE of 1.38 ± 0.76 mm. The results show that LTI, TGD, ISI, CDI and PTI can be measured with excellent reliability (ICC > 0.9), and SA, TFA and MISI can be measured with good reliability (ICC > 0.75), with the proposed framework. CONCLUSIONS This study proposes a reliable approach with promising applicability for automatic patellar height and trochlear dysplasia assessment, assisting the radiologists in their clinical practice. CLINICAL RELEVANCE STATEMENT The objective knee landmarks detection on MRI images provided by artificial intelligence may improve the reproducibility and reliability of the imaging evaluation of trochlear anatomy and patellar height, assisting radiologists in their clinical practice in the patellofemoral instability assessment. KEY POINTS • Imaging evaluation of patellofemoral instability is subjective and vulnerable to substantial intra and interobserver variability. • Patellar height and trochlear dysplasia are reliably assessed in MRI by means of artificial intelligence (AI). • The developed AI framework provides an objective evaluation of patellar height and trochlear dysplasia enhancing the clinical practice of the radiologists.
Collapse
Affiliation(s)
- Roberto M Barbosa
- Center of MicroElectroMechanical Systems (CMEMS), University of Minho, Guimarães, Portugal.
- MIT Portugal Program, School of Engineering, University of Minho, Guimarães, Portugal.
| | - Luís Serrador
- Center of MicroElectroMechanical Systems (CMEMS), University of Minho, Guimarães, Portugal
| | | | | | - Cristina P Santos
- Center of MicroElectroMechanical Systems (CMEMS), University of Minho, Guimarães, Portugal
- LABBELS - Associate Laboratory, Braga/Guimarães, Portugal
| |
Collapse
|
22
|
Arnold R, Prassl AJ, Neic A, Thaler F, Augustin CM, Gsell MAF, Gillette K, Manninger M, Scherr D, Plank G. pyCEPS: A cross-platform electroanatomic mapping data to computational model conversion platform for the calibration of digital twin models of cardiac electrophysiology. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 254:108299. [PMID: 38959599 DOI: 10.1016/j.cmpb.2024.108299] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/18/2024] [Revised: 06/10/2024] [Accepted: 06/19/2024] [Indexed: 07/05/2024]
Abstract
BACKGROUND AND OBJECTIVE Data from electro-anatomical mapping (EAM) systems are playing an increasingly important role in computational modeling studies for the patient-specific calibration of digital twin models. However, data exported from commercial EAM systems are challenging to access and parse. Converting to data formats that are easily amenable to be viewed and analyzed with commonly used cardiac simulation software tools such as openCARP remains challenging. We therefore developed an open-source platform, pyCEPS, for parsing and converting clinical EAM data conveniently to standard formats widely adopted within the cardiac modeling community. METHODS AND RESULTS pyCEPS is an open-source Python-based platform providing the following functions: (i) access and interrogate the EAM data exported from clinical mapping systems; (ii) efficient browsing of EAM data to preview mapping procedures, electrograms (EGMs), and electro-cardiograms (ECGs); (iii) conversion to modeling formats according to the openCARP standard, to be amenable to analysis with standard tools and advanced workflows as used for in silico EAM data. Documentation and training material to facilitate access to this complementary research tool for new users is provided. We describe the technological underpinnings and demonstrate the capabilities of pyCEPS first, and showcase its use in an exemplary modeling application where we use clinical imaging data to build a patient-specific anatomical model. CONCLUSION With pyCEPS we offer an open-source framework for accessing EAM data, and converting these to cardiac modeling standard formats. pyCEPS provides the core functionality needed to integrate EAM data in cardiac modeling research. We detail how pyCEPS could be integrated into model calibration workflows facilitating the calibration of a computational model based on EAM data.
Collapse
Affiliation(s)
- Robert Arnold
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria
| | - Anton J Prassl
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria
| | | | - Franz Thaler
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria; Institute of Computer Graphics and Vision, Graz University of Technology, Graz, Austria
| | - Christoph M Augustin
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria
| | - Matthias A F Gsell
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria
| | - Karli Gillette
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria; BioTechMed-Graz, Graz, Austria
| | - Martin Manninger
- Division of Cardiology, Department of Medicine, Medical University of Graz, Graz, Austria
| | - Daniel Scherr
- Division of Cardiology, Department of Medicine, Medical University of Graz, Graz, Austria
| | - Gernot Plank
- Gottfried Schatz Research Center, Division of Medical Physics and Biophysics, Medical University of Graz, Graz, Austria; NumeriCor GmbH, Graz, Austria; BioTechMed-Graz, Graz, Austria.
| |
Collapse
|
23
|
Quan Q, Yao Q, Zhu H, Wang Q, Zhou SK. Which images to label for few-shot medical image analysis? Med Image Anal 2024; 96:103200. [PMID: 38801797 DOI: 10.1016/j.media.2024.103200] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2023] [Revised: 03/26/2024] [Accepted: 05/06/2024] [Indexed: 05/29/2024]
Abstract
The success of deep learning methodologies hinges upon the availability of meticulously labeled extensive datasets. However, when dealing with medical images, the annotation process for such abundant training data often necessitates the involvement of experienced radiologists, thereby consuming their limited time resources. In order to alleviate this burden, few-shot learning approaches have been developed, which manage to achieve competitive performance levels with only several labeled images. Nevertheless, a crucial yet previously overlooked problem in few-shot learning is about the selection of template images for annotation before learning, which affects the final performance. In this study, we propose a novel TEmplate Choosing Policy (TECP) that aims to identify and select "the most worthy" images for annotation, particularly within the context of multiple few-shot medical tasks, including landmark detection, anatomy detection, and anatomy segmentation. TECP is composed of four integral components: (1) Self-supervised training, which entails training a pre-existing deep model to extract salient features from radiological images; (2) Alternative proposals for localizing informative regions within the images; and (3) Representative Score Estimation, which involves the evaluation and identification of the most representative samples or templates. (4) Ranking, which rank all candidates and select one with highest representative score. The efficacy of the TECP approach is demonstrated through a series of comprehensive experiments conducted on multiple public datasets. Across all three medical tasks, the utilization of TECP yields noticeable improvements in model performance.
Collapse
Affiliation(s)
- Quan Quan
- Institute of Computing Technology, Chinese Academy of Sciences (CAS), Beijing, 100080, China; University of Chinese Academy of Sciences (UCAS), Beijing, 101408, China
| | - Qingsong Yao
- Institute of Computing Technology, Chinese Academy of Sciences (CAS), Beijing, 100080, China; University of Chinese Academy of Sciences (UCAS), Beijing, 101408, China
| | - Heqin Zhu
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, China
| | - Qiyuan Wang
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, China
| | - S Kevin Zhou
- Institute of Computing Technology, Chinese Academy of Sciences (CAS), Beijing, 100080, China; School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, China; Center for Medical Imaging, Robotics, Analytic Computing; Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215000, China; Key Laboratory of Precision and Intelligent Chemistry, USTC, Hefei, 230026, China.
| |
Collapse
|
24
|
Zhang Q, Zhao F, Zhang Y, Huang M, Gong X, Deng X. Automated measurement of lumbar pedicle screw parameters using deep learning algorithm on preoperative CT scans. J Bone Oncol 2024; 47:100627. [PMID: 39188420 PMCID: PMC11345936 DOI: 10.1016/j.jbo.2024.100627] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Revised: 07/13/2024] [Accepted: 07/22/2024] [Indexed: 08/28/2024] Open
Abstract
Purpose This study aims to devise and assess an automated measurement framework for lumbar pedicle screw parameters leveraging preoperative computed tomography (CT) scans and a deep learning algorithm. Methods A deep learning model was constructed employing a dataset comprising 1410 axial preoperative CT images of lumbar pedicles sourced from 282 patients. The model was trained to predict several screw parameters, including the axial angle and width of pedicles, the length of pedicle screw paths, and the interpedicular distance. The mean values of these parameters, as determined by two radiologists and one spinal surgeon, served as the reference standard. Results The deep learning model achieved high agreement with the reference standard for the axial angle of the left pedicle (ICC = 0.92) and right pedicle (ICC = 0.93), as well as for the length of the left pedicle screw path (ICC = 0.82) and right pedicle (ICC = 0.87). Similarly, high agreement was observed for pedicle width (left ICC = 0.97, right ICC = 0.98) and interpedicular distance (ICC = 0.91). Overall, the model's performance paralleled that of manual determination of lumbar pedicle screw parameters. Conclusion The developed deep learning-based model demonstrates proficiency in accurately identifying landmarks on preoperative CT scans and autonomously generating parameters relevant to lumbar pedicle screw placement. These findings suggest its potential to offer efficient and precise measurements for clinical applications.
Collapse
Affiliation(s)
- Qian Zhang
- Department of Radiology, The 901st Hospital of the Joint Logistics Support Force of PLA, Hefei 230031, China
- Soochow University, Soochow 215000, China
- Department of Radiology, Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital), Hangzhou Medical College, Hangzhou 310014, China
| | - Fanfan Zhao
- Department of Radiology, Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital), Hangzhou Medical College, Hangzhou 310014, China
| | - Yu Zhang
- Department of Radiology, The 901st Hospital of the Joint Logistics Support Force of PLA, Hefei 230031, China
| | - Man Huang
- Department of Radiology, The 901st Hospital of the Joint Logistics Support Force of PLA, Hefei 230031, China
| | - Xiangyang Gong
- Soochow University, Soochow 215000, China
- Department of Radiology, Zhejiang Provincial People’s Hospital (Affiliated People’s Hospital), Hangzhou Medical College, Hangzhou 310014, China
| | - Xuefei Deng
- Department of Anatomy, Anhui Medical University, Hefei 230032, China
| |
Collapse
|
25
|
Tao L, Zhang X, Yang Y, Cheng M, Zhang R, Qian H, Wen Y, Yu H. Craniomaxillofacial landmarks detection in CT scans with limited labeled data via semi-supervised learning. Heliyon 2024; 10:e34583. [PMID: 39130473 PMCID: PMC11315087 DOI: 10.1016/j.heliyon.2024.e34583] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Revised: 05/21/2024] [Accepted: 07/11/2024] [Indexed: 08/13/2024] Open
Abstract
Background Three-dimensional cephalometric analysis is crucial in craniomaxillofacial assessment, with landmarks detection in craniomaxillofacial (CMF) CT scans being a key component. However, creating robust deep learning models for this task typically requires extensive CMF CT datasets annotated by experienced medical professionals, a process that is time-consuming and labor-intensive. Conversely, acquiring large volume of unlabeled CMF CT data is relatively straightforward. Thus, semi-supervised learning (SSL), leveraging limited labeled data supplemented by sufficient unlabeled dataset, could be a viable solution to this challenge. Method We developed an SSL model, named CephaloMatch, based on a strong-weak perturbation consistency framework. The proposed SSL model incorporates a head position rectification technique through coarse detection to enhance consistency between labeled and unlabeled datasets and a multilayers perturbation method which is employed to expand the perturbation space. The proposed SSL model was assessed using 362 CMF CT scans, divided into a training set (60 scans), a validation set (14 scans), and an unlabeled set (288 scans). Result The proposed SSL model attained a detection error of 1.60 ± 0.87 mm, significantly surpassing the performance of conventional fully supervised learning model (1.94 ± 1.12 mm). Notably, the proposed SSL model achieved equivalent detection accuracy (1.91 ± 1.00 mm) with only half the labeled dataset, compared to the fully supervised learning model. Conclusions The proposed SSL model demonstrated exceptional performance in landmarks detection using a limited labeled CMF CT dataset, significantly reducing the workload of medical professionals and enhances the accuracy of 3D cephalometric analysis.
Collapse
Affiliation(s)
- Leran Tao
- Department of Oral and Cranio-Maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| | - Xu Zhang
- Mechanical College, Shanghai Dianji University, Shanghai, 201306, China
| | - Yang Yang
- Shanghai Lanhui Medical Technology Co., Ltd, Shanghai, 200333, China
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200030, China
| | - Mengjia Cheng
- Department of Oral and Cranio-Maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| | - Rongbin Zhang
- College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200125, China
| | | | - Yaofeng Wen
- Shanghai Lanhui Medical Technology Co., Ltd, Shanghai, 200333, China
| | - Hongbo Yu
- Department of Oral and Cranio-Maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| |
Collapse
|
26
|
Huang Z, Zhao R, Leung FHF, Banerjee S, Lam KM, Zheng YP, Ling SH. Landmark Localization From Medical Images With Generative Distribution Prior. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2679-2692. [PMID: 38421850 DOI: 10.1109/tmi.2024.3371948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/02/2024]
Abstract
In medical image analysis, anatomical landmarks usually contain strong prior knowledge of their structural information. In this paper, we propose to promote medical landmark localization by modeling the underlying landmark distribution via normalizing flows. Specifically, we introduce the flow-based landmark distribution prior as a learnable objective function into a regression-based landmark localization framework. Moreover, we employ an integral operation to make the mapping from heatmaps to coordinates differentiable to further enhance heatmap-based localization with the learned distribution prior. Our proposed Normalizing Flow-based Distribution Prior (NFDP) employs a straightforward backbone and non-problem-tailored architecture (i.e., ResNet18), which delivers high-fidelity outputs across three X-ray-based landmark localization datasets. Remarkably, the proposed NFDP can do the job with minimal additional computational burden as the normalizing flows module is detached from the framework on inferencing. As compared to existing techniques, our proposed NFDP provides a superior balance between prediction accuracy and inference speed, making it a highly efficient and effective approach. The source code of this paper is available at https://github.com/jacksonhzx95/NFDP.
Collapse
|
27
|
Wang H, Jin Q, Li S, Liu S, Wang M, Song Z. A comprehensive survey on deep active learning in medical image analysis. Med Image Anal 2024; 95:103201. [PMID: 38776841 DOI: 10.1016/j.media.2024.103201] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/20/2023] [Revised: 04/25/2024] [Accepted: 05/06/2024] [Indexed: 05/25/2024]
Abstract
Deep learning has achieved widespread success in medical image analysis, leading to an increasing demand for large-scale expert-annotated medical image datasets. Yet, the high cost of annotating medical images severely hampers the development of deep learning in this field. To reduce annotation costs, active learning aims to select the most informative samples for annotation and train high-performance models with as few labeled samples as possible. In this survey, we review the core methods of active learning, including the evaluation of informativeness and sampling strategy. For the first time, we provide a detailed summary of the integration of active learning with other label-efficient techniques, such as semi-supervised, self-supervised learning, and so on. We also summarize active learning works that are specifically tailored to medical image analysis. Additionally, we conduct a thorough comparative analysis of the performance of different AL methods in medical image analysis with experiments. In the end, we offer our perspectives on the future trends and challenges of active learning and its applications in medical image analysis. An accompanying paper list and code for the comparative analysis is available in https://github.com/LightersWang/Awesome-Active-Learning-for-Medical-Image-Analysis.
Collapse
Affiliation(s)
- Haoran Wang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Qiuye Jin
- Computational Bioscience Research Center (CBRC), King Abdullah University of Science and Technology (KAUST), Thuwal 23955, Saudi Arabia
| | - Shiman Li
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Siyu Liu
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China
| | - Manning Wang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China.
| | - Zhijian Song
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai 200032, China; Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Shanghai 200032, China.
| |
Collapse
|
28
|
Choi E, Park D, Son G, Bak S, Eo T, Youn D, Hwang D. Weakly supervised deep learning for diagnosis of multiple vertebral compression fractures in CT. Eur Radiol 2024; 34:3750-3760. [PMID: 37973631 DOI: 10.1007/s00330-023-10394-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2022] [Revised: 08/08/2023] [Accepted: 09/11/2023] [Indexed: 11/19/2023]
Abstract
OBJECTIVE This study aims to develop a weakly supervised deep learning (DL) model for vertebral-level vertebral compression fracture (VCF) classification using image-level labelled data. METHODS The training set included 815 patients with normal (n = 507, 62%) or VCFs (n = 308, 38%). Our proposed model was trained on image-level labelled data for vertebral-level classification. Another supervised DL model was trained with vertebral-level labelled data to compare the performance of the proposed model. RESULTS The test set included 227 patients with normal (n = 117, 52%) or VCFs (n = 110, 48%). For a fair comparison of the two models, we compared sensitivities with the same specificities of the proposed model and the vertebral-level supervised model. The specificity for overall L1-L5 performance was 0.981. The proposed model may outperform the vertebral-level supervised model with sensitivities of 0.770 vs 0.705 (p = 0.080), respectively. For vertebral-level analysis, the specificities for each L1-L5 were 0.974, 0.973, 0.970, 0.991, and 0.995, respectively. The proposed model yielded the same or better sensitivity than the vertebral-level supervised model in L1 (0.750 vs 0.694, p = 0.480), L3 (0.793 vs 0.586, p < 0.05), L4 (0.833 vs 0.667, p = 0.480), and L5 (0.600 vs 0.600, p = 1.000), respectively. The proposed model showed lower sensitivity than the vertebral-level supervised model for L2, but there was no significant difference (0.775 vs 0.825, p = 0.617). CONCLUSIONS The proposed model may have a comparable or better performance than the supervised model in vertebral-level VCF classification. CLINICAL RELEVANCE STATEMENT Vertebral-level vertebral compression fracture classification aids in devising patient-specific treatment plans by identifying the precise vertebrae affected by compression fractures. KEY POINTS • Our proposed weakly supervised method may have comparable or better performance than the supervised method for vertebral-level vertebral compression fracture classification. • The weakly supervised model could have classified cases with multiple vertebral compression fractures at the vertebral-level, even if the model was trained with image-level labels. • Our proposed method could help reduce radiologists' labour because it enables vertebral-level classification from image-level labels.
Collapse
Affiliation(s)
- Euijoon Choi
- Department of Artificial Intelligence, Yonsei University, Seoul, Republic of Korea
| | - Doohyun Park
- School of Electrical and Electronic Engineering, Yonsei University, Seoul, Republic of Korea
| | - Geonhui Son
- School of Electrical and Electronic Engineering, Yonsei University, Seoul, Republic of Korea
| | | | - Taejoon Eo
- School of Electrical and Electronic Engineering, Yonsei University, Seoul, Republic of Korea
| | - Daemyung Youn
- School of Management of Technology, Yonsei University, Seoul, Republic of Korea
| | - Dosik Hwang
- School of Electrical and Electronic Engineering, Yonsei University, Seoul, Republic of Korea.
- Center for Healthcare Robotics, Korea Institute of Science and Technology, 5, Hwarang-Ro 14-Gil, Seongbuk-Gu, Seoul, 02792, Republic of Korea.
- Department of Oral and Maxillofacial Radiology, Yonsei University College of Dentistry, Seoul, Republic of Korea.
- Department of Radiology and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, Seoul, Republic of Korea.
| |
Collapse
|
29
|
Tan Z, Feng J, Lu W, Yin Y, Yang G, Zhou J. Multi-task global optimization-based method for vascular landmark detection. Comput Med Imaging Graph 2024; 114:102364. [PMID: 38432060 DOI: 10.1016/j.compmedimag.2024.102364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2023] [Revised: 12/04/2023] [Accepted: 02/22/2024] [Indexed: 03/05/2024]
Abstract
Vascular landmark detection plays an important role in medical analysis and clinical treatment. However, due to the complex topology and similar local appearance around landmarks, the popular heatmap regression based methods always suffer from the landmark confusion problem. Vascular landmarks are connected by vascular segments and have special spatial correlations, which can be utilized for performance improvement. In this paper, we propose a multi-task global optimization-based framework for accurate and automatic vascular landmark detection. A multi-task deep learning network is exploited to accomplish landmark heatmap regression, vascular semantic segmentation, and orientation field regression simultaneously. The two auxiliary objectives are highly correlated with the heatmap regression task and help the network incorporate the structural prior knowledge. During inference, instead of performing a max-voting strategy, we propose a global optimization-based post-processing method for final landmark decision. The spatial relationships between neighboring landmarks are utilized explicitly to tackle the landmark confusion problem. We evaluated our method on a cerebral MRA dataset with 564 volumes, a cerebral CTA dataset with 510 volumes, and an aorta CTA dataset with 50 volumes. The experiments demonstrate that the proposed method is effective for vascular landmark localization and achieves state-of-the-art performance.
Collapse
Affiliation(s)
- Zimeng Tan
- Department of Automation, Tsinghua University, Beijing, China
| | - Jianjiang Feng
- Department of Automation, Tsinghua University, Beijing, China.
| | - Wangsheng Lu
- UnionStrong (Beijing) Technology Co.Ltd, Beijing, China
| | - Yin Yin
- UnionStrong (Beijing) Technology Co.Ltd, Beijing, China
| | | | - Jie Zhou
- Department of Automation, Tsinghua University, Beijing, China
| |
Collapse
|
30
|
Surendran A, Daigavane P, Shrivastav S, Kamble R, Sanchla AD, Bharti L, Shinde M. The Future of Orthodontics: Deep Learning Technologies. Cureus 2024; 16:e62045. [PMID: 38989357 PMCID: PMC11234326 DOI: 10.7759/cureus.62045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2024] [Accepted: 06/09/2024] [Indexed: 07/12/2024] Open
Abstract
Deep learning has emerged as a revolutionary technical advancement in modern orthodontics, offering novel methods for diagnosis, treatment planning, and outcome prediction. Over the past 25 years, the field of dentistry has widely adopted information technology (IT), resulting in several benefits, including decreased expenses, increased efficiency, decreased need for human expertise, and reduced errors. The transition from preset rules to learning from real-world examples, particularly machine learning (ML) and artificial intelligence (AI), has greatly benefited the organization, analysis, and storage of medical data. Deep learning, a type of AI, enables robots to mimic human neural networks, allowing them to learn and make decisions independently without the need for explicit programming. Its ability to automate cephalometric analysis and enhance diagnosis through 3D imaging has revolutionized orthodontic operations. Deep learning models have the potential to significantly improve treatment outcomes and reduce human errors by accurately identifying anatomical characteristics on radiographs, thereby expediting analytical processes. Additionally, the use of 3D imaging technologies such as cone-beam computed tomography (CBCT) can facilitate precise treatment planning, allowing for comprehensive examinations of craniofacial architecture, tooth movements, and airway dimensions. In today's era of personalized medicine, deep learning's ability to customize treatments for individual patients has propelled the field of orthodontics forward tremendously. However, it is essential to address issues related to data privacy, model interpretability, and ethical considerations before orthodontic practices can use deep learning in an ethical and responsible manner. Modern orthodontics is evolving, thanks to the ability of deep learning to deliver more accurate, effective, and personalized orthodontic treatments, improving patient care as technology develops.
Collapse
Affiliation(s)
- Aathira Surendran
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Pallavi Daigavane
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Sunita Shrivastav
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Ranjit Kamble
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Abhishek D Sanchla
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Lovely Bharti
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| | - Mrudula Shinde
- Department of Orthodontics & Dentofacial Orthopaedics, Sharad Pawar Dental College & Hospital, Wardha, IND
| |
Collapse
|
31
|
Serrador L, Villani FP, Moccia S, Santos CP. Knowledge distillation on individual vertebrae segmentation exploiting 3D U-Net. Comput Med Imaging Graph 2024; 113:102350. [PMID: 38340574 DOI: 10.1016/j.compmedimag.2024.102350] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Revised: 02/01/2024] [Accepted: 02/01/2024] [Indexed: 02/12/2024]
Abstract
Recent advances in medical imaging have highlighted the critical development of algorithms for individual vertebral segmentation on computed tomography (CT) scans. Essential for diagnostic accuracy and treatment planning in orthopaedics, neurosurgery and oncology, these algorithms face challenges in clinical implementation, including integration into healthcare systems. Consequently, our focus lies in exploring the application of knowledge distillation (KD) methods to train shallower networks capable of efficiently segmenting vertebrae in CT scans. This approach aims to reduce segmentation time, enhance suitability for emergency cases, and optimize computational and memory resource efficiency. Building upon prior research in the field, a two-step segmentation approach was employed. Firstly, the spine's location was determined by predicting a heatmap, indicating the probability of each voxel belonging to the spine. Subsequently, an iterative segmentation of vertebrae was performed from the top to the bottom of the CT volume over the located spine, using a memory instance to record the already segmented vertebrae. KD methods were implemented by training a teacher network with performance similar to that found in the literature, and this knowledge was distilled to a shallower network (student). Two KD methods were applied: (1) using the soft outputs of both networks and (2) matching logits. Two publicly available datasets, comprising 319 CT scans from 300 patients and a total of 611 cervical, 2387 thoracic, and 1507 lumbar vertebrae, were used. To ensure dataset balance and robustness, effective data augmentation methods were applied, including cleaning the memory instance to replicate the first vertebra segmentation. The teacher network achieved an average Dice similarity coefficient (DSC) of 88.22% and a Hausdorff distance (HD) of 7.71 mm, showcasing performance similar to other approaches in the literature. Through knowledge distillation from the teacher network, the student network's performance improved, with an average DSC increasing from 75.78% to 84.70% and an HD decreasing from 15.17 mm to 8.08 mm. Compared to other methods, our teacher network exhibited up to 99.09% fewer parameters, 90.02% faster inference time, 88.46% shorter total segmentation time, and 89.36% less associated carbon (CO2) emission rate. Regarding our student network, it featured 75.00% fewer parameters than our teacher, resulting in a 36.15% reduction in inference time, a 33.33% decrease in total segmentation time, and a 42.96% reduction in CO2 emissions. This study marks the first exploration of applying KD to the problem of individual vertebrae segmentation in CT, demonstrating the feasibility of achieving comparable performance to existing methods using smaller neural networks.
Collapse
Affiliation(s)
- Luís Serrador
- Center for MicroElectroMechanical Systems (CMEMS), University of Minho, Guimaraes, Portugal; Clinical Academic Center of Braga (2CA-Braga), Hospital of Braga, Braga, Portugal.
| | | | - Sara Moccia
- The BioRobotics Institute and Department of Excellence in Robotics & AI, Scuola Superiore Sant'Anna, Italy
| | - Cristina P Santos
- Center for MicroElectroMechanical Systems (CMEMS), University of Minho, Guimaraes, Portugal; Clinical Academic Center of Braga (2CA-Braga), Hospital of Braga, Braga, Portugal
| |
Collapse
|
32
|
S R, S S, S Murthy P, Deshmukh S. Landmark annotation through feature combinations: a comparative study on cephalometric images with in-depth analysis of model's explainability. Dentomaxillofac Radiol 2024; 53:115-126. [PMID: 38166356 DOI: 10.1093/dmfr/twad011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 11/16/2023] [Accepted: 11/16/2023] [Indexed: 01/04/2024] Open
Abstract
OBJECTIVES The objectives of this study are to explore and evaluate the automation of anatomical landmark localization in cephalometric images using machine learning techniques, with a focus on feature extraction and combinations, contextual analysis, and model interpretability through Shapley Additive exPlanations (SHAP) values. METHODS We conducted extensive experimentation on a private dataset of 300 lateral cephalograms to thoroughly study the annotation results obtained using pixel feature descriptors including raw pixel, gradient magnitude, gradient direction, and histogram-oriented gradient (HOG) values. The study includes evaluation and comparison of these feature descriptions calculated at different contexts namely local, pyramid, and global. The feature descriptor obtained using individual combinations is used to discern between landmark and nonlandmark pixels using classification method. Additionally, this study addresses the opacity of LGBM ensemble tree models across landmarks, introducing SHAP values to enhance interpretability. RESULTS The performance of feature combinations was assessed using metrics like mean radial error, standard deviation, success detection rate (SDR) (2 mm), and test time. Remarkably, among all the combinations explored, both the HOG and gradient direction operations demonstrated significant performance across all context combinations. At the contextual level, the global texture outperformed the others, although it came with the trade-off of increased test time. The HOG in the local context emerged as the top performer with an SDR of 75.84% compared to others. CONCLUSIONS The presented analysis enhances the understanding of the significance of different features and their combinations in the realm of landmark annotation but also paves the way for further exploration of landmark-specific feature combination methods, facilitated by explainability.
Collapse
Affiliation(s)
- Rashmi S
- Dept. of Computer Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru, 570006, India
| | - Srinath S
- Dept. of Computer Science and Engineering, Sri Jayachamarajendra College of Engineering, JSS Science and Technology University, Mysuru, 570006, India
| | - Prashanth S Murthy
- Dept. of Pediatric & Preventive Dentistry, JSS Dental College & Hospital, JSS Academy of Higher Education & Research, Mysuru, 570015, India
| | - Seema Deshmukh
- Dept. of Pediatric & Preventive Dentistry, JSS Dental College & Hospital, JSS Academy of Higher Education & Research, Mysuru, 570015, India
| |
Collapse
|
33
|
Zhang X, Sun K, Wu D, Xiong X, Liu J, Yao L, Li S, Wang Y, Feng J, Shen D. An Anatomy- and Topology-Preserving Framework for Coronary Artery Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:723-733. [PMID: 37756173 DOI: 10.1109/tmi.2023.3319720] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/29/2023]
Abstract
Coronary artery segmentation is critical for coronary artery disease diagnosis but challenging due to its tortuous course with numerous small branches and inter-subject variations. Most existing studies ignore important anatomical information and vascular topologies, leading to less desirable segmentation performance that usually cannot satisfy clinical demands. To deal with these challenges, in this paper we propose an anatomy- and topology-preserving two-stage framework for coronary artery segmentation. The proposed framework consists of an anatomical dependency encoding (ADE) module and a hierarchical topology learning (HTL) module for coarse-to-fine segmentation, respectively. Specifically, the ADE module segments four heart chambers and aorta, and thus five distance field maps are obtained to encode distance between chamber surfaces and coarsely segmented coronary artery. Meanwhile, ADE also performs coronary artery detection to crop region-of-interest and eliminate foreground-background imbalance. The follow-up HTL module performs fine segmentation by exploiting three hierarchical vascular topologies, i.e., key points, centerlines, and neighbor connectivity using a multi-task learning scheme. In addition, we adopt a bottom-up attention interaction (BAI) module to integrate the feature representations extracted across hierarchical topologies. Extensive experiments on public and in-house datasets show that the proposed framework achieves state-of-the-art performance for coronary artery segmentation.
Collapse
|
34
|
Park JA, Kim D, Yang S, Kang JH, Kim JE, Huh KH, Lee SS, Yi WJ, Heo MS. Automatic detection of posterior superior alveolar artery in dental cone-beam CT images using a deeply supervised multi-scale 3D network. Dentomaxillofac Radiol 2024; 53:22-31. [PMID: 38214942 PMCID: PMC11003607 DOI: 10.1093/dmfr/twad002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/24/2023] [Revised: 09/15/2023] [Accepted: 10/18/2023] [Indexed: 01/13/2024] Open
Abstract
OBJECTIVES This study aimed to develop a robust and accurate deep learning network for detecting the posterior superior alveolar artery (PSAA) in dental cone-beam CT (CBCT) images, focusing on the precise localization of the centre pixel as a critical centreline pixel. METHODS PSAA locations were manually labelled on dental CBCT data from 150 subjects. The left maxillary sinus images were horizontally flipped. In total, 300 datasets were created. Six different deep learning networks were trained, including 3D U-Net, deeply supervised 3D U-Net (3D U-Net DS), multi-scale deeply supervised 3D U-Net (3D U-Net MSDS), 3D Attention U-Net, 3D V-Net, and 3D Dense U-Net. The performance evaluation involved predicting the centre pixel of the PSAA. This was assessed using mean absolute error (MAE), mean radial error (MRE), and successful detection rate (SDR). RESULTS The 3D U-Net MSDS achieved the best prediction performance among the tested networks, with an MAE measurement of 0.696 ± 1.552 mm and MRE of 1.101 ± 2.270 mm. In comparison, the 3D U-Net showed the lowest performance. The 3D U-Net MSDS demonstrated a SDR of 95% within a 2 mm MAE. This was a significantly higher result than other networks that achieved a detection rate of over 80%. CONCLUSIONS This study presents a robust deep learning network for accurate PSAA detection in dental CBCT images, emphasizing precise centre pixel localization. The method achieves high accuracy in locating small vessels, such as the PSAA, and has the potential to enhance detection accuracy and efficiency, thus impacting oral and maxillofacial surgery planning and decision-making.
Collapse
Affiliation(s)
- Jae-An Park
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - DaEl Kim
- Interdisciplinary Program in Bioengineering, Graduate School of Engineering, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul, 08826, South Korea
| | - Su Yang
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, 1 Gwanak-ro, Gwanak-gu, Seoul, 08826, South Korea
| | - Ju-Hee Kang
- Department of Oral and Maxillofacial Radiology, Seoul National University Dental Hospital, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - Jo-Eun Kim
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - Kyung-Hoe Huh
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - Sam-Sun Lee
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - Won-Jin Yi
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| | - Min-Suk Heo
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, 101 Daehak-ro, Jongno-gu, Seoul, 03080, South Korea
| |
Collapse
|
35
|
Zhai H, Chen Z, Li L, Tao H, Wang J, Li K, Shao M, Cheng X, Wang J, Wu X, Wu C, Zhang X, Kettunen L, Wang H. Two-stage multi-task deep learning framework for simultaneous pelvic bone segmentation and landmark detection from CT images. Int J Comput Assist Radiol Surg 2024; 19:97-108. [PMID: 37322299 DOI: 10.1007/s11548-023-02976-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2022] [Accepted: 05/23/2023] [Indexed: 06/17/2023]
Abstract
PURPOSE Pelvic bone segmentation and landmark definition from computed tomography (CT) images are prerequisite steps for the preoperative planning of total hip arthroplasty. In clinical applications, the diseased pelvic anatomy usually degrades the accuracies of bone segmentation and landmark detection, leading to improper surgery planning and potential operative complications. METHODS This work proposes a two-stage multi-task algorithm to improve the accuracy of pelvic bone segmentation and landmark detection, especially for the diseased cases. The two-stage framework uses a coarse-to-fine strategy which first conducts global-scale bone segmentation and landmark detection and then focuses on the important local region to further refine the accuracy. For the global stage, a dual-task network is designed to share the common features between the segmentation and detection tasks, so that the two tasks mutually reinforce each other's performance. For the local-scale segmentation, an edge-enhanced dual-task network is designed for simultaneous bone segmentation and edge detection, leading to the more accurate delineation of the acetabulum boundary. RESULTS This method was evaluated via threefold cross-validation based on 81 CT images (including 31 diseased and 50 healthy cases). The first stage achieved DSC scores of 0.94, 0.97, and 0.97 for the sacrum, left and right hips, respectively, and an average distance error of 3.24 mm for the bone landmarks. The second stage further improved the DSC of the acetabulum by 5.42%, and this accuracy outperforms the state-of-the-arts (SOTA) methods by 0.63%. Our method also accurately segmented the diseased acetabulum boundaries. The entire workflow took ~ 10 s, which was only half of the U-Net run time. CONCLUSION Using the multi-task networks and the coarse-to-fine strategy, this method achieved more accurate bone segmentation and landmark detection than the SOTA method, especially for diseased hip images. Our work contributes to accurate and rapid design of acetabular cup prostheses.
Collapse
Affiliation(s)
- Haoyu Zhai
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116024, China
| | - Zhonghua Chen
- Faculty of Information Technology, University of Jyväskylä, 40100, Jyvaskyla, Finland
| | - Lei Li
- Department of Vascular Surgery, The Second Affiliated Hospital of Dalian Medical University, Dalian, 116024, China
| | - Hairong Tao
- Shanghai Key Laboratory of Orthopaedic Implants, Shanghai, 200011, China
- Department of Orthopaedic Surgery, Shanghai Ninth People's Hospital, Shanghai, 200011, China
- Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
| | - Jinwu Wang
- Department of Orthopaedic Surgery, Shanghai Ninth People's Hospital, Shanghai, 200011, China
- Department of Orthopaedics and Bone and Joint Research Center, Shanghai Jiaotong University School of Medicine, Shanghai, 200011, China
| | - Kang Li
- West China Biomedical Big Data Center, West China Hospital, Sichuan University, Chengdu, 610041, China
| | - Moyu Shao
- Jiangsu Yunqianbai Digital Technology Co., LTD, Xuzhou, 221000, China
| | - Xiaomin Cheng
- Jiangsu Yunqianbai Digital Technology Co., LTD, Xuzhou, 221000, China
| | - Jing Wang
- School of Chemical Engineering and Technology, Xi'an JiaoTong University, Xi'an, 710049, China
| | - Xiang Wu
- School of Medical Information and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Chuan Wu
- School of Medical Information and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Xiao Zhang
- School of Medical Information and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Lauri Kettunen
- Faculty of Information Technology, University of Jyväskylä, 40100, Jyvaskyla, Finland
| | - Hongkai Wang
- School of Biomedical Engineering, Faculty of Medicine, Dalian University of Technology, Dalian, 116024, China.
- Liaoning Key Laboratory of Integrated Circuit and Biomedical Electronic System, Dalian, 116024, China.
| |
Collapse
|
36
|
Hadzic A, Urschler M, Press JNA, Riedl R, Rugani P, Štern D, Kirnbauer B. Evaluating a Periapical Lesion Detection CNN on a Clinically Representative CBCT Dataset-A Validation Study. J Clin Med 2023; 13:197. [PMID: 38202204 PMCID: PMC10779652 DOI: 10.3390/jcm13010197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2023] [Revised: 12/20/2023] [Accepted: 12/25/2023] [Indexed: 01/12/2024] Open
Abstract
The aim of this validation study was to comprehensively evaluate the performance and generalization capability of a deep learning-based periapical lesion detection algorithm on a clinically representative cone-beam computed tomography (CBCT) dataset and test for non-inferiority. The evaluation involved 195 CBCT images of adult upper and lower jaws, where sensitivity and specificity metrics were calculated for all teeth, stratified by jaw, and stratified by tooth type. Furthermore, each lesion was assigned a periapical index score based on its size to enable a score-based evaluation. Non-inferiority tests were conducted with proportions of 90% for sensitivity and 82% for specificity. The algorithm achieved an overall sensitivity of 86.7% and a specificity of 84.3%. The non-inferiority test indicated the rejection of the null hypothesis for specificity but not for sensitivity. However, when excluding lesions with a periapical index score of one (i.e., very small lesions), the sensitivity improved to 90.4%. Despite the challenges posed by the dataset, the algorithm demonstrated promising results. Nevertheless, further improvements are needed to enhance the algorithm's robustness, particularly in detecting very small lesions and the handling of artifacts and outliers commonly encountered in real-world clinical scenarios.
Collapse
Affiliation(s)
- Arnela Hadzic
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria; (A.H.); (R.R.)
| | - Martin Urschler
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria; (A.H.); (R.R.)
| | - Jan-Niclas Aaron Press
- Division of Oral Surgery and Orthodontics, Medical University of Graz, 8010 Graz, Austria (P.R.); (B.K.)
| | - Regina Riedl
- Institute for Medical Informatics, Statistics and Documentation, Medical University of Graz, 8036 Graz, Austria; (A.H.); (R.R.)
| | - Petra Rugani
- Division of Oral Surgery and Orthodontics, Medical University of Graz, 8010 Graz, Austria (P.R.); (B.K.)
| | - Darko Štern
- Institute of Computer Graphics and Vision, Graz University of Technology, 8010 Graz, Austria
| | - Barbara Kirnbauer
- Division of Oral Surgery and Orthodontics, Medical University of Graz, 8010 Graz, Austria (P.R.); (B.K.)
| |
Collapse
|
37
|
Roth T, Sigrist B, Wieczorek M, Schilling N, Hodel S, Walker J, Somm M, Wein W, Sutter R, Vlachopoulos L, Snedeker JG, Fucentese SF, Fürnstahl P, Carrillo F. An automated optimization pipeline for clinical-grade computer-assisted planning of high tibial osteotomies under consideration of weight-bearing. Comput Assist Surg (Abingdon) 2023; 28:2211728. [PMID: 37191179 DOI: 10.1080/24699322.2023.2211728] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/17/2023] Open
Abstract
3D preoperative planning for high tibial osteotomies (HTO) has increasingly replaced 2D planning but is complex, time-consuming and therefore expensive. Several interdependent clinical objectives and constraints have to be considered, which often requires multiple rounds of revisions between surgeons and biomedical engineers. We therefore developed an automated preoperative planning pipeline, which takes imaging data as an input to generate a ready-to-use, patient-specific planning solution. Deep-learning based segmentation and landmark localization was used to enable the fully automated 3D lower limb deformity assessment. A 2D-3D registration algorithm allowed the transformation of the 3D bone models into the weight-bearing state. Finally, an optimization framework was implemented to generate ready-to use preoperative plannings in a fully automated fashion, using a genetic algorithm to solve the multi-objective optimization (MOO) problem based on several clinical requirements and constraints. The entire pipeline was evaluated on a large clinical dataset of 53 patient cases who previously underwent a medial opening-wedge HTO. The pipeline was used to automatically generate preoperative solutions for these patients. Five experts blindly compared the automatically generated solutions to the previously generated manual plannings. The overall mean rating for the algorithm-generated solutions was better than for the manual solutions. In 90% of all comparisons, they were considered to be equally good or better than the manual solution. The combined use of deep learning approaches, registration methods and MOO can reliably produce ready-to-use preoperative solutions that significantly reduce human workload and related health costs.
Collapse
Affiliation(s)
- Tabitha Roth
- Institute for Biomechanics, ETH Zurich, Zurich, Switzerland
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Bastian Sigrist
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | | | | | - Sandro Hodel
- Department of Orthopedics, Balgrist University Hospital, University of Zurich, Zürich, Switzerland
| | - Jonas Walker
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Mario Somm
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | | | - Reto Sutter
- Department of Radiology, Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Lazaros Vlachopoulos
- Department of Orthopedics, Balgrist University Hospital, University of Zurich, Zürich, Switzerland
| | | | - Sandro F Fucentese
- Department of Orthopedics, Balgrist University Hospital, University of Zurich, Zürich, Switzerland
| | - Philipp Fürnstahl
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Fabio Carrillo
- Research in Orthopedic Computer Science (ROCS), Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| |
Collapse
|
38
|
Nicolaes J, Skjødt MK, Raeymaeckers S, Smith CD, Abrahamsen B, Fuerst T, Debois M, Vandermeulen D, Libanati C. Towards Improved Identification of Vertebral Fractures in Routine Computed Tomography (CT) Scans: Development and External Validation of a Machine Learning Algorithm. J Bone Miner Res 2023; 38:1856-1866. [PMID: 37747147 DOI: 10.1002/jbmr.4916] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/17/2023] [Revised: 09/06/2023] [Accepted: 09/17/2023] [Indexed: 09/26/2023]
Abstract
Vertebral fractures (VFs) are the hallmark of osteoporosis, being one of the most frequent types of fragility fracture and an early sign of the disease. They are associated with significant morbidity and mortality. VFs are incidentally found in one out of five imaging studies, however, more than half of the VFs are not identified nor reported in patient computed tomography (CT) scans. Our study aimed to develop a machine learning algorithm to identify VFs in abdominal/chest CT scans and evaluate its performance. We acquired two independent data sets of routine abdominal/chest CT scans of patients aged 50 years or older: a training set of 1011 scans from a non-interventional, prospective proof-of-concept study at the Universitair Ziekenhuis (UZ) Brussel and a validation set of 2000 subjects from an observational cohort study at the Hospital of Holbaek. Both data sets were externally reevaluated to identify reference standard VF readings using the Genant semiquantitative (SQ) grading. Four independent models have been trained in a cross-validation experiment using the training set and an ensemble of four models has been applied to the external validation set. The validation set contained 15.3% scans with one or more VF (SQ2-3), whereas 663 of 24,930 evaluable vertebrae (2.7%) were fractured (SQ2-3) as per reference standard readings. Comparison of the ensemble model with the reference standard readings in identifying subjects with one or more moderate or severe VF resulted in an area under the receiver operating characteristic curve (AUROC) of 0.88 (95% confidence interval [CI], 0.85-0.90), accuracy of 0.92 (95% CI, 0.91-0.93), kappa of 0.72 (95% CI, 0.67-0.76), sensitivity of 0.81 (95% CI, 0.76-0.85), and specificity of 0.95 (95% CI, 0.93-0.96). We demonstrated that a machine learning algorithm trained for VF detection achieved strong performance on an external validation set. It has the potential to support healthcare professionals with the early identification of VFs and prevention of future fragility fractures. © 2023 UCB S.A. and The Authors. Journal of Bone and Mineral Research published by Wiley Periodicals LLC on behalf of American Society for Bone and Mineral Research (ASBMR).
Collapse
Affiliation(s)
- Joeri Nicolaes
- Department of Electrical Engineering (ESAT), Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
- UCB Pharma, Brussels, Belgium
| | - Michael Kriegbaum Skjødt
- Department of Medicine, Hospital of Holbaek, Holbaek, Denmark
- OPEN-Open Patient Data Explorative Network, Department of Clinical Research, University of Southern Denmark and Odense University Hospital, Odense, Denmark
| | | | - Christopher Dyer Smith
- OPEN-Open Patient Data Explorative Network, Department of Clinical Research, University of Southern Denmark and Odense University Hospital, Odense, Denmark
| | - Bo Abrahamsen
- Department of Medicine, Hospital of Holbaek, Holbaek, Denmark
- OPEN-Open Patient Data Explorative Network, Department of Clinical Research, University of Southern Denmark and Odense University Hospital, Odense, Denmark
- NDORMS, Nuffield Department of Orthopaedics, Rheumatology and Musculoskeletal Sciences, Oxford University Hospitals, Oxford, UK
| | | | | | - Dirk Vandermeulen
- Department of Electrical Engineering (ESAT), Center for Processing Speech and Images, KU Leuven, Leuven, Belgium
| | | |
Collapse
|
39
|
Suna A, Davidson A, Weil Y, Joskowicz L. Automated computation of radiographic parameters of distal radial metaphyseal fractures in forearm X-rays. Int J Comput Assist Radiol Surg 2023; 18:2179-2189. [PMID: 37097517 DOI: 10.1007/s11548-023-02907-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2022] [Accepted: 04/03/2023] [Indexed: 04/26/2023]
Abstract
PURPOSE Radiographic parameters (RPs) provide objective support for effective decision making in determining clinical treatment of distal radius fractures (DRFs). This paper presents a novel automatic RP computation pipeline for computing the six anatomical RPs associated with DRFs in anteroposterior (AP) and lateral (LAT) forearm radiographs. METHODS The pipeline consists of: (1) segmentation of the distal radius and ulna bones with six 2D Dynamic U-Net deep learning models; (2) landmark points detection and distal radius axis computation from the segmentations with geometric methods; (3) RP computation and generation of a quantitative DRF report and composite AP and LAT radiograph images. This hybrid approach combines the advantages of deep learning and model-based methods. RESULTS The pipeline was evaluated on 90 AP and 93 LAT radiographs for which ground truth distal radius and ulna segmentations and RP landmarks were manually obtained by expert clinicians. It achieves an accuracy of 94 and 86% on the AP and LAT RPs, within the observer variability, and an RP measurement difference of 1.4 ± 1.2° for the radial angle, 0.5 ± 0.6 mm for the radial length, 0.9 ± 0.7 mm for the radial shift, 0.7 ± 0.5 mm for the ulnar variance, 2.9 ± 3.3° for the palmar tilt and 1.2 ± 1.0 mm for the dorsal shift. CONCLUSION Our pipeline is the first fully automatic method that accurately and robustly computes the RPs for a wide variety of clinical forearm radiographs from different sources, hand orientations, with and without cast. The computed accurate and reliable RF measurements may support fracture severity assessment and clinical management.
Collapse
Affiliation(s)
- Avigail Suna
- School of Computer Science and Engineering, The Hebrew University of Jerusalem, Israel, Edmond J. Safra Campus, Givat Ram, 9190401, Jerusalem, Israel
| | - Amit Davidson
- Department of Orthopedics, Hadassah University Medical Center, Jerusalem, Israel
| | - Yoram Weil
- Department of Orthopedics, Hadassah University Medical Center, Jerusalem, Israel
| | - Leo Joskowicz
- School of Computer Science and Engineering, The Hebrew University of Jerusalem, Israel, Edmond J. Safra Campus, Givat Ram, 9190401, Jerusalem, Israel.
| |
Collapse
|
40
|
Tao L, Li M, Zhang X, Cheng M, Yang Y, Fu Y, Zhang R, Qian D, Yu H. Automatic craniomaxillofacial landmarks detection in CT images of individuals with dentomaxillofacial deformities by a two-stage deep learning model. BMC Oral Health 2023; 23:876. [PMID: 37978486 PMCID: PMC10657133 DOI: 10.1186/s12903-023-03446-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2023] [Accepted: 09/22/2023] [Indexed: 11/19/2023] Open
Abstract
BACKGROUND Accurate cephalometric analysis plays a vital role in the diagnosis and subsequent surgical planning in orthognathic and orthodontics treatment. However, manual digitization of anatomical landmarks in computed tomography (CT) is subject to limitations such as low accuracy, poor repeatability and excessive time consumption. Furthermore, the detection of landmarks has more difficulties on individuals with dentomaxillofacial deformities than normal individuals. Therefore, this study aims to develop a deep learning model to automatically detect landmarks in CT images of patients with dentomaxillofacial deformities. METHODS Craniomaxillofacial (CMF) CT data of 80 patients with dentomaxillofacial deformities were collected for model development. 77 anatomical landmarks digitized by experienced CMF surgeons in each CT image were set as the ground truth. 3D UX-Net, the cutting-edge medical image segmentation network, was adopted as the backbone of model architecture. Moreover, a new region division pattern for CMF structures was designed as a training strategy to optimize the utilization of computational resources and image resolution. To evaluate the performance of this model, several experiments were conducted to make comparison between the model and manual digitization approach. RESULTS The training set and the validation set included 58 and 22 samples respectively. The developed model can accurately detect 77 landmarks on bone, soft tissue and teeth with a mean error of 1.81 ± 0.89 mm. Removal of region division before training significantly increased the error of prediction (2.34 ± 1.01 mm). In terms of manual digitization, the inter-observer and intra-observer variations were 1.27 ± 0.70 mm and 1.01 ± 0.74 mm respectively. In all divided regions except Teeth Region (TR), our model demonstrated equivalent performance to experienced CMF surgeons in landmarks detection (p > 0.05). CONCLUSIONS The developed model demonstrated excellent performance in detecting craniomaxillofacial landmarks when considering manual digitization work of expertise as benchmark. It is also verified that the region division pattern designed in this study remarkably improved the detection accuracy.
Collapse
Affiliation(s)
- Leran Tao
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| | - Meng Li
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| | - Xu Zhang
- Mechanical college, Shanghai Dianji University, Shanghai, 201306, China
| | - Mengjia Cheng
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China
| | - Yang Yang
- Shanghai Lanhui Medical Technology Co., Ltd, Shanghai, 200333, China
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200030, China
| | - Yijiao Fu
- College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200125, China
| | - Rongbin Zhang
- College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200125, China
| | - Dahong Qian
- School of Biomedical Engineering, Shanghai Jiao Tong University, Shanghai, 200030, China.
| | - Hongbo Yu
- Department of Oral and Cranio-maxillofacial Surgery, Shanghai Ninth People's Hospital, College of Stomatology, Shanghai Jiao Tong University School of Medicine, Shanghai, 200011, China.
- National Center for Stomatology & National Clinical Research Center for Oral Diseases, Shanghai, 200011, China.
- Shanghai Key Laboratory of Stomatology & Shanghai Research Institute of Stomatology, Shanghai, 200011, China.
| |
Collapse
|
41
|
Hong W, Kim SM, Choi J, Ahn J, Paeng JY, Kim H. Automated Cephalometric Landmark Detection Using Deep Reinforcement Learning. J Craniofac Surg 2023; 34:2336-2342. [PMID: 37622568 DOI: 10.1097/scs.0000000000009685] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Accepted: 06/25/2023] [Indexed: 08/26/2023] Open
Abstract
Accurate cephalometric landmark detection leads to accurate analysis, diagnosis, and surgical planning. Many studies on automated landmark detection have been conducted, however reinforcement learning-based networks have not yet been applied. This is the first study to apply deep Q-network (DQN) and double deep Q-network (DDQN) to automated cephalometric landmark detection to the best of our knowledge. The performance of the DQN-based network for cephalometric landmark detection was evaluated using the IEEE International Symposium of Biomedical Imaging (ISBI) 2015 Challenge data set and compared with the previously proposed methods. Furthermore, the clinical applicability of DQN-based automated cephalometric landmark detection was confirmed by testing the DQN-based and DDQN-based network using 500-patient data collected in a clinic. The DQN-based network demonstrated that the average mean radius error of 19 landmarks was smaller than 2 mm, that is, the clinically accepted level, without data augmentation and additional preprocessing. Our DQN-based and DDQN-based approaches tested with the 500-patient data set showed the average success detection rate of 67.33% and 66.04% accuracy within 2 mm, respectively, indicating the feasibility and potential of clinical application.
Collapse
Affiliation(s)
- Woojae Hong
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Seong-Min Kim
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Joongyeon Choi
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| | - Jaemyung Ahn
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Seoul, Republic of Korea
| | - Jun-Young Paeng
- Department of Oral and Maxillofacial Surgery, Samsung Medical Center, Seoul, Republic of Korea
| | - Hyunggun Kim
- Department of Biomechatronic Engineering, Sungkyunkwan University, Suwon, Gyeonggi
| |
Collapse
|
42
|
Yang S, Song ES, Lee ES, Kang SR, Yi WJ, Lee SP. Ceph-Net: automatic detection of cephalometric landmarks on scanned lateral cephalograms from children and adolescents using an attention-based stacked regression network. BMC Oral Health 2023; 23:803. [PMID: 37884918 PMCID: PMC10604948 DOI: 10.1186/s12903-023-03452-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2023] [Accepted: 09/25/2023] [Indexed: 10/28/2023] Open
Abstract
BACKGROUND The success of cephalometric analysis depends on the accurate detection of cephalometric landmarks on scanned lateral cephalograms. However, manual cephalometric analysis is time-consuming and can cause inter- and intra-observer variability. The purpose of this study was to automatically detect cephalometric landmarks on scanned lateral cephalograms with low contrast and resolution using an attention-based stacked regression network (Ceph-Net). METHODS The main body of Ceph-Net compromised stacked fully convolutional networks (FCN) which progressively refined the detection of cephalometric landmarks on each FCN. By embedding dual attention and multi-path convolution modules in Ceph-Net, the network learned local and global context and semantic relationships between cephalometric landmarks. Additionally, the intermediate deep supervision in each FCN further boosted the training stability and the detection performance of cephalometric landmarks. RESULTS Ceph-Net showed a superior detection performance in mean radial error and successful detection rate, including accuracy improvements in cephalometric landmark detection located in low-contrast soft tissues compared with other detection networks. Moreover, Ceph-Net presented superior detection performance on the test dataset split by age from 8 to 16 years old. CONCLUSIONS Ceph-Net demonstrated an automatic and superior detection of cephalometric landmarks by successfully learning local and global context and semantic relationships between cephalometric landmarks in scanned lateral cephalograms with low contrast and resolutions.
Collapse
Affiliation(s)
- Su Yang
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea
| | - Eun Sun Song
- Department of Oral Anatomy, Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Eun Seung Lee
- Department of Oral Anatomy, Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea
| | - Se-Ryong Kang
- Department of Biomedical Radiation Sciences, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea
| | - Won-Jin Yi
- Department of Applied Bioengineering, Graduate School of Convergence Science and Technology, Seoul National University, Seoul, South Korea.
- Department of Oral and Maxillofacial Radiology and Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea.
| | - Seung-Pyo Lee
- Department of Oral Anatomy, Dental Research Institute, School of Dentistry, Seoul National University, Seoul, South Korea.
| |
Collapse
|
43
|
Liu J, Zhang C, Shan Z. Application of Artificial Intelligence in Orthodontics: Current State and Future Perspectives. Healthcare (Basel) 2023; 11:2760. [PMID: 37893833 PMCID: PMC10606213 DOI: 10.3390/healthcare11202760] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 10/11/2023] [Accepted: 10/16/2023] [Indexed: 10/29/2023] Open
Abstract
In recent years, there has been the notable emergency of artificial intelligence (AI) as a transformative force in multiple domains, including orthodontics. This review aims to provide a comprehensive overview of the present state of AI applications in orthodontics, which can be categorized into the following domains: (1) diagnosis, including cephalometric analysis, dental analysis, facial analysis, skeletal-maturation-stage determination and upper-airway obstruction assessment; (2) treatment planning, including decision making for extractions and orthognathic surgery, and treatment outcome prediction; and (3) clinical practice, including practice guidance, remote care, and clinical documentation. We have witnessed a broadening of the application of AI in orthodontics, accompanied by advancements in its performance. Additionally, this review outlines the existing limitations within the field and offers future perspectives.
Collapse
Affiliation(s)
- Junqi Liu
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| | - Chengfei Zhang
- Division of Restorative Dental Sciences, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| | - Zhiyi Shan
- Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China;
| |
Collapse
|
44
|
Liu J, Xing F, Shaikh A, French B, Linguraru MG, Porras AR. Joint Cranial Bone Labeling and Landmark Detection in Pediatric CT Images Using Context Encoding. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3117-3126. [PMID: 37216247 PMCID: PMC10760565 DOI: 10.1109/tmi.2023.3278493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/24/2023]
Abstract
Image segmentation, labeling, and landmark detection are essential tasks for pediatric craniofacial evaluation. Although deep neural networks have been recently adopted to segment cranial bones and locate cranial landmarks from computed tomography (CT) or magnetic resonance (MR) images, they may be hard to train and provide suboptimal results in some applications. First, they seldom leverage global contextual information that can improve object detection performance. Second, most methods rely on multi-stage algorithm designs that are inefficient and prone to error accumulation. Third, existing methods often target simple segmentation tasks and have shown low reliability in more challenging scenarios such as multiple cranial bone labeling in highly variable pediatric datasets. In this paper, we present a novel end-to-end neural network architecture based on DenseNet that incorporates context regularization to jointly label cranial bone plates and detect cranial base landmarks from CT images. Specifically, we designed a context-encoding module that encodes global context information as landmark displacement vector maps and uses it to guide feature learning for both bone labeling and landmark identification. We evaluated our model on a highly diverse pediatric CT image dataset of 274 normative subjects and 239 patients with craniosynostosis (age 0.63 ± 0.54 years, range 0-2 years). Our experiments demonstrate improved performance compared to state-of-the-art approaches.
Collapse
|
45
|
Pepe A, Egger J, Codari M, Willemink MJ, Gsaxner C, Li J, Roth PM, Schmalstieg D, Mistelbauer G, Fleischmann D. Automated cross-sectional view selection in CT angiography of aortic dissections with uncertainty awareness and retrospective clinical annotations. Comput Biol Med 2023; 165:107365. [PMID: 37647783 DOI: 10.1016/j.compbiomed.2023.107365] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2023] [Revised: 07/20/2023] [Accepted: 08/12/2023] [Indexed: 09/01/2023]
Abstract
Surveillance imaging of patients with chronic aortic diseases, such as aneurysms and dissections, relies on obtaining and comparing cross-sectional diameter measurements along the aorta at predefined aortic landmarks, over time. The orientation of the cross-sectional measuring planes at each landmark is currently defined manually by highly trained operators. Centerline-based approaches are unreliable in patients with chronic aortic dissection, because of the asymmetric flow channels, differences in contrast opacification, and presence of mural thrombus, making centerline computations or measurements difficult to generate and reproduce. In this work, we present three alternative approaches - INS, MCDS, MCDbS - based on convolutional neural networks and uncertainty quantification methods to predict the orientation (ϕ,θ) of such cross-sectional planes. For the monitoring of chronic aortic dissections, we show how a dataset of 162 CTA volumes with overall 3273 imperfect manual annotations routinely collected in a clinic can be efficiently used to accomplish this task, despite the presence of non-negligible interoperator variabilities in terms of mean absolute error (MAE) and 95% limits of agreement (LOA). We show how, despite the large limits of agreement in the training data, the trained model provides faster and more reproducible results than either an expert user or a centerline method. The remaining disagreement lies within the variability produced by three independent expert annotators and matches the current state of the art, providing a similar error, but in a fraction of the time.
Collapse
Affiliation(s)
- Antonio Pepe
- Graz University of Technology, Institute of Computer Graphics and Vision, Inffeldgasse 16/II, 8010 Graz, Austria; Stanford University, School of Medicine, 3D and Quantitative Imaging Lab, 300 Pasteur Drive Stanford, CA 94305, USA; Computer Algorithms for Médicine (Café) Laboratory, Graz, Austria.
| | - Jan Egger
- Computer Algorithms for Médicine (Café) Laboratory, Graz, Austria; University Medicine Essen, Institute for AI in Medicine (IKIM), Girardetstraße 2, 45131 Essen, Germany.
| | - Marina Codari
- Stanford University, School of Medicine, 3D and Quantitative Imaging Lab, 300 Pasteur Drive Stanford, CA 94305, USA.
| | - Martin J Willemink
- Stanford University, School of Medicine, 3D and Quantitative Imaging Lab, 300 Pasteur Drive Stanford, CA 94305, USA.
| | - Christina Gsaxner
- Graz University of Technology, Institute of Computer Graphics and Vision, Inffeldgasse 16/II, 8010 Graz, Austria; Computer Algorithms for Médicine (Café) Laboratory, Graz, Austria.
| | - Jianning Li
- Computer Algorithms for Médicine (Café) Laboratory, Graz, Austria; University Medicine Essen, Institute for AI in Medicine (IKIM), Girardetstraße 2, 45131 Essen, Germany.
| | - Peter M Roth
- Graz University of Technology, Institute of Computer Graphics and Vision, Inffeldgasse 16/II, 8010 Graz, Austria.
| | - Dieter Schmalstieg
- Graz University of Technology, Institute of Computer Graphics and Vision, Inffeldgasse 16/II, 8010 Graz, Austria.
| | - Gabriel Mistelbauer
- Stanford University, School of Medicine, 3D and Quantitative Imaging Lab, 300 Pasteur Drive Stanford, CA 94305, USA.
| | - Dominik Fleischmann
- Stanford University, School of Medicine, 3D and Quantitative Imaging Lab, 300 Pasteur Drive Stanford, CA 94305, USA.
| |
Collapse
|
46
|
Katakol S, Baker TJ, Bian Z, Lu Y, Spahlinger G, Hatt CR, Burris NS. Fully automated pipeline for measurement of the thoracic aorta using joint segmentation and localization neural network. J Med Imaging (Bellingham) 2023; 10:051810. [PMID: 37915405 PMCID: PMC10617550 DOI: 10.1117/1.jmi.10.5.051810] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2023] [Revised: 09/14/2023] [Accepted: 10/24/2023] [Indexed: 11/03/2023] Open
Abstract
Purpose Diagnosis and surveillance of thoracic aortic aneurysm (TAA) involves measuring the aortic diameter at various locations along the length of the aorta, often using computed tomography angiography (CTA). Currently, measurements are performed by human raters using specialized software for three-dimensional analysis, a time-consuming process, requiring 15 to 45 min of focused effort. Thus, we aimed to develop a convolutional neural network (CNN)-based algorithm for fully automated and accurate aortic measurements. Approach Using 212 CTA scans, we trained a CNN to perform segmentation and localization of key landmarks jointly. Segmentation mask and landmarks are subsequently used to obtain the centerline and cross-sectional diameters of the aorta. Subsequently, a cubic spline is fit to the aortic boundary at the sinuses of Valsalva to avoid errors related inclusions of coronary artery origins. Performance was evaluated on a test set of 60 scans with automated measurements compared against expert manual raters. Result Compared to training separate networks for each task, joint training yielded higher accuracy for segmentation, especially at the boundary (p < 0.001 ), but a marginally worse (0.2 to 0.5 mm) accuracy for landmark localization (p < 0.001 ). Mean absolute error between human and automated was ≤ 1 mm at six of nine standard clinical measurement locations. However, higher errors were noted in the aortic root and arch regions, ranging between 1.4 and 2.2 mm, although agreement of manual raters was also lower in these regions. Conclusion Fully automated aortic diameter measurements in TAA are feasible using a CNN-based algorithm. Automated measurements demonstrated low errors that are comparable in magnitude to those with manual raters; however, measurement error was highest in the aortic root and arch.
Collapse
Affiliation(s)
- Sudeep Katakol
- University of Michigan, Department of Electrical and Computer Engineering, Ann Arbor, Michigan, United States
- University of Michigan, Department of Radiology, Ann Arbor, Michigan, United States
| | - Timothy J. Baker
- University of Michigan, Department of Radiology, Ann Arbor, Michigan, United States
| | - Zhangxing Bian
- Johns Hopkins University, Department of Electrical and Computer Engineering, Baltimore, Maryland, United States
| | - Yanglong Lu
- University of Michigan, Department of Radiology, Ann Arbor, Michigan, United States
| | - Greg Spahlinger
- University of Michigan, Department of Radiology, Ann Arbor, Michigan, United States
| | | | - Nicholas S. Burris
- University of Michigan, Department of Radiology, Ann Arbor, Michigan, United States
| |
Collapse
|
47
|
Nishimoto S, Saito T, Ishise H, Fujiwara T, Kawai K, Kakibuchi M. Three-Dimensional Craniofacial Landmark Detection in Series of CT Slices Using Multi-Phased Regression Networks. Diagnostics (Basel) 2023; 13:diagnostics13111930. [PMID: 37296782 DOI: 10.3390/diagnostics13111930] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Revised: 05/26/2023] [Accepted: 05/28/2023] [Indexed: 06/12/2023] Open
Abstract
Geometrical assessments of human skulls have been conducted based on anatomical landmarks. If developed, the automatic detection of these landmarks will yield both medical and anthropological benefits. In this study, an automated system with multi-phased deep learning networks was developed to predict the three-dimensional coordinate values of craniofacial landmarks. Computed tomography images of the craniofacial area were obtained from a publicly available database. They were digitally reconstructed into three-dimensional objects. Sixteen anatomical landmarks were plotted on each of the objects, and their coordinate values were recorded. Three-phased regression deep learning networks were trained using ninety training datasets. For the evaluation, 30 testing datasets were employed. The 3D error for the first phase, which tested 30 data, was 11.60 px on average (1 px = 500/512 mm). For the second phase, it was significantly improved to 4.66 px. For the third phase, it was further significantly reduced to 2.88. This was comparable to the gaps between the landmarks, as plotted by two experienced practitioners. Our proposed method of multi-phased prediction, which conducts coarse detection first and narrows down the detection area, may be a possible solution to prediction problems, taking into account the physical limitations of memory and computation.
Collapse
Affiliation(s)
- Soh Nishimoto
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| | - Takuya Saito
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| | - Hisako Ishise
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| | - Toshihiro Fujiwara
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| | - Kenichiro Kawai
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| | - Masao Kakibuchi
- Department of Plastic Surgery, Hyogo Medical University, Nishinomiya 663-8501, Japan
| |
Collapse
|
48
|
de Queiroz Tavares Borges Mesquita G, Vieira WA, Vidigal MTC, Travençolo BAN, Beaini TL, Spin-Neto R, Paranhos LR, de Brito Júnior RB. Artificial Intelligence for Detecting Cephalometric Landmarks: A Systematic Review and Meta-analysis. J Digit Imaging 2023; 36:1158-1179. [PMID: 36604364 PMCID: PMC10287619 DOI: 10.1007/s10278-022-00766-w] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2022] [Revised: 11/19/2022] [Accepted: 12/19/2022] [Indexed: 01/07/2023] Open
Abstract
Using computer vision through artificial intelligence (AI) is one of the main technological advances in dentistry. However, the existing literature on the practical application of AI for detecting cephalometric landmarks of orthodontic interest in digital images is heterogeneous, and there is no consensus regarding accuracy and precision. Thus, this review evaluated the use of artificial intelligence for detecting cephalometric landmarks in digital imaging examinations and compared it to manual annotation of landmarks. An electronic search was performed in nine databases to find studies that analyzed the detection of cephalometric landmarks in digital imaging examinations with AI and manual landmarking. Two reviewers selected the studies, extracted the data, and assessed the risk of bias using QUADAS-2. Random-effects meta-analyses determined the agreement and precision of AI compared to manual detection at a 95% confidence interval. The electronic search located 7410 studies, of which 40 were included. Only three studies presented a low risk of bias for all domains evaluated. The meta-analysis showed AI agreement rates of 79% (95% CI: 76-82%, I2 = 99%) and 90% (95% CI: 87-92%, I2 = 99%) for the thresholds of 2 and 3 mm, respectively, with a mean divergence of 2.05 (95% CI: 1.41-2.69, I2 = 10%) compared to manual landmarking. The menton cephalometric landmark showed the lowest divergence between both methods (SMD, 1.17; 95% CI, 0.82; 1.53; I2 = 0%). Based on very low certainty of evidence, the application of AI was promising for automatically detecting cephalometric landmarks, but further studies should focus on testing its strength and validity in different samples.
Collapse
Affiliation(s)
| | - Walbert A Vieira
- Department of Restorative Dentistry, Endodontics Division, School of Dentistry of Piracicaba, State University of Campinas, Piracicaba, São Paulo, Brazil
| | | | | | - Thiago Leite Beaini
- Department of Preventive and Community Dentistry, School of Dentistry, Federal University of Uberlândia, Campus Umuarama Av. Pará, 1720, Bloco 2G, sala 1, 38405-320, Uberlândia, Minas Gerais, Brazil
| | - Rubens Spin-Neto
- Department of Dentistry and Oral Health, Section for Oral Radiology, Aarhus University, Aarhus C, Denmark
| | - Luiz Renato Paranhos
- Department of Preventive and Community Dentistry, School of Dentistry, Federal University of Uberlândia, Campus Umuarama Av. Pará, 1720, Bloco 2G, sala 1, 38405-320, Uberlândia, Minas Gerais, Brazil.
| | | |
Collapse
|
49
|
Lu H, Li M, Yu K, Zhang Y, Yu L. Lumbar spine segmentation method based on deep learning. J Appl Clin Med Phys 2023:e13996. [PMID: 37082799 DOI: 10.1002/acm2.13996] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Revised: 02/13/2023] [Accepted: 04/04/2023] [Indexed: 04/22/2023] Open
Abstract
Aiming at the difficulties of lumbar vertebrae segmentation in computed tomography (CT) images, we propose an automatic lumbar vertebrae segmentation method based on deep learning. The method mainly includes two parts: lumbar vertebra positioning and lumbar vertebrae segmentation. First of all, we propose a lumbar spine localization network of Unet network, which can directly locate the lumbar spine part in the image. Then, we propose a three-dimensional XUnet lumbar vertebrae segmentation method to achieve automatic lumbar vertebrae segmentation. The method proposed in this paper was validated on the lumbar spine CT images on the public dataset VerSe 2020 and our hospital dataset. Through qualitative comparison and quantitative analysis, the experimental results show that the method proposed in this paper can obtain good lumbar vertebrae segmentation performance, which can be further applied to detection of spinal anomalies and surgical treatment.
Collapse
Affiliation(s)
- Hongjiang Lu
- Department of Radiology, 903 Hospital of the Joint Service Support Force of the Chinese People's Liberation Army, Hangzhou, Zhejiang, China
| | - Mingying Li
- Department of Radiology, 903 Hospital of the Joint Service Support Force of the Chinese People's Liberation Army, Hangzhou, Zhejiang, China
| | - Kun Yu
- Department of Head and Neck & Thyroid Surgery, Zhejiang Provincial People's Hospital, Affiliated People's Hospital, Hangzhou Medical College, Hangzhou, Zhejiang, China
| | - Yingjiao Zhang
- Department of Gastroenterology, 903 Hospital of the Joint Service Support Force of the Chinese People's Liberation Army, Hangzhou, Zhejiang, China
| | - Liang Yu
- Department of Radiology, Zhejiang Provincial People's Hospital, Affiliated People's Hospital, Hangzhou Medical College, Hangzhou, Zhejiang, China
| |
Collapse
|
50
|
Li J, Chen J, Tang Y, Wang C, Landman BA, Zhou SK. Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives. Med Image Anal 2023; 85:102762. [PMID: 36738650 PMCID: PMC10010286 DOI: 10.1016/j.media.2023.102762] [Citation(s) in RCA: 54] [Impact Index Per Article: 27.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/02/2022] [Revised: 01/18/2023] [Accepted: 01/27/2023] [Indexed: 02/01/2023]
Abstract
Transformer, one of the latest technological advances of deep learning, has gained prevalence in natural language processing or computer vision. Since medical imaging bear some resemblance to computer vision, it is natural to inquire about the status quo of Transformers in medical imaging and ask the question: can the Transformer models transform medical imaging? In this paper, we attempt to make a response to the inquiry. After a brief introduction of the fundamentals of Transformers, especially in comparison with convolutional neural networks (CNNs), and highlighting key defining properties that characterize the Transformers, we offer a comprehensive review of the state-of-the-art Transformer-based approaches for medical imaging and exhibit current research progresses made in the areas of medical image segmentation, recognition, detection, registration, reconstruction, enhancement, etc. In particular, what distinguishes our review lies in its organization based on the Transformer's key defining properties, which are mostly derived from comparing the Transformer and CNN, and its type of architecture, which specifies the manner in which the Transformer and CNN are combined, all helping the readers to best understand the rationale behind the reviewed approaches. We conclude with discussions of future perspectives.
Collapse
Affiliation(s)
- Jun Li
- Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China
| | - Junyu Chen
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins Medical Institutes, Baltimore, MD, USA
| | - Yucheng Tang
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN, USA
| | - Ce Wang
- Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China
| | - Bennett A Landman
- Department of Electrical and Computer Engineering, Vanderbilt University, Nashville, TN, USA
| | - S Kevin Zhou
- Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing 100190, China; School of Biomedical Engineering & Suzhou Institute for Advanced Research, Center for Medical Imaging, Robotics, and Analytic Computing & Learning (MIRACLE), University of Science and Technology of China, Suzhou 215123, China.
| |
Collapse
|