Artificial intelligence-based automated preprocessing and classification of impacted maxillary canines in panoramic radiographs.
Dentomaxillofac Radiol 2024;
53:173-177. [PMID:
38374464 PMCID:
PMC11003657 DOI:
10.1093/dmfr/twae005]
[Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 02/04/2024] [Accepted: 02/15/2024] [Indexed: 02/21/2024] Open
Abstract
OBJECTIVES
Automating the digital workflow for diagnosing impacted canines using panoramic radiographs (PRs) is challenging. This study explored feature extraction, automated cropping, and classification of impacted and nonimpacted canines as a first step.
METHODS
A convolutional neural network with SqueezeNet architecture was first trained to classify two groups of PRs (91with and 91without impacted canines) on the MATLAB programming platform. Based on results, the need to crop the PRs was realized. Next, artificial intelligence (AI) detectors were trained to identify specific landmarks (maxillary central incisors, lateral incisors, canines, bicuspids, nasal area, and the mandibular ramus) on the PRs. Landmarks were then explored to guide cropping of the PRs. Finally, improvements in classification of automatically cropped PRs were studied.
RESULTS
Without cropping, the area under the curve (AUC) of the receiver operating characteristic (ROC) curve for classifying impacted and nonimpacted canine was 84%. Landmark training showed that detectors could correctly identify upper central incisors and the ramus in ∼98% of PRs. The combined use of the mandibular ramus and maxillary central incisors as guides for cropping yielded the best results (∼10% incorrect cropping). When automatically cropped PRs were used, the AUC-ROC improved to 96%.
CONCLUSIONS
AI algorithms can be automated to preprocess PRs and improve the identification of impacted canines.
Collapse