Abstract
Evaluating the mandibular canal proximity is crucial for planning mandibular third molar extractions. Panoramic radiography is commonly used for radiological examinations before third molar extraction but has limitations in assessing the true contact relationship between the third molars and the mandibular canal. Therefore, the true relationship between the mandibular canal and molars can be determined only through additional cone-beam computed tomography (CBCT) imaging. In this study, we aimed to develop an automatic diagnosis method based on a deep learning model that can determine the true proximity between the mandibular canal and third molars using only panoramic radiographs. A total of 901 third molars shown on panoramic radiographs were examined with CBCT imaging to ascertain whether true proximity existed between the mandibular canal and the third molar by two radiologists (450 molars: true contact, 451 molars: true non-contact). Three deep learning models (RetinaNet, YOLOv3, and EfficientDet) were developed, with performance metrics of accuracy, sensitivity, and specificity. EfficientDet showed the highest performance, with an accuracy of 78.65%, sensitivity of 82.02%, and specificity of 75.28%. The proposed deep learning method can be helpful when clinicians must evaluate the proximity of the mandibular canal and a third molar using only panoramic radiographs without CBCT.
Similar content being viewed by others
Introduction
The extraction of third molars is the most common surgical procedure performed by dentists and maxillofacial surgeons. Third molar extractions are typically recommended for reasons such as dental caries, periodontal disease, pain, infection, and damage to adjacent teeth1. However, the procedure can carry some risks and complications, including swelling, infection, and osteomyelitis. These risks are particularly heightened when a mandibular third molar is located near the mandibular canal, which can result in nerve damage such as nerve paralysis and numbness in the lower lip, gingiva, or jaw2,3. The rate of inferior alveolar nerve damage due to third molar extraction has been reported to range from 0.44 to 8.4%5, with permanent nerve damage reportedly occurring in fewer than 1% of cases6,7. Nerve damage can lead to medico-legal disputes and reduced quality of life for the patient. As such, it is crucial to accurately evaluate the relationship between the tooth and the adjacent mandibular canal prior to performing the surgical procedure.
Panoramic radiographs are used in dental screening to assess various oral issues, as they offer a comprehensive view of the entire oral and maxillofacial region. Moreover, this modality is widely used in dentistry because it involves low radiation exposure compared to other types of radiological images. In a previous study investigating the relationship between the third molar and the inferior alveolar nerve on panoramic radiographs, the accuracy of evaluation of true contact relationships was low, ranging from 52.68 to 69.64%, when assessed by six oral and maxillofacial surgery specialists8. Beginners may have lower accuracy and higher variability in diagnostic ability. On panoramic images, the superior and inferior borders of the mandibular canal typically appear as radiopaque lines; however, these borders may be only partially visible or not visible at all9. When the mandibular canal is unclear, it can be challenging to locate it or to distinguish it from other structures. Panoramic radiographs are two-dimensional (2D) images, leading to limitations in accurately assessing the positions and relationships of certain structures. Additionally, structures may appear to overlap, making interpretation difficult at times. In particular, molars situated buccal or lingual to the mandibular canal may overlap on panoramic radiographs, causing the appearance of contact between the mandibular canal and the molar. Furthermore, even when the mandibular canal is clearly visible, the true contact relationship should be evaluated using additional cone-beam computed tomography (CBCT) images10. However, compared to panoramic radiographs, CBCT involves a higher dose of ionizing radiation, higher costs, and longer imaging times, necessitating caution with pregnant women or young patients. The interpretation of CBCT images also requires specialized training to enable dentists to make the correct diagnosis11.
With advances in artificial intelligence technology over the past few years, numerous studies have focused on the automatic detection, classification, and segmentation of various anatomical structures in medical12 and dental images using deep learning models13,14,15. In particular, in the dental domain, deep learning has become a popular diagnostic tool for automatically identifying lesions on panoramic radiographs16,17. Kwon et al.16 developed a deep convolution neural network that automatically diagnosed jaw cysts and tumors with 95.6% accuracy on panoramic radiographs. Ha et al.17 reported an artificial intelligence model that automatically detected mesiodens with 96.2% accuracy on panoramic radiographs. Liu et al.18 reported an accuracy of 93.3% in evaluating the contact relationship between the mandibular canal and the third molar on CBCT using ResNet-34.
We focused on the principle that objects closer to the imaging plate show distinct images (for example, when evaluating periodontal disease, the lingual bone height is seen more clearly)9. It was hypothesized that deep learning could make it possible to diagnose the actual contact relationship between the mandibular canal and the third molar using only panoramic radiographs at the pixel level. In this study, we aimed to develop three deep learning models that enable clinicians to determine the true contact relationship between the third molar and the mandibular canal using only panoramic radiographs, without the need for CBCT images prior to extraction. If the true contact relationship can be automatically diagnosed through artificial intelligence using solely panoramic radiographs, it could provide a second opinion for clinicians in their diagnosis and reduce the risk of nerve damage.
Materials and methods
Data preparation
This study was approved by the Institutional Review Board of Yonsei University Health System, Severance Hospital (No. 2-2022-0071), and informed consent was waived due to its retrospective nature. The research was carried out in compliance with pertinent guidelines and ethical regulations, with all data anonymized prior to transmission to the investigators to prevent patient identification. All images were collected from images taken from January 2021 to December 2022, and the image review period was from January to March 2023.
For algorithm development, panoramic radiograph data were obtained from six instruments at four different institutions. These instruments included Cranex 3+ (Soredex Orion Co., Helsinki, Finland), RAYSCAN Alpha (Ray Co., Ltd., Hwaseong-si, Korea), and PaX-i3D (Vatech Co., Ltd., Hwaseong-si, Korea) at Yonsei University Dental Hospital, as well as Osstem-T1-CS (Osstem Co., Ltd., Seoul, Korea), PaX-i3D OP X-ray (Vatech Co., Ltd.), and PHT-35LHS (Vatech Co., Ltd.) at three dental clinics. A total of 901 third molars were assessed from 518 patients who underwent both panoramic radiography and CBCT. All data were classified into Group A (contact) and Group B (non-contact) by two oral radiologists with 25 and 20 years of clinical experience, respectively, according to the true contact relationship between the third molar and the mandibular canal using CBCT images, regardless of the findings on panoramic radiography. Two radiologists evaluated the images after calibration with 10% of the total data (90 third molars) evaluation. In 10% of the total data set, the intra-rater reliability was 0.952 and 0.946. The inter-rater reliability was 0.927. Group A consisted of cases where the third molar and the mandibular canal were in true contact, as evidenced by both panoramic radiographs and CBCT images. Group B consisted of cases in which the third molar and the mandibular canal had no true contact, including both cases in which the third molar and mandibular canal appeared to have no contact and cases where they appeared to be in contact on the panoramic radiographs (Fig. 1). Group A contained 450 third molars and Group B included 451 third molars. The datasets were split into 6:2:2 ratios for training, validation, and test (Table 1).
Development of three deep learning models
The collected images were in bitmap format, and the size ranged from 1976 to 2988 pixels in width and from 976 to 1500 pixels in height. All images were resized to 1024 (width) and 1024 (height) pixels for model training. An oral radiologist manually labeled the bounding boxes with a size range from 145 to 380 pixels in width and from 200 to 405 pixels in height, which sufficiently included the apex or crown of the third molar and the mandibular canal, using LabelImg software (ver 1.8.4, available at https://github.com/tzutalin/labelImg) as an annotation tool. Additionally, if the mandibular canal was not clearly visible on the panoramic radiographs, it was identified through CBCT images and labeled to include the mandibular canal. These labels served as gold standards for network training. The annotation files contained information about the class names (contact or non-contact) and bounding boxes used to describe the location of the third molar and mandibular canal. Figure 2 shows the overall process of developing a deep learning-based diagnostic system for the third molar and mandibular canal relationship.
The detection and classification model of the relationship between the third molar and mandibular canal was developed by using three deep learning algorithms (RetinaNet, YOLOv3, and EfficientDet) on panoramic radiographs. RetinaNet19 is designed to detect features of various sizes in images by employing ResNet50 and the feature pyramid network (FPN) model, a multi-scale feature pyramid, as its backbone. The feature maps for each pyramid level extracted from the backbone are input into the classification subnetwork and bounding box regression subnetwork, respectively, to detect the target. This model was the first for which the use of a focal loss function, focused on class imbalance problems to improve object detection performance, was proposed. In YOLOv320, darknet 53 is used as the backbone and binary cross-entropy as the loss function, demonstrating rapid inference time and robust performance. This algorithm is widely used in medical imaging. It was trained to automatically detect the relationship between the third molar and mandibular canal on panoramic radiographs by using 53 convolutional layers and the leaky rectified linear unit (ReLU) activation function, which generated three feature maps with different resolutions. EfficientDet21 employs EfficientNet-B0 to B6 as the backbone, focusing on improving accuracy and efficiency with fewer parameters than previous detectors. In the present study, the EfficientDet-D4 model was used due to computer resource limitations. Similar to RetinaNet, this model uses a focal loss function but employs the bidirectional FPN (BiFPN) model to detect objects of various sizes when extracting features from images. Unlike the conventional top-down pathway FPN, a bottom-up pathway was added for bidirectional use, and weights for different features were considered by connecting input and output nodes. Behind the BiFPN layer, a classification sub-network and a box prediction subnetwork are used to detect objects (Fig. 3).
All models were trained using transfer learning on the COCO dataset22 for 300 epochs with a batch size of 2 and momentum of 0.9. During training, predicted bounding boxes were considered correct if the detected region had an intersection-over-union value of 0.5 or greater compared to the annotation. During testing, when the trained model detected the third molar and mandibular canal, the detected area on the panoramic radiograph was marked with a box and class name. Group A (contact) is shown in blue boxes, while Group B (non-contact) is shown in yellow boxes. If no third molar was detected, the input image was returned without a box. All networks were implemented in Python 3 using the Keras library with a TensorFlow backend and trained on an NVIDIA Titan RTX graphics card (NVIDIA, Santa Clara, CA, USA).
Evaluation of the three deep learning models
The three models successfully detected the third molars and mandibular canals for all test data sets. Consequently, the performance of these models was evaluated using the classification performance indicators of accuracy, sensitivity, and specificity. The accuracy, sensitivity, and specificity were determined based on the classification of the relationship between the third molar and the mandibular canal, using the following equations:
where TP, TN, FP, and FN represent true positive, true negative, false positive, and false negative, respectively.
Results
Table 2 illustrates the diagnostic performance of the three implemented models in relation to the association between the third molar and the mandibular canal. EfficientDet-D4 showed an accuracy of 78.65%, a sensitivity of 82.02%, and a specificity of 75.28%. YOLOv3 showed an accuracy of 73.03%, a sensitivity of 82.02%, and a specificity of 64.04%, and RetinaNet showed the highest specificity (88.76%). RetinaNet, YOLOv3, and EfficientDet-D4 incorrectly predicted Group A (contact) samples as Group B (non-contact) at rates of 69.66%, 17.98%, and 17.98%, respectively. In contrast, the rates of incorrectly predicting Group B (non-contact) samples as Group A (contact) for RetinaNet, YOLOv3, and EfficientDet-D4 were 11.24%, 35.96%, and 24.72%, respectively, with a higher number of contact cases observed on the panoramic radiographs (Fig. 4). Figure 5 displays correctly predicted examples from the three models using the test dataset.
Discussion
We developed three deep learning models to determine the true contact relationship between the third molar and the mandibular canal using only panoramic radiographs, with the gold standard being two radiologists’ evaluations of the contact relationship from CBCT. Compared to specialists and deep learning in previous studies, our study demonstrated superior accuracy in automatically diagnosing the true proximity of the third molar and mandibular canal on panoramic radiographs. Conducted using images from various devices across multiple institutions, this study is believed to be highly applicable in clinical practice. Dentists perform extractions several times a day, and extracting mandibular third molars is a challenging task due to their proximity to the mandibular canal. Before extraction, the closeness of the mandibular canal and a third molar is primarily assessed through panoramic radiographs. Identifying the mandibular canal on these images can be difficult, and due to the limitations of 2D imaging, two structures separated by the buccal and lingual sides may appear to overlap and make contact. Diagnosing true proximity on panoramic radiographs without CBCT is extremely challenging, even for specialists.
Several studies have recently attempted to automatically diagnose the contact relationship between third molars and the mandibular canal using deep learning on panoramic radiographs8,23,24. Our study is particularly robust, as we collected data from six devices across multiple institutions and included all possible cases to improve real-world usability. Furthermore, our models were developed to diagnose true contact on original panoramic radiographs rather than cropped images, resulting in improved accuracy compared to previous studies. Fukuda et al.23 conducted an experiment in which they cropped 600 panoramic radiographs to 70 \(\times\) 70 and 140 \(\times\) 140 pixels, and they classified contact and non-contact groups of the third molar and mandibular canal on images using AlexNet, GoogLeNet, and VGG16. The diagnostic performance was highest, at 0.92, after training with 70 \(\times\) 70 pixels on GoogLeNet. Although the accuracy was as high as 0.92, which exceeded that of the present study, that experiment was performed using panoramic radiographs cropped to include the apex of the third molar and the mandibular canal as input images. This approach has a major drawback in that it is unknown whether the third molar truly comes into contact with the mandibular canal. Choi et al.8 proposed the ResNet50 model to determine the true contact relationship between the third molar and the mandibular canal using 571 panoramic radiographs and compared the performance of the artificial intelligence model to that of oral and maxillofacial specialists. In that study, true contact analysis was classified using only cases in which the third molar and mandibular canal overlapped on images. The accuracy of the deep learning model was 0.63 for true contact and 0.76 for the buccolingual position (non-contact), constituting better performance than six specialists (52.68–69.64% and 32.26–51.61%, respectively). However, the mandibular canal may not be identified on such images, and some clinicians may not accurately evaluate the contact between the third molar and the mandibular canal on an image. Sukegawa et al.24 reported a deep learning model using ResNet50 and RestNet50v2, which was developed to classify the contact state between the mandibular canal and the third molar. The dataset included 1279 panoramic radiographs cropped to 250 \(\times\) 200 pixels. Two experiments were conducted: contact analysis on panoramic image and true contact analysis between the third molar and the mandibular canal. In both sets of experimental results, the ResNet50v2 model showed the best performance, with accuracies of 0.860 and 0.766, respectively. The true contact analysis was similar to our study but differed in that it involved manually cropped images. In our study, the entire image was used without manual cropping, and RetinaNet, YOLOv3, and EfficientDet-D4, which differ from the models used in previous studies, were employed. Among the three models, EfficientDet demonstrated the highest accuracy at 78.65%, which exceeded the result of the true contact analysis by Sukegawa et al. (76.6%).
In the present study, EfficientDet-D4 demonstrated the highest accuracy, at 78.65%, for determining the relationship between the third molar and the mandibular canal, while YOLOv3 and RetinaNet had accuracies of 73.03% and 59.55%, respectively. Among the three models, RetinaNet had the lowest accuracy (59.55%), but it exhibited the highest specificity at 88.76%. Overall, we observed that the sensitivity (i.e., correctly predicting true contact) was higher than the specificity (i.e., correctly predicting non-contact) on panoramic radiographs. Cases where actual non-contact is incorrectly predicted as contact are not overly problematic from a practical standpoint, as the clinician can be particularly careful when extracting the third molar, but predicting contact as non-contact may increase the risk of nerve damage. Therefore, high sensitivity is more important than high specificity.
In cases where CBCT access is limited, the model developed in this study can provide a second opinion about the actual proximity of the third molar and the mandibular canal using only panoramic radiographs.
The present study had several strengths. First, we fully automated the analysis of the true contact relationship between the third molar and the mandibular canal using only panoramic radiographs and a deep learning model. Unlike previous studies utilizing cropped images, our algorithm was developed with original panoramic radiographs, accounting for the real clinical environment. Second, we developed models reflecting the characteristics of real-world data by using images acquired from various devices across multiple institutions. This showcases the potential for application in the dental field. However, the accuracy of the current model reaches only up to 78.65%, indicating that further efforts are necessary to enhance the accuracy above 80% by adjusting parameters and investigating other deep learning models.
Conclusion
Various deep learning models have been developed to automatically identify the true contact relationship between the third molar and the mandibular canal on panoramic radiographs. By utilizing the automatic true proximity analysis model developed here, it is feasible to determine the true contact relationship between the third molar and the mandibular canal using only panoramic radiographs, without the need for CBCT images. This algorithm can offer diagnostic assistance to inexperienced dentists.
Data availability
The data generated and analyzed in this study are not publicly available due to Korean privacy laws and policies, but they are available from the corresponding author upon reasonable request.
References
Chu, F. et al. Prevalence of impacted teeth and associated pathologies-a radiographic study of the Hong Kong Chinese population. Hong Kong Med. J. (2003).
Sukegawa, S. et al. What are the risk factors for postoperative infections of third molar extraction surgery: A retrospective clinical study-?. Med. Oral Patol. Oral Cir. Bucal. 24, e123. https://doi.org/10.4317/medoral.22556 (2019).
Nguyen, E., Grubor, D. & Chandu, A. Risk factors for permanent injury of inferior alveolar and lingual nerves during third molar surgery. J. Oral Maxillofac. Surg. 72, 2394–2401. https://doi.org/10.1016/j.joms.2014.06.451 (2014).
Alling, C. C. III. Dysesthesia of the lingual and inferior alveolar nerves following third molar surgery. J. Oral Maxillofac. Surg. 44, 454–457. https://doi.org/10.1016/S0278-2391(86)80010-6 (1986).
Lopes, V., Mumenya, R., Feinmann, C. & Harris, M. Third molar surgery: An audit of the indications for surgery, post-operative complaints and patient satisfaction. Br. J. Oral Maxillofac. Surg. 33, 33–35. https://doi.org/10.1016/0266-4356(95)90083-7 (1995).
Wofford, D. T. & Miller, R. I. Prospective study of dysesthesia following odontectomy of impacted mandibular third molars. J. Oral Maxillofac. Surg. 45, 15–19. https://doi.org/10.1016/0278-2391(87)90080-2 (1987).
Bruce, R. A., Frederickson, G. C. & Small, G. S. Age of patients and morbidity associated with mandibular third molar surgery. J. Am. Dent. Assoc. 101, 240–245. https://doi.org/10.14219/jada.archive.1980.0183 (1980).
Choi, E. et al. Artificial intelligence in positioning between mandibular third molar and inferior alveolar nerve on panoramic radiography. Sci. Rep. 12, 2456. https://doi.org/10.1038/s41598-022-06483-2 (2022).
White, S. C. & Pharoah, M. J. White and Pharoah’s Oral Radiology: Principles and Interpretation 201–327 (Elsevier Health Sciences, 2018).
Scarfe, W. C., Farman, A. G. & Sukovic, P. Clinical applications of cone-beam computed tomography in dental practice. J. Can. Dent. Assoc. 72, 75 (2006).
Brown, J. et al. Basic training requirements for the use of dental CBCT by dentists: a position paper prepared by the European Academy of DentoMaxilloFacial Radiology. Dentomaxillofac. Radiol. 43, 20130291. https://doi.org/10.1259/dmfr.20130291 (2014).
Shen, D., Wu, G. & Suk, H.-I. Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 19, 221–248. https://doi.org/10.1146/annurev-bioeng-071516-044442 (2017).
Jeon, K. J., Ha, E.-G., Choi, H., Lee, C. & Han, S.-S. Performance comparison of three deep learning models for impacted mesiodens detection on periapical radiographs. Sci. Rep. 12, 15402. https://doi.org/10.1038/s41598-022-19753-w (2022).
Ha, E.-G. et al. Automatic diagnosis of retention pseudocyst in the maxillary sinus on panoramic radiographs using a convolutional neural network algorithm. Sci. Rep. 13, 2734. https://doi.org/10.1038/s41598-023-29890-5 (2023).
Choi, H. et al. Deep learning-based fully automatic segmentation of the maxillary sinus on cone-beam computed tomographic images. Sci. Rep. 12, 1–9. https://doi.org/10.1038/s41598-022-18436-w (2022).
Kwon, O. et al. Automatic diagnosis for cysts and tumors of both jaws on panoramic radiographs using a deep convolution neural network. Dentomaxillofac. Radiol. 49, 20200185. https://doi.org/10.1259/dmfr.20200185 (2020).
Ha, E.-G., Jeon, K. J., Kim, Y. H., Kim, J.-Y. & Han, S.-S. Automatic detection of mesiodens on panoramic radiographs using artificial intelligence. Sci. Rep. 11, 23061. https://doi.org/10.1038/s41598-021-02571-x (2021).
Liu, M.-Q. et al. Deep learning-based evaluation of the relationship between mandibular third molar and mandibular canal on CBCT. Clin. Oral Investig. https://doi.org/10.1007/s00784-021-04082-5 (2022).
Lin, T.-Y., Goyal, P., Girshick, R., He, K. & Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV) 2980–2988. (2017).
Redmon, J. & Farhadi, A. Yolov3: An incremental improvement. arXiv preprint https://doi.org/10.48550/arXiv.1804.02767 (2018).
Tan, M., Pang, R. & Le, Q. V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (CVPR) 10781–10790 (2020).
Lin, T.-Y. et al. Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6–12, 2014, Proceedings, Part V 13 740–755 (Springer). https://doi.org/10.1007/978-3-319-10602-1_48.
Fukuda, M. et al. Comparison of 3 deep learning neural networks for classifying the relationship between the mandibular third molar and the mandibular canal on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 130, 336–343. https://doi.org/10.1016/j.oooo.2020.04.005 (2020).
Sukegawa, S. et al. Deep learning model for analyzing the relationship between mandibular third molar and inferior alveolar nerve in panoramic radiography. Sci. Rep. 12, 16925. https://doi.org/10.1038/s41598-022-21408-9 (2022).
Acknowledgements
This work was supported by a grant from the National Research Foundation of Korea, funded by the Korean government (NRF-2022R1A2B5B01002517).
Author information
Authors and Affiliations
Contributions
All authors gave their final approval and agreed to be accountable for all aspects of the work. S.H. proposed the ideas; K.J., H.C. collected data; K.J., H.C., and C.L. contributed to data analysis; H.C. designed the deep learning model; K.J., H.C., C.L., and S.H. drafted and critically revised the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Jeon, K.J., Choi, H., Lee, C. et al. Automatic diagnosis of true proximity between the mandibular canal and the third molar on panoramic radiographs using deep learning. Sci Rep 13, 22022 (2023). https://doi.org/10.1038/s41598-023-49512-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-023-49512-4
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.