Imaging Sci Dent. 2023 Sep;53(3):199-208. English.
Published online Aug 02, 2023.
Copyright © 2023 by Korean Academy of Oral and Maxillofacial Radiology
Original Article

Determining the reliability of diagnosis and treatment using artificial intelligence software with panoramic radiographs

Kaan Orhan,1 Ceren Aktuna Belgin,2 David Manulis,3 Maria Golitsyna,3 Seval Bayrak,4 Secil Aksoy,5 Alex Sanders,3 Merve Önder,1 Matvey Ezhov,3 Mamat Shamshiev,3 Maxim Gusarev,3 and Vladislav Shlenskii3
    • 1Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Ankara University, Ankara, Turkey.
    • 2Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Hatay Mustafa Kemal University, Hatay, Turkey.
    • 3Diagnocat Inc., San Francisco, CA, USA.
    • 4Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Abant İzzet Baysal University, Bolu, Turkey.
    • 5Department of Dentomaxillofacial Radiology, Faculty of Dentistry, Near East University, Nicosia, Cyprus.
Received May 24, 2023; Revised July 07, 2023; Accepted July 10, 2023.

This is an Open Access article distributed under the terms of the Creative Commons Attribution Non-Commercial License (http://creativecommons.org/licenses/by-nc/3.0) which permits unrestricted non-commercial use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

Purpose

The objective of this study was to evaluate the accuracy and effectiveness of an artificial intelligence (AI) program in identifying dental conditions using panoramic radiographs (PRs), as well as to assess the appropriateness of its treatment recommendations.

Material and Methods

PRs from 100 patients (representing 4497 teeth) with known clinical examination findings were randomly selected from a university database. Three dentomaxillofacial radiologists and the Diagnocat AI software evaluated these PRs. The evaluations were focused on various dental conditions and treatments, including canal filling, caries, cast post and core, dental calculus, fillings, furcation lesions, implants, lack of interproximal tooth contact, open margins, overhangs, periapical lesions, periodontal bone loss, short fillings, voids in root fillings, overfillings, pontics, root fragments, impacted teeth, artificial crowns, missing teeth, and healthy teeth.

Results

The AI demonstrated almost perfect agreement (exceeding 0.81) in most of the assessments when compared to the ground truth. The sensitivity was very high (above 0.8) for the evaluation of healthy teeth, artificial crowns, dental calculus, missing teeth, fillings, lack of interproximal contact, periodontal bone loss, and implants. However, the sensitivity was low for the assessment of caries, periapical lesions, pontic voids in the root canal, and overhangs.

Conclusion

Despite the limitations of this study, the synthesized data suggest that AI-based decision support systems can serve as a valuable tool in detecting dental conditions, when used with PR for clinical dental applications.

Keywords
Artificial Intelligence; Radiography, Panoramic; Deep Learning; Dentistry

Introduction

Artificial intelligence (AI) is a scientific field that is focused on replicating human neurological processes, such as problem-solving, object recognition, and decision-making, through machines.1 Machine learning, a subset of AI, draws inferences from the information stored in a database. The more data it analyzes, the more it improves its future performance.2

Neural networks were among the earliest types of AI algorithms developed. The computational power of these networks hinges on the quality and volume of training data, which enable the networks to refine their connections. Networks with a simple structure and only a few layers are referred to as “shallow” learning neural networks, while those with numerous and larger layers are termed “deep” learning neural networks.3 Deep learning is a subset of machine learning. Deep learning constructs known as convolutional neural networks (CNNs) analyze and categorize data, thereby supporting deep learning capabilities, much like neurons in the human brain. They are primarily utilized to process large and intricate images.4, 5

One AI software application, Diagnocat (Diagnocat Inc., San Francisco, CA, USA), employs a deep CNN that incorporates U-Net-like and Mask R-CNN architectural models for segmentation and diagnosis.6 The U-Net model is known for its semantic segmentation capabilities, while the Mask R-CNN model is utilized for instance segmentation. The latter enables the measurement of the classification performance of individual objects using classification metrics. U-Net-like models are considered state-of-the-art architectures for object detection and segmentation tasks.7

Deep learning methods have found applications in a variety of medical fields. These include determining pediatric bone age from direct radiographs,8 detecting breast cancer in mammography,9 identifying skin cancer,10 and detecting brain tumors.11 CNNs can be utilized in dental radiology on images with complex layers, such as cone-beam computed tomography scans, as well as on panoramic and periapical radiographs.12, 13, 14 In the field of dentistry, AI has been employed for a range of tasks. These include tooth numbering, assessing the relationships of third molars to the mandibular canal, planning dental implants, evaluating periapical pathologies, detecting dental caries, evaluating osteoporotic changes, and examining jaw tumors.13, 15, 16, 17, 18, 19, 20, 21, 22, 23

The objective of this study was to evaluate the accuracy and efficacy of an AI program in identifying dental conditions using panoramic radiographs (PRs), as well as to assess the appropriateness of its treatment recommendations.

Materials and Methods

This study received approval from the ethics committee of the Near East University Faculty of Medicine, under protocol number YDU/2022/99-1473. The research was conducted in line with the principles outlined in the Declaration of Helsinki.

In this retrospective study, 100 PRs (representing 4497 teeth) of patients between 20 and 67 years old, taken between October 2017 and January 2020, were randomly selected from the Near East University Faculty of Dentistry’s database. The patients exhibited a variety of dental conditions, including caries, missing teeth, temporomandibular disorders, periodontal disease, and so on. Clinical examination records of the selected patients were carefully secured to verify the existence of the dental issues identified on the radiographs. To increase the complexity of the program, the study included panoramic images of patients with multiple dental conditions. Patients with fixed prosthetics, implants, caries, periapical lesions, furcation lesions, missing or restored teeth, root canal treatment, and periodontal disease were included. However, images of patients with mixed dentition, radiographs of insufficient quality for diagnostic purposes, and images including artifacts were excluded.

Imaging and evaluation

Regarding PR image generation, a Planmeca CC Proline device (Planmeca, Helsinki, Finland) was used to obtain digital radiographs. The radiographs were acquired in strict adherence to a standardized protocol, which involved stabilizing the natural head position with ear rods, as recommended by the manufacturer. Images were exposed at a peak kilovoltage of 70–84 kVp and a current of 12–14 mA, for 12–16 s each at a magnification of ×1.25.

All PR data were anonymized in Digital Imaging and Communications in Medicine format. Three oral and maxillofacial radiologists, with varying levels of clinical experience (7 years, 12 years, and 15 years), evaluated the images. Prior to the study, research calibration was conducted via videoconference. A list of dental states to be evaluated on the radiographs was also prepared (Table 1). Subsequently, the PRs were uploaded to the deep CNN software (Diagnocat Inc.). The diagnosis and treatment plans for each participant were then reassessed by comparing them with the issues identified by the observers (Fig. 1).

Fig. 1
A representative panoramic image shows a diagnosis automatically generated by Diagnocat.

Table 1
List of evaluated dental states

Model for pathology detection on panoramic images

Detecting pathologies on PRs can pose a challenge for AI, given that each image presents an entire set of teeth. At this scale, the pathologies are minute and can be difficult to detect. Furthermore, the process of annotating all pathologies is intricate and time-consuming, resulting in a training dataset with numerous missing labels.

The proposed solution was a pipeline involving 2-stage detectors. One of these was used to identify areas of interest, another to prepare pseudo-labels, and the last one to predict pathology. In addition to PRs, an annotated dataset of spot X-rays was utilized to enhance the base dataset.

The total dataset contained a substantial quantity of partially annotated data, which were not immediately usable. This is because based on these data, the model would be unable to differentiate between true negatives and false negatives and could fail to learn to identify poorly labeled pathologies. To overcome this issue, this study suggests using the model’s predictions, trained on a subset of data, as pseudo-labels. These pseudo-labels are then incorporated into the final model during training. In this study, a Mask R-CNN architectural model24 was utilized for labeling and segmentation tasks. This model was equipped with multiple heads, each trained to detect a specific group of labels. Each label group was defined in such a way that the data containing this set of labels were fully annotated. Therefore, during training, each sample was processed only through the heads that predicted the labels annotated on that particular image. The prediction of the model was presented as a set of pathologies, each accompanied by a probability, bounding box, and mask.

The model employed to identify regions of interest was separately trained for tooth detection. The training dataset was composed of 4500 images featuring annotated teeth, inclusive of missing teeth. The objective of these models was to identify teeth by segmenting their masks and defining their numbering. A 2-stage detector, Mask R-CNN, was utilized for this task, with a pre-trained ResNet-10125 serving as the backbone. The predictions generated by the trained model were used to designate the mouth area as the region of interest. The model’s output consisted of bounding boxes and segmentation masks with predicted numbers for each tooth. The coordinates of the mouth area were determined by the minimum and maximum values of the x and y coordinates for all detected teeth, expanded by a selected number of pixels (Fig. 2).

Fig. 2
Image illustrating the predictions made by the trained model, used to define the mouth region as the area of interest within the panoramic radiograph.

This model was trained using approximately 5430 partially annotated panoramic images from various manufacturers, featuring diverse pathologies. The dataset was further expanded by an additional 18000 intraoral X-rays. The model was designed to detect 21 different states. Initially, all panoramic images were supplemented with pseudo-labels from previously trained models. Each image was then cropped based on predictions from the ROI detector, and these cropped images were subsequently fed into the model. The chosen architecture for the model was the Cascade R-CNN,26 which differs from the Mask R-CNN in that it iteratively refines box prediction and takes class predictions as the average outcome for each cascade layer. This approach improves the prediction quality and reduces overfitting. To enhance the generalization of the model and improve performance, various augmentations were applied to the input data. These included randomized crop, rotation, brightness, contrast, image downscaling, blur, and noise; optical distortion; grid distortion; and contrast-limited adaptive histogram equalization.27 For X-ray images, this study also utilized mosaic augmentation, which produced images as depicted in Figure 3. Mosaic augmentation is a technique employed in dental radiography to enhance object detection. It involves extracting foreground ROIs from sparse samples and combining them to form a new image. To address scale variation, chips are magnified and suitable regions are chosen using sliding windows, with the zoom factor serving as a guide. This augmentation is applied to all training samples, while general samples do not require rescaling. Mosaic augmentation introduces complex backgrounds, which assists in object detection in a variety of contexts. It significantly enhances accuracy and robustness in the analysis of dental radiographs.28

Fig. 3
Mosaic data augmentation teaches the model to recognize objects in different locations, eliminating the need for a single specific context. This method of data augmentation combines numerous training images into a single image, using random proportions.

In the inference process, a panoramic image must initially undergo ROI detection, similar to the data preprocessing stage. Subsequently, it is processed through the cascade model. The predictions of the model are then calibrated and presented as the final output.

Statistical analysis

To evaluate the statistical reliability of the Diagnocat reports, the analyses of the 3 investigators were compared. If 2 or 3 investigators made the same diagnosis, that diagnosis was accepted as the ground truth. Then, the compatibility of the Diagnocat results with this ground truth was evaluated using the kappa test. The sensitivity and specificity assessments were analyzed for each pathology. Kappa values were interpreted as follows: less than 0 indicated no agreement, 0–0.20 slight agreement, 0.21–0.40 fair agreement, 0.41–0.60 moderate agreement, 0.61–0.80 substantial agreement, and 0.81-1 almost perfect agreement. Statistical analysis was performed with SPSS Statistics (IBM Corp., Armonk, NY, USA). The significance level was set at 0.05, and P-values of less than 0.05 were considered to indicate significant agreement.

Results

A total of 4497 teeth with panoramic radiographs from 100 patients were analyzed by all 3 researchers. As depicted in Table 2, instances where both the software and the researchers reached the same diagnosis are labeled as “correctly diagnosed (true positive).” Situations where the researchers made a correct diagnosis but the AI misdiagnosed the condition are referred to as “misdiagnosed (false negative).” Conversely, cases where the researchers incorrectly diagnosed the condition but the AI correctly identified the issue are termed “over-diagnosed (false positive).” Table 3 presents a compatibility chart between the researchers and the software, detailing the detection of all dental conditions for each dental state.

Table 2
Sensitivity and specificity of assessment using Diagnocat software

Table 3
Kappa values for inter-observer assessments and comparisons to the ground truth for each condition evaluated

The AI demonstrated almost perfect agreement (above 0.81) with the ground truth in most assessments (Table 2). The sensitivity was very high (exceeding 0.8) for the assessment of healthy teeth, artificial crowns, dental calculus, missing teeth, fillings, lack of interproximal contact, missing teeth, periodontal bone loss, and implants. Substantial sensitivity (above 0.6) was observed for cast posts and posts, short fillings, open margins, furcation lesions, and impacted teeth (Table 2).

The sensitivity was found to be low for the assessment of caries, periapical lesions, pontics, voids in the root canal filling, and overhangs (Table 2).

Statistical analysis revealed high kappa values for all inter-examiner assessments, demonstrating the strong reliability of the ground truth and the reproducibility of the reports (Table 3). The statistical evaluation indicated good reliability between the ground truth and AI software results, with the exception of assessments for caries (kappa: 0.365), periapical lesions, (kappa: 0.540), pontics (kappa: 0.517), overhangs (kappa: 0.604), and furcation lesions (kappa: 0.361) (Table 3).

Discussion

The rise in AI software usage in medical fields has extended to dentistry, with new programs being developed specifically for this sector. This study examined one such software, Diagnocat, and its capacity to detect various dental conditions in PR. The software demonstrated the highest accuracy in identifying dental states associated with missing teeth, while its sensitivity was lowest for missed canals. The Diagnocat program also correctly identified primary teeth in 2 patients, although these findings were not included in the statistical analysis. Previous research has indicated that deep learning methods supported by CNNs are highly acceptable in areas such as classification, scanning, and segmentation in medical image analysis.29 Various CNN architectures have been employed across different imaging modalities in the literature. Those with a U-Net-like architecture are typically favored for segmentation and volumetric data.30 The literature also suggests that software utilizing a U-Net-like architecture yields impressive results in image segmentation. However, training the U-Net network requires an excessive number of parameters.31 Diagnocat AI software employs a deep CNN with U-Net-like and Mask R-CNN architectural models for diagnosis.6

Various published studies have utilized Diagnocat software to explore different aspects of dental practice.15, 16, 17 One such study investigated the ability of deep CNNs to detect periapical lesions, and the results demonstrated success.15 In another study, Orhan et al.16 examined the effectiveness of AI in diagnosing impacted third molars, assessing their relationship with neighboring anatomical structures, and determining the number of root canals. Using cone-beam computed tomography images as the gold standard, the study indicated that the deep CNN method was successful in detecting impacted third molars and evaluating their relationship with anatomical structures. Kurt Bayrakdar et al.17 evaluated the use of AI in implant planning. In bone height measurements, they found no statistical difference between manual measurements and those made by AI. Consequently, they suggested that AI applications should be developed for implant planning. Ezhov et al.32 conducted a study to test the clinical performance, accuracy, and time efficiency of Diagnocat in diagnosing anatomical landmarks and pathologies. They also evaluated the clinical effectiveness and reliability of these diagnoses. The study concluded that Diagnocat software could potentially enhance the decision-making processes of clinicians.

In their study, Lee et al.18 utilized the GoogLeNet Inception v3 CNN network for preprocessing and transfer learning. The research yielded caries diagnostic accuracy levels of 89.0% for the premolar model, 88.0% for the molar model, and 82.0% for a model incorporating both premolars and molars. In a separate study, Kositbowornchai et al.33 employed a probabilistic neural network design to evaluate vertical root fracture. They reported that the highest sensitivity, specificity, and accuracy rates for detecting vertical root fractures were 98%, 90.5%, and 95.7%, respectively.

The efficacy of AI has been explored in distinct areas, such as the identification of dental caries,18 jaw tumors,20 and root fractures,33 as well as the examination of root canal morphology,34 periodontal diseases,35 and periapical lesions.15 However, no study has yet consolidated these topics into a single comprehensive investigation. Another strength of the present research is the preference for patient radiographs with established clinical examination results. This approach allowed us to compare the accuracy of AI diagnoses with actual clinical findings.

In the literature, various studies have explored the use of CNN architecture in different dental imaging modalities. Lin et al.36 examined the capacity of AI to identify and categorize teeth on dental bitewing radiographs, while Hosntalab et al.37 applied the technology to multi-slice computed tomography images. Chen et al.14 investigated the use of AI in detecting and numbering teeth on periapical radiographs. Arik et al.38 studied the role of AI in defining anatomical landmarks on cephalometric radiographs, reporting a high success rate in landmark detection. The present study found the lowest reliability in the assessment of caries, periapical lesions, pontic voids in the root canal, and overhangs. The diagnosis of caries is a clinical judgment concerning the presence of the disease. The detection of caries lesions involves identifying signs of caries, either clinically or radiographically. The use of a dental probe can enhance visual evaluation by providing tactile feedback. However, no consensus exists on the benefits of probing, as this action can cause the weakened enamel to collapse and the lesion to progress.39, 40 Furthermore, a dental probe can potentially transport bacteria from one surface to another, serving as a source of contamination. Radiographic detection involves identifying a radiolucency, interpreted as a caries lesion, on a dental radiograph.40 Visual evaluation of the approximal surfaces of the posterior teeth can be challenging. However, bitewing radiography can offer valuable information about the caries status of the approximal surfaces of the tooth crowns.39, 40, 41 Nonetheless, in bitewing radiographs, the proximal surfaces can overlap due to angulation errors, and the status of the occlusal surfaces remains unclear.42

AI techniques, such as Bayesian networks, artificial neural networks, or deep learning, can be utilized to develop computer-based systems. These systems can assist physicians in making decisions about their patients’ care. The increasing volume of data resulting from digitization in the radiology workflow provides an opportunity to maximize the use of machine learning techniques. This study evaluated the effectiveness of AI in detecting various dental conditions, including caries.

Lee et al.18 evaluated the efficacy of the CNN algorithm in detecting dental caries in periapical radiographs, finding it to be highly accurate. The study revealed that the AI system was significantly more sensitive than dentist evaluation, while its specificity was not significantly lower.43 A systematic review also evaluated the neural networks used in detecting and diagnosing dental caries. The review examined the conditions of these studies and their variable parameters, but the authors were unable to draw definitive conclusions. The definition of detected caries was not detailed in all studies, and not all specified the type of caries. Each study included in the review utilized a different neural network. These variations complicated the conclusions about the subject and the reliability, or lack thereof, of a neural network in the detection and diagnosis of caries.43, 44

The inability to conduct a histopathological examination to verify caries diagnosis was a notable limitation of the present study. This limitation applies to the assessments made by both observers and AI. In the present study, PR was the preferred imaging method due to its common use in dental radiographic examinations. PR allows for the scanning of a large anatomical region while requiring a relatively low dose of radiation. However, the quality and magnification of dental PR images can vary based on the patient's positioning.45 Therefore, the use of trans-hospital or hybrid datasets from multiple machines and conditions is important for achieving meaningfully high accuracy in the clinical application of deep learning.46 Notably, in this study, PRs were obtained from a single device in a single center, which constitutes a second limitation of this research.

In conclusion, despite the limitations of the study, the synthesized data suggest that AI-based decision support systems can serve as a valuable tool in detecting dental conditions, using PR for clinical dental applications.

Notes

Conflicts of Interest:The authors have no disclosures to report regarding funding, disclaimer statements, presentations of the research at conferences or symposia, or postings of the work on a preprint server, website, or other location. Kaan Orhan serves as a Scientific Advisor for Diagnocat Inc., based in San Francisco, CA, USA.

References

    1. Chartrand G, Cheng PM, Vorontsov E, Drozdzal M, Turcotte S, Pal CJ, et al. Deep learning: a primer for radiologists. Radiographics 2017;37:2113–2131.
    1. Hung K, Montalvao C, Tanaka R, Kawai T, Bornstein MM. The use and performance of artificial intelligence applications in dental and maxillofacial radiology: a systematic review. Dentomaxillofac Radiol 2020;49:20190107
    1. Burt JR, Torosdagli N, Khosravan N, RaviPrakash H, Mortazi A, Tissavirasingham F, et al. Deep learning beyond cats and dogs: recent advances in diagnosing breast cancer with deep neural networks. Br J Radiol 2018;91:20170545
    1. Shin HC, Roth HR, Gao M, Lu L, Xu Z, Nogues I, et al. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging 2016;35:1285–1298.
    1. Kim DH, MacKinnon T. Artificial intelligence in fracture detection: transfer learning from deep convolutional neural networks. Clin Radiol 2018;73:439–445.
    1. Shelhamer E, Long J, Darrell T. Fully convolutional networks for semantic segmentation. IEEE Trans Pattern Anal Mach Intell 2017;39:640–651.
    1. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional networks for biomedical image segmentation. arXiv:1505.04597 [Preprint]. 2015 [cited 2023 May 20].
    1. Lee H, Tajmir S, Lee J, Zissen M, Yeshiwas BA, Alkasab TK, et al. Fully automated deep learning system for bone age assessment. J Digit Imaging 2017;30:427–441.
    1. Becker AS, Marcon M, Ghafoor S, Wurnig MC, Frauenfelder T, Boss A. Deep learning in mammography: diagnostic accuracy of a multipurpose image analysis software in the detection of breast cancer. Invest Radiol 2017;52:434–440.
    1. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017;542:115–118.
    1. Rezaei M, Yang H, Meinel C. Deep neural network with l2-norm unit for brain lesions detection. arXiv:1708.05221 [Preprint]. 2017 [cited 2023 May 20].
    1. Miki Y, Muramatsu C, Hayashi T, Zhou X, Hara T, Katsumata A, et al. Classification of teeth in cone-beam CT using deep convolutional neural network. Comput Biol Med 2017;80:24–29.
    1. Tuzoff DV, Tuzova LN, Bornstein MM, Krasnov AS, Kharchenko MA, Nikolenko SI, et al. Tooth detection and numbering in panoramic radiographs using convolutional neural networks. Dentomaxillofac Radiol 2019;48:20180051
    1. Chen H, Zhang K, Lyu P, Li H, Zhang L, Wu J, et al. A deep learning approach to automatic teeth detection and numbering based on object detection in dental periapical films. Sci Rep 2019;9:3840
    1. Orhan K, Bayrakdar IS, Ezhov M, Kravtsov A, Özyürek T. Evaluation of artificial intelligence for detecting periapical pathosis on cone-beam computed tomography scans. Int Endod J 2020;53:680–689.
    1. Orhan K, Bilgir E, Bayrakdar IS, Ezhov M, Gusarev M, Shumilov E. Evaluation of artificial intelligence for detecting impacted third molars on cone-beam computed tomography scans. J Stomatol Oral Maxillofac Surg 2021;122:333–337.
    1. Kurt Bayrakdar S, Orhan K, Bayrakdar IS, Bilgir E, Ezhov M, Gusarev M, et al. A deep learning approach for dental implant planning in cone-beam computed tomography images. BMC Med Imaging 2021;21:86
    1. Lee JH, Kim DH, Jeong SN, Choi SH. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J Dent 2018;77:106–111.
    1. Lee JS, Adhikari S, Liu L, Jeong HG, Kim H, Yoon SJ. Osteoporosis detection in panoramic radiographs using a deep convolutional neural network-based computer-assisted diagnosis system: a preliminary study. Dentomaxillofac Radiol 2019;48:20170344
    1. Poedjiastoeti W, Suebnukarn S. Application of convolutional neural network in the diagnosis of jaw tumors. Health Inform Res 2018;24:236–241.
    1. Song IS, Shin HK, Kang JH, Kim JE, Huh KH, Yi WJ, et al. Deep learning-based apical lesion segmentation from panoramic radiographs. Imaging Sci Dent 2022;52:351–357.
    1. Widyaningrum R, Candradewi I, Aji NR, Aulianisa R. Comparison of Multi-Label U-Net and Mask R-CNN for panoramic radiograph segmentation to detect periodontitis. Imaging Sci Dent 2022;52:383–391.
    1. Khazaei M, Mollabashi V, Khotanlou H, Farhadian M. Sex determination from lateral cephalometric radiographs using an automated deep learning convolutional neural network. Imaging Sci Dent 2022;52:239–244.
    1. He K, Gkioxari G, Dollar P, Girshick R. Mask R-CNN. IEEE Trans Pattern Anal Mach Intell 2020;42:386–397.
    1. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. arXiv: 1512.03385 [Preprint]. 2017 [cited 2023 May 20].
    1. Zhaowei C, Vasconcelos N. Cascade R-CNN: Delving into high quality object detection. arXiv: 1712.00726 [Preprint]. 2018 [cited 2023 May 20].
    1. Mishra A. Contrast limited adaptive histogram equalization (CLAHE) approach for enhancement of the microstructures of friction stir welded joints. arXiv: 2109.00886 [Preprint]. 2021 [cited 2023 May 20].
    1. Wei Z, Duan C, Song X, Tian Y, Wang H. AMRNet: chips augmentation in aerial images object detection. arXiv:2009.07168 [Preprint]. 2020 [cited 2023 May 20].
    1. Anwar SM, Majid M, Qayyum A, Awais M, Alnowami M, Khan MK. Medical image analysis using convolutional neural networks: a review. J Med Syst 2018;42:226
    1. Cicek O, Abdulkadir A, Lienkamp SS, Brox T, Ronneberger O. 3D U-Net learning dense volumetric segmentation from sparse annotation. arXiv: 1606.06650 [Preprint]. 2016 [cited 2023 May 20].
    1. Minnema J, van Eijnatten M, Hendriksen AA, Liberton N, Pelt DM, Batenburg KJ, et al. Segmentation of dental cone-beam CT scans affected by metal artifacts using a mixed-scale dense convolutional neural network. Med Phys 2019;46:5027–5035.
    1. Ezhov M, Gusarev M, Golitsyna M, Yates JM, Kushnerev E, Tamimi D, et al. Clinically applicable artificial intelligence system for dental diagnosis with CBCT. Sci Rep 2021;11:15006
    1. Kositbowornchai S, Plermkamon S, Tangkosol T. Performance of an artificial neural network for vertical root fracture detection: an ex vivo study. Dent Traumatol 2013;29:151–155.
    1. Hiraiwa T, Ariji Y, Fukuda M, Kise Y, Nakata K, Katsumata A, et al. A deep learning artificial intelligence system for assessment of root morphology of the mandibular first molar on panoramic radiography. Dentomaxillofac Radiol 2019;48:20180218
    1. Kim J, Lee HS, Song IS, Jung KH. DeNTNet: Deep Neural Transfer Network for the detection of periodontal bone loss using panoramic dental radiographs. Sci Rep 2019;9:17615
    1. Lin PL, Lai YH, Huang PW. An effective classification and numbering system for dental bitewing radiographs using teeth region and contour information. Pattern Recognit 2010;43:1380–1392.
    1. Hosntalab M, Aghaeizadeh Zoroofi R, Abbaspour Tehrani-Fard A, Shirani G. Classification and numbering of teeth in multi-slice CT images using wavelet-Fourier descriptor. Int J Comput Assist Radiol Surg 2010;5:237–249.
    1. Arık SÖ, Ibragimov B, Xing L. Fully automated quantitative cephalometry using convolutional neural networks. J Med Imaging (Bellingham) 2017;4:014501
    1. Machiulskiene V, Campus G, Carvalho JC, Dige I, Ekstrand KR, Jablonski-Momeni A, et al. Terminology of dental caries and dental caries management: consensus report of a workshop organized by ORCA and Cariology Research Group of IADR. Caries Res 2020;54:7–14.
    1. Mathur VP, Dhillon JK. Dental caries: a disease which needs attention. Indian J Pediatr 2018;85:202–206.
    1. Pitts NB, Zero DT, Marsh PD, Ekstrand K, Weintraub JA, Ramos-Gomez F, et al. Dental caries. Nat Rev Dis Primers 2017;3:17030
    1. Yılmaz H, Keleş S. Recent methods for diagnosis of dental caries in dentistry. Meandros Med Dent J 2018;19:1–8.
    1. Cantu AG, Gehrung S, Krois J, Chaurasia A, Rossi JG, Gaudin R, et al. Detecting caries lesions of different radiographic extension on bitewings using deep learning. J Dent 2020;100:103425
    1. Prados-Privado M, García Villalón J, Martínez-Martínez CH, Ivorra C, Prados-Frutos JC. Dental caries diagnosis and detection using neural networks: a systematic review. J Clin Med 2020;9:3579
    1. Devlin H, Yuan J. Object position and image magnification in dental panoramic radiography: a theoretical analysis. Dentomaxillofac Radiol 2013;42:29951683
    1. Hwang JJ, Jung YH, Cho BH, Heo MS. An overview of deep learning in the field of dentistry. Imaging Sci Dent 2019;49:1–7.

Metrics
Share
Links to
Figures

1 / 3

Tables

1 / 3

PERMALINK