Next Article in Journal
Water–Fat Separated T1 Mapping in the Liver and Correlation to Hepatic Fat Fraction
Next Article in Special Issue
The U-Net Approaches to Evaluation of Dental Bite-Wing Radiographs: An Artificial Intelligence Study
Previous Article in Journal
Diagnostic Algorithm for Surgical Management of Limbal Stem Cell Deficiency
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Deep Learning-Based Approach for Segmentation of Different Type Caries Lesions on Panoramic Radiographs

by
Burak Dayı
1,
Hüseyin Üzen
2,
İpek Balıkçı Çiçek
3 and
Şuayip Burak Duman
4,*
1
Department of Restorative Dentistry, Faculty of Dentistry, Inonu University, Malatya 44280, Turkey
2
Department of Computer Engineering, Bingol University, Bingol 12000, Turkey
3
Department of Biostatistics and Medical Informatics, Faculty of Medicine, Inonu University, Malatya 44280, Turkey
4
Department of Oral and Maxillofacial Radiology, Faculty of Dentistry, Inonu University, Malatya 44280, Turkey
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(2), 202; https://doi.org/10.3390/diagnostics13020202
Submission received: 4 December 2022 / Revised: 30 December 2022 / Accepted: 3 January 2023 / Published: 5 January 2023

Abstract

:
The study aims to evaluate the diagnostic performance of an artificial intelligence system based on deep learning for the segmentation of occlusal, proximal and cervical caries lesions on panoramic radiographs. The study included 504 anonymous panoramic radiographs obtained from the radiology archive of Inonu University Faculty of Dentistry’s Department of Oral and Maxillofacial Radiology from January 2018 to January 2020. This study proposes Dental Caries Detection Network (DCDNet) architecture for dental caries segmentation. The main difference between DCDNet and other segmentation architecture is that the last part of DCDNet contains a Multi-Predicted Output (MPO) structure. In MPO, the final feature map split into three different paths for detecting occlusal, proximal and cervical caries. Extensive experimental analyses were executed to analyze the DCDNet network architecture performance. In these comparison results, while the proposed model achieved an average F1-score of 62.79%, the highest average F1-score of 15.69% was achieved with the state-of-the-art segmentation models. These results show that the proposed artificial intelligence-based model can be one of the indispensable auxiliary tools of dentists in the diagnosis and treatment planning of carious lesions by enabling their detection in different locations with high success.

1. Introduction

Dental caries is one of the most common chronic diseases and affects more than three billion people worldwide [1]. Clinicians need to detect caries at the enamel and dentin levels before these turn into irreversible lesions [2]. Dental radiographs are an essential diagnostic tool for helping clinicians to diagnose caries. One type of extra-oral radiograph is the digital panoramic, which is widely used in many areas of dentistry and provides safer, more accurate and relatively cheaper results today [3]. Contrary to intra-oral radiographs, it has minimal spatial resolution and is open to significant, unpredictable geometric distortion [4]. For this and many other reasons (different tooth anatomical morphologies and restorative forms, etc.), even though dental radiography and explorer (also known as a dental probe) are often employed and considered very reliable diagnostic techniques for the identification of dental caries, a large proportion of screening and final diagnosis frequently relies on empirical evidence [5]. However, this approach is highly subjective and in crowded clinics mistakes or underdiagnoses may happen depending on the knowledge and focus of the clinicians. This can be avoided by integrating artificial intelligence software with radiographs to assist the clinician during the diagnosis and treatment phases.
Artificial intelligence (AI) is a generic phrase used to describe the development of computer systems that can carry out functions that typically require human intellect. The applications of AI in dentistry are primarily virtual, using AI algorithms to separate lesions from healthy structures, rank risk variables and simulate and assess future outcomes [6]. Convolutional neural networks (CNNs) have been effectively used in recent years to clarify numerous healthcare difficulties involving various forms of medical imaging. Skin cancer classification from dermoscopic pictures, breast cancer detection from thermal images, Alzheimer’s disease diagnosis using SPECT data and automated identification and quantification of COVID-19 from chest computed tomography records are a few new techniques employing CNNs [7,8,9,10].
Radiology is a key part of the diagnostic process in dentistry. Every year, a considerable number of images, including panoramic, bitewing, periapical and cephalometric radiographs, are acquired in dental radiology [11]. Given this massive number of picture records, CNNs appear to have enormous clinical evaluation and diagnostic potential. Deep learning researchers have just begun to investigate this potential in the realm of dental radiography. CNNs have been utilized effectively in periapical radiographs to diagnose periodontal bone loss [12], bitewing radiographs to detect carious lesions [13] and panoramic radiographs to detect apical lesions [14]. A thorough examination of the use of CNNs in dental radiology is provided. Furthermore, it is feasible to envision how the application of artificial intelligence in dentistry might yield gratifying results, particularly in the field of caries diagnostics [15].
Briefly, in connection with various medical and dental practices, the deep learning system, one of the most promising artificial intelligence models, has been developed [16,17]. Deep learning systems can automatically classify datasets and, with the aid of multilayer CNNs, they can learn in-depth about the features present in the data [18].
Tooth segmentation is crucial for the automated detection of tooth-related disorders on dental radiologic images, but manual annotation is a laborious and time-consuming operation. Therefore, the first (and most difficult) step in creating automated interpretable diagnostic procedures for dental images is the automation of tooth segmentation. CNN has recently been widely used in dentistry to overcome the constraints associated with traditional segmentation algorithms [19,20,21].
The application of CNNs in dentistry, particularly for the detection of caries, is a novel technique [13]. The aim of this study is to develop a new deep-learning system for the segmentation of different types of carious lesions on panoramic radiographs. In this direction, the Dental Caries Detection Network (DCDNet) architecture was developed to detect different types of caries lesions and performance was compared with other existing models.

2. Material and Methods

2.1. Image Dataset

Inonu University Non-Interventional Clinical Research Ethics Board (number 2022/3774) authorized the study protocol. The principles of the Helsinki Declaration were followed. The study included anonymous panoramic radiographs obtained from the radiology archive of Inonu University Faculty of Dentistry’s Department of Oral and Maxillofacial Radiology from January 2018 to January 2020. The age range of the patients to whom the radiographs belonged was between 14 and 80. The percentage of males and females was 29.5 and 70.5, respectively. Radiographs were obtained using the Planmeca Promax 2D Panoramic System (Planmeca, Helsinki, Finland) with image acquisition parameters of 68 kVp, 14 mA and 12 s.

2.2. Image Evaluation and Labeling

Radiographs with artifacts and in which no caries could be detected were not included in the analysis. Occlusal caries detected on the panoramic radiographs were labeled as Type I, proximal caries as Type II and caries in the cervical region as Type III with the Plainsight Software System (San Francisco, CA, USA). Labeling was performed with the joint decision of an 8-year restorative dental specialist and a 5-year oral and maxillofacial radiologist. In case of instability, the relevant images were excluded from the study. As a result, 504 panoramic radiography images were used in experimental studies and 746 occlusal caries (Type I), 1627 proximal caries (Type II) and 378 cervical caries (Type III) labels were made in these images (Figure 1).

2.3. Construction of the Models

Pre-trained network architectures are used to ensure the high performance of our proposed model. In this study, MobileNetV2, VGG16, ResNet50 and EfficientNet and Inception network architectures, which have recently provided high performance, have been used. In addition, the decoder module of the proposed model was inspired by the Feature Pyramid Network. The architectures used are briefly described below.

2.3.1. MobileNetV2

MobileNetV1 is a convolutional architecture designed for low-cost or mobile devices that minimizes network cost and size. This has resulted in the ease of use of image processing and deep network categorization in mobile devices. The MobileNetV2 model was based on the MobileNetV1 and tackled difficulties relating to nonlinearities in the model’s thin layers containing building blocks [22]. The MobileNetV2 model can perform classification, segmentation and object identification and adds two additional features to its predecessor. The first is that some bottlenecks might appear linearly between layers; the second is the development of shortcuts between bottlenecks [23].
The MobileNetV2 design includes depth wise (dw) separable filters and combination stages. This model employs a deep convolution filter for each layer input with a resolution of 1 × 1 pixel. Depth wise separable convolutional filters investigate inputs by dividing them into two distinct layers. This reduces both the model’s speed and cost. The features obtained by filter separation are combined in the combining stages and a new layer is formed. Batchnorm and Rectified Linear Unit (ReLU) linearity are used in constructing the MobilNetV2 model [24].

2.3.2. VGG16

In 2014, Zisserman and Simonyan presented VGG16 as a VGGNET network structure. It is a more extensive network constructed on top of the AlexNet network. It can more correctly describe the data collection properties while recognizing and classifying images and outperforms other methods when dealing with big datasets and complicated backdrop recognition tasks. The network topology consists of three fully connected layers, 13 convolutional layers and five pool layers. Compared to other networks, the VGG16’s 13 convolutional layers employ a medium-sized 3 × 3 matrix with a moving step of 1. The number of convolution kernels steadily rose from 64 in the first layer to 128 to 256 in the second layer and then to 512 in the final layer. The convolution kernel in the pooling layer has a size of 2 × 2 and a step size of 2. With a convolution kernel size of 5 × 5, it outperforms other networks in terms of extracted features [25].

2.3.3. ResNet50

The name ResNet is an abbreviation for residual neural networks. It is an improved version of CNN with a large number of convolutional neural networks. ResNet tries to solve saturation and accuracy loss in the deep CNN training process [26]. ResNet50 is a residual network with 50 layers. Having transitions between layers deepens the network in ResNet models. The deterioration that may occur in the deepening network is prevented thanks to these transitions. Moreover, ResNet models employ a mixture of multiple-sized convolution filters to counteract degradation and minimize training time due to deep structures. These models use blocks called bottlenecks for rapid training [27].

2.3.4. EfficientNet

EfficientNet is an algorithm that uses convolutional neural networks. EfficientNet has an architecture that focuses on improving the efficiency of models as well as their accuracy. EfficientNet consists of 8 different models between B0-B7. The model scales use three separate parameters. These parameters are depth, width and resolution. The depth parameter measures how deep the networks are, while the width parameter is the number of neurons in the layers. The resolution parameter expresses the resolution of the dataset on which the model will be trained. Within the scope of the study, the classification process was carried out with EfficientNet, unlike existing CNN models, which employ a novel activation function termed Swish rather than a ReLU activation function [28]. EfficientNet also provides more efficient outcomes in other cutting-edge models by consistently scaling depth, breadth and resolution while scaling down the model. The first stage in the compound scaling approach is to find a grid that will assist in determining the relationship between the various scaling dimensions of the baseline network while working with a fixed resource limitation. Thus, an appropriate scaling factor for the depth, breadth and resolution parameters is calculated. The coefficients are then applied to scale the baseline network to the desired target network [29].

2.3.5. Feature Pyramid Network

A Feature Pyramid Network (FPNet) is a fully convolutional feature extractor that takes as input a single-scale image of any size and outputs proportionally sized feature maps at multiple levels. This process does not depend on the architecture of the convolutional backbone, so may be used as a universal approach for building feature pyramids within deep convolutional networks for applications such as object recognition [30].
A bottom-up and a top-down path are used to build the pyramid. The bottom-up pathway is the backbone of ConvNet’s feedforward computation. It builds a feature hierarchy out of feature maps at different scales with a scaling step of 2. For each step of the feature pyramid, one pyramid level is defined. As a reference set of feature maps, the output of the last layer of each stage is used [31]. The top-down path gives the impression of higher resolution features by upsampling feature maps from higher pyramid levels that are geographically coarser but have better meaning. Then, using lateral connections, these traits are added to traits from the bottom-up pathway. Each lateral link is made up of equal-sized feature maps from the top-down and bottom-up pathways. The bottom-up feature map has lower-level meanings, but its activations have been more accurately localized because there was less subsampling [32].

2.4. Proposed Network Architecture

In this study, Dental Caries Detection Network (DCDNet) architecture is proposed for dental caries detection. This network architecture is encoder-decoder based as shown in Figure 2. The Encoder part consists of pre-trained backbone network architectures such as VGG16, MobileNet and EfficientNet. In the proposed model, initial features are obtained from these backbone networks. The decoder part of the proposed DCDNet network consists of two parts, the Multi-level Features Concatenation (MFC) module and the Multi-Predicted Output (MPO) Block. It is fed with features from different levels of the backbone network in the MFC model. In this section, new powerful feature maps are obtained for caries detection. The second part of the decoder generates the prediction map for the three caries types from the final feature map in the MPO block. For this process, the final feature map proceeds in three ways and convolution and sigmoid activation are applied.

2.4.1. Encoder

Pre-trained backbone networks are used in the encoder part of the proposed model. The backbone networks used in experimental studies are EfficientNet, ResNet50, VGG16, Inception-V3 and MobileNetV2 networks, respectively. Each backbone network has connection points, as shown in Table 1. Features taken from these ports are transferred to the decoder section.
The primary purpose of using backbone networks shown in Table 1 is to obtain initial features for a limited dataset. In this way, a more efficient training procedure takes place.

2.4.2. Decoder

The decoder of the proposed model consists of two stages; the first is the feature integration module, MFC. The MFC model was inspired by the FPNet model. As shown in Figure 2, the MFC module consists of four levels. ConvBNReLU blocks have been applied to the input features at the beginning of each level. Then the resulting output was combined with the skip connection. In the merging process, the Pointwise Convolution (PC) layer was primarily applied to the feature maps. In this way, the two feature maps were brought to the same size. The element-based sum operation was then applied to these feature maps. These processes were repeated for the four levels of the MFC.
At the last level of MFC, feature maps of all levels were combined. In this way, essential details obtained in the intermediate layers were preserved. The combined feature maps were transferred to the MPO unit to obtain dental caries prediction output.
The MPO block divided the final feature map into three paths, as shown in Figure 2. Each path represents one type of caries. Here, the final feature map, like Unet and FPN, were not used directly. The main reason for this is to prevent the interaction of caries types that are very closely related to each other. Thus, it detects with higher scores than other single-output architectures.

2.5. Training Procedure

In transfer learning architectures based on the CNN model, high-performance results are achieved through learning transfer. These models have provided significant improvements, especially in the field of image processing. The EfficientNet architecture and other deep learning models were trained using the transfer learning approach.
The initial parameters of the backbone networks used in the proposed DCDNet network were used with ImageNet-trained parameters. In addition, the other layers in the decoder section were initialized with random values. Detailed layer structures, connections and other details of the proposed DCDNet architecture are given in Table 2.
The Binary Cross-Entropy loss function was used in the training of the DCDNet network, defined as in Equation (1).
L k = i , j M y i × j l o g ( P i × j ) + ( 1 y i × j ) l o g ( 1 P i × j ) L = 1 N k N L k
Here, y i × j and P i × j represent the actual and predicted values of the pixel at position i × j, respectively. M and N represent the total number of samples and the number of caries types, respectively. The L k value shows the mean error value obtained for k-type ( k { 1 ,   2 ,   3 } ) caries and the L result error value.

3. Results

In this section, experimental studies have been carried out to analyze the caries detection performance of the proposed DCDNet network. Firstly, the backbone networks of the proposed model were compared. Then, the proposed model was compared with other models, such as Unet and FPNet.

3.1. Application Details

To evaluate the DCDNet network, the Dental Caries dataset produced in this study was used. Panoramic radiography images in the dataset are 900 × 1700 in size. As seen in Figure 1 in these high-resolution images, there are external parts, such as the other parts of the lower and upper jaw apart from the teeth. These external parts, except the teeth, are unimportant for caries detection. Therefore, the images were cropped at 540 × 1300 size with reference to the middle part of each image covering the teeth (Figure 3). Then, these cropped images were reduced to 256 × 512 size to give the input of the network architecture.
For the training phase, 75% of the panoramic radiography images that marked dental caries and their types were used. The remaining images were used for testing. The training and testing processes of the proposed model were carried out using the TensorFlow-Keras library in a python environment. In training the model, the Batch size was set to 8, the learning rate to 0.001 and the training epoch to 100. Adam’s optimization method was employed to update the network parameters. All experiments were performed under the Ubuntu 18.04 system using the Intel Xeon CPU, 128GB RAM and Nvidia P40 (24 GB).
In experimental studies, Precision , Recall, F1-score and mIoU (mean intersection over union), and metrics were used to evaluate the proposed model performance. This metric is defined as follows:
Precision = T P T P + F P Recall = T P T P + F N F 1 score = 2 × precision × recall precision + recall mIoU   = p r   G T p r     G T
where TP, FP and FN refer to true positives, false positives and false negatives, respectively. GT, pr, mIoU and F1-score denote the ground truth, the prediction map, mean intersection-over-union and F1-score metric, respectively.

3.2. Dental Caries Detection Results

The proposed DCDNet model consists of a pre-trained backbone network, MCF and MPO structures. Four different models were chosen as the backbone network. In this way, the highest performances are aimed for. The results obtained with the different backbone networks of the DCDNet network are given in Table 3. In addition, sample images from the dataset and the estimation results of the models are given in Figure 3.
As can be seen in Table 3, the highest performance was obtained with the ResNet50-DCDNet architecture for the average F1-score values. In addition, an average of F1-score (62.67%) was achieved when the EfficientNet backbone network, which has achieved successful performance in many areas and recent articles, was used. On the other hand, on average, the lowest F1-score was obtained with the VGG16-DCDNet network.
Detailed performance results of the proposed models are given in Table 4.
When considering each type of caries separately in Table 4, the methods yielded a precision, recall and F1-score of over 60% for Type I and Type II caries, while Type III gave generally low results.
In the image results given in Figure 3, it was observed that the proposed models produced effective results for Type I and Type II caries detection. However, the proposed DCDNet model for detecting Type III caries remained weak.

3.3. Comparison of the Proposed Model with Other Methods

This section compares the proposed DCDNet network with other Unet, FPNet, Mobile-UNet and Eff-Unet models in the literature. For comparisons, the same experimental studies were performed using Unet, FPNet, Mobile-UNet and Eff-Unet networks. The results obtained in the experimental studies are given in Table 5.
As can be seen in the results given in Table 5, Unet, FPNet, Mobile-UNet and Eff-Unet models produced results below 20%.

3.4. Comparison of Models in Terms of Time Consumption

The DCDNet network proposed in this study produced effective results for caries detection. As shown in the previous sections, it produced high performance against other models thanks to its three-output structure. In this experimental study, the comparison of the proposed model in terms of time consumption is structured. The times for each model to process an image are given in Table 6 in seconds and milliseconds.
As seen in Table 6, the time consumption of DCDNet models varies between 80.0 and 98.3 milliseconds. The fact that the proposed DCDNet models have three outputs requires more time compared to other models. On the other hand, single-output models Mobile-Unet, Unet, Eff-Unet and FPNet have become more economical. However, compared to the results given in Section 3.3, the success of these models is quite low. In comparison, the DCDNet models, although costlier in terms of time consumption, achieved much higher success (see Section 2 and Section 3) compared to other models.
In the time comparison of DCDNet models, the fastest model was the Mobilenet-DCDNet model. In addition, this model provided the highest scores for Type 1 caries detection, as seen in the results in Table 3. However, this model produced lower results than the ResNet50-DCDNet model in detecting other types of caries. On the other hand, the ResNet50-DCDNet model required 0.093 s to process an image. Although the ResNet50-DCDNet mode is costly in terms of time consumption, it is the model with the highest average F1 score (Table 3). As a result, less than 0.1 s is required to process an image on all recommended DCDNet models. These results are sufficient for the detection of dental caries.

4. Discussion

Accurate, fast and timely diagnosis of dental caries is vital for both the physician and the patient in terms of restoring teeth without further treatment. Advanced treatments, besides causing time and economic loss, may adversely affect the success and the duration of the tooth in the mouth. The availability of automatic detection software that will help and guide physicians in the detection of dental caries and other pathological conditions will add a different dimension to dentistry. In the present study, a deep learning model used to segment dental caries on panoramic radiographs successfully detected occlusal and proximal caries.
To evaluate the performance of artificial intelligence programs, studies on root morphology [18,33], automatic tooth detection and numbering caries detection, [34,35,36] and classification on panoramic radiographs [37,38] were carried out. Caries detection and classification studies with artificial intelligence in panoramic radiographs are limited in number. Lee et al. [39] detected dental anomalies, including the category of dental caries, using artificial intelligence on panoramic radiographs. The study divided dental caries into four groups according to their clinical features: dental caries, cervical caries, proximal caries and secondary caries. In this study, we examined caries in three groups: occlusal caries (Type I), proximal caries (Type II) and cervical caries (Type III).
In this study, MobileNetV2, VGG16, ResNet50 and EfficientNet and Inception network architectures, backbone networks of the DCDNet network architecture which is known to have highly successful performance in recent studies in the segmentation and classification of dental caries, were compared. As a result of the comparison, the ResNet50-DCDNet network architecture gave the highest F1-score (62.79%). Then, the ResNet50-DCDNet network architecture was compared with the Unet, FPNet, Mobile-UNet and Eff-Unet models. As a result, the compared models showed a very low F1-score (highest Eff-Unet: 15.69%) compared to the model we proposed. The main reason for this may be that these models, which are the only way of distinguishing between softmax and caries, cannot provide adequate detection by not being able to distinguish varieties of caries, since the types of caries in the study are similar to each other. The DCDNet network, whose development is based on this problem, produces separate outputs with an MPO block for each caries as shown in Figure 2. Thus, higher success was achieved by creating a separate mechanism for caries. Vinayahalingam et al. [38] achieved high success (accuracy of 0.87) with the MobileNet V2 deep learning model, which they used to classify carious lesions in third molar tooth images cropped from 253 panoramic X-rays. Zhu et al. [40] segmented shallow, moderate and deep caries in panoramic X-rays with Caries Net deep learning architecture. They achieved a mean 93.64% Dice coefficient and 93.61% accuracy at three different caries levels. Haghanifar et al. [41] detected dental caries on (470) panoramic X-ray images close to those in our study and achieved 90.52% caries detection accuracy. In our study, 70.79% F1-score success was found in the detection of occlusal caries and 67.65% F1-score in the detection of proximal caries with the ResNet50-DCDNet learning architecture. Lian et al. [37] compared the caries depths classified by expert dentists on 1160 panoramic X-rays with the efficiency of the artificial intelligence program. As a result of the study, they found the performance of the artificial intelligence program and expert dentists to be similar.
Studies show that even experienced dentists are not consistent in diagnosing proximal caries [42]. When early lesions are missed, the chance to perform minimally invasive procedures may be lost. Li et al. [43] marked 953 pits and fissures in 712 intraoral photographs and 1002 approximal caries. While they found over 88% sensitivity in pit and fissure caries and approximal caries detected by the deep learning-based prototype artificial intelligence system, this compares with an over 67% F1-score in the same type of caries in our study. One of the reasons for the 18.64% F1-score success rate in the detection of cervical caries in our study may be that the proposed DCDNet architecture could not learn this type of caries adequately due to the insufficiency of the dataset in this type, which we call type III. Another reason may be that panoramic radiographs are 2D images and this type of caries is usually located in the vestibule and lingual of the teeth, so it overlaps with the denser healthy tooth tissue and, at the same time, it can be challenging to detect because of overlapping with the radiolucent reflection of the pulp chamber on the radiograph.
The limitations of this study are the insufficient dataset, especially for cervical caries. In addition, the panoramic radiographs used in the study could not be supported by clinical examination of the patients. As a result, dental caries whose cavitation has just begun to form in the mouth may have been overlooked in the panoramic radiographs and some dental caries detected by artificial intelligence may have been evaluated as false positive by experts when they were present in the mouth, and vice versa. At the same time, the superposition of the teeth, which is frequently seen in panoramic radiographs, may have caused some dental caries not to be detected.
It is indisputable that the use of artificial intelligence as an auxiliary diagnostic tool in dentistry will be of great benefit in reducing the workload of the dentist and in making an accurate diagnosis. There is a need for further studies on the use of artificial intelligence in different areas of dentistry, its possible negative effects and the development of artificial intelligence programs and deep learning methods that will enable higher levels of success to be achieved.

5. Conclusions

Within the limitations of this study, while the deep learning-based artificial intelligence system successfully detected occlusal and proximal caries, it showed low performance in detecting cervical caries. More powerful datasets and new network architectures could enable the detection of cavities in different locations with higher success rates. Thus, deep learning-based artificial intelligence systems may become one of the favorite elements in dental clinics by increasing the success rate of dentists in diagnosis and treatment and providing dentists with ease of work.

Author Contributions

Conceptualization, B.D., H.Ü., İ.B.Ç. and Ş.B.D.; Methodology, B.D., H.Ü. and Ş.B.D.; Software, H.Ü. and İ.B.Ç.; Validation, İ.B.Ç; Formal Analysis, H.Ü. and İ.B.Ç.; Data Curation, H.Ü. and İ.B.Ç.; Writing—Original Draft Preparation, B.D., H.Ü., İ.B.Ç. and Ş.B.D.; Writing—Review & Editing, B.D., H.Ü., İ.B.Ç. and Ş.B.D.; Supervision, B.D., Ş.B.D.; Project Administration, B.D., Ş.B.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

All procedures followed were following the ethical standards of the responsible committee on human experimentation (institutional and national) and with the Helsinki Declaration of 1975, as revised in 2008. Inonu University Non-Interventional Clinical Research Ethics Board (decision number:2022/3774) was authorized the study protocol.

Informed Consent Statement

Informed consent was obtained from all patients being included in the study.

Data Availability Statement

Data is available from the corresponding author upon reasonable request.

Acknowledgments

We are grateful to Cemil Colak for his valuable contributions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kassebaum, N.J.; Smith, A.G.; Bernabé, E.; Fleming, T.D.; Reynolds, A.E.; Vos, T.; Murray, C.; Marcenes, W.; Collaborators, G.O.H. Global, regional, and national prevalence, incidence, and disability-adjusted life years for oral conditions for 195 countries, 1990–2015: A systematic analysis for the global burden of diseases, injuries, and risk factors. J. Dent. Res. 2017, 96, 380–387. [Google Scholar] [CrossRef] [PubMed]
  2. Baelum, V. What is an appropriate caries diagnosis? Acta. Odontol. Scand. 2010, 68, 65–79. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Keerthana, S.; Mohammed, H.; Hariprasad, A.; Anand, M.; Ayesha, S. Comparative evaluation of condylar guidance obtained by three different interocclusal recording materials in a semi-adjustable articulator and digital panoramic radiographic images in dentate patients: An in vivo study. J. Indian Prosthodont. Soc. 2021, 21, 397. [Google Scholar]
  4. Shah, N.; Bansal, N.; Logani, A. Recent advances in imaging technologies in dentistry. World J. Radiol. 2014, 6, 794. [Google Scholar] [CrossRef] [PubMed]
  5. Lee, J.-H.; Kim, D.-H.; Jeong, S.-N.; Choi, S.-H. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J. Dent. 2018, 77, 106–111. [Google Scholar] [CrossRef]
  6. Shan, T.; Tay, F.; Gu, L. Application of artificial intelligence in dentistry. J. Dent. Res. 2021, 100, 232–244. [Google Scholar] [CrossRef]
  7. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  8. Świetlik, D.; Białowąs, J. Application of artificial neural networks to identify alzheimer’s disease using cerebral perfusion SPECT data. Int. J. Environ. Res. Public Health 2019, 16, 1303. [Google Scholar] [CrossRef] [Green Version]
  9. Zhang, H.-T.; Zhang, J.-S.; Zhang, H.-H.; Nan, Y.-D.; Zhao, Y.; Fu, E.-Q.; Xie, Y.-H.; Liu, W.; Li, W.-P.; Zhang, H.-J. Automated detection and quantification of COVID-19 pneumonia: CT imaging analysis by a deep learning-based software. Eur. J. Nucl. Med. Mol. Imaging 2020, 47, 2525–2532. [Google Scholar] [CrossRef]
  10. Kakileti, S.T.; Madhu, H.J.; Krishnan, L.; Manjunath, G.; Sampangi, S.; Ramprakash, H. Observational study to evaluate the clinical efficacy of Thermalytix for detecting breast cancer in symptomatic and asymptomatic women. JCO Glob. Oncol. 2020, 6, 1472–1480. [Google Scholar] [CrossRef]
  11. Schwendicke, F.; Golla, T.; Dreher, M.; Krois, J. Convolutional neural networks for dental image diagnostics: A scoping review. J. Dent. 2019, 91, 103226. [Google Scholar] [CrossRef] [PubMed]
  12. Lee, J.-H.; Kim, D.-H.; Jeong, S.-N.; Choi, S.-H. Diagnosis and prediction of periodontally compromised teeth using a deep learning-based convolutional neural network algorithm. J. Periodontal Implant. Sci. 2018, 48, 114–123. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Bayraktar, Y.; Ayan, E. Diagnosis of interproximal caries lesions with deep convolutional neural network in digital bitewing radiographs. Clin. Oral Investig. 2022, 26, 623–632. [Google Scholar] [CrossRef] [PubMed]
  14. Ekert, T.; Krois, J.; Meinhold, L.; Elhennawy, K.; Emara, R.; Golla, T.; Schwendicke, F. Deep Learning for the Radiographic Detection of Apical Lesions. J. Endod. 2019, 45, 917–922.e5. [Google Scholar] [CrossRef]
  15. Schwendicke, F.; Elhennawy, K.; Paris, S.; Friebertshäuser, P.; Krois, J. Deep learning for caries lesion detection in near-infrared light transillumination images: A pilot study. J. Dent. 2020, 92, 103260. [Google Scholar] [CrossRef] [PubMed]
  16. Thakur, A.; Mishra, A.P.; Panda, B.; Rodríguez, D.; Gaurav, I.; Majhi, B. Application of artificial intelligence in pharmaceutical and biomedical studies. Curr. Pharm. Des. 2020, 26, 3569–3578. [Google Scholar] [CrossRef]
  17. Ossowska, A.; Kusiak, A.; Świetlik, D. Artificial intelligence in dentistry—Narrative review. Int. J. Environ. Res. Public Health 2022, 19, 3449. [Google Scholar] [CrossRef]
  18. Hiraiwa, T.; Ariji, Y.; Fukuda, M.; Kise, Y.; Nakata, K.; Katsumata, A.; Fujita, H.; Ariji, E. A deep-learning artificial intelligence system for assessment of root morphology of the mandibular first molar on panoramic radiography. Dentomaxillofac. Radiol. 2019, 48, 20180218. [Google Scholar] [CrossRef]
  19. Silva, G.; Oliveira, L.; Pithon, M. Automatic segmenting teeth in X-ray images: Trends, a novel data set, benchmarking and future perspectives. Expert Syst. Appl. 2018, 107, 15–31. [Google Scholar] [CrossRef] [Green Version]
  20. Lee, J.-H.; Han, S.-S.; Kim, Y.H.; Lee, C.; Kim, I. Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2020, 129, 635–642. [Google Scholar] [CrossRef]
  21. Duman, Ş.B.; Syed, A.Z.; Celik Ozen, D.; Bayrakdar, İ.Ş.; Salehi, H.S.; Abdelkarim, A.; Celik, Ö.; Eser, G.; Altun, O.; Orhan, K. Convolutional Neural Network Performance for Sella Turcica Segmentation and Classification Using CBCT Images. Diagnostics 2022, 12, 2244. [Google Scholar] [CrossRef]
  22. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  23. Toğaçar, M.; Cömert, Z.; Ergen, B. Intelligent skin cancer detection applying autoencoder, MobileNetV2 and spiking neural networks. Chaos Solitons Fractals 2021, 144, 110714. [Google Scholar] [CrossRef]
  24. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  25. Krishnaswamy Rangarajan, A.; Purushothaman, R. Disease classification in eggplant using pre-trained VGG16 and MSVM. Sci. Rep. 2020, 10, 2322. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  26. Dash, P.B.; Behera, H.; Senapati, M.R. Deep Learning Based Framework for Breast Cancer Mammography Classification Using Resnet50. In Proceedings of the International Conference on Computational Intelligence in Pattern Recognition, Las Vegas, NV, USA, 14–16 December 2022; pp. 625–633. [Google Scholar]
  27. Sharma, N.; Jain, V.; Mishra, A. An analysis of convolutional neural networks for image classification. Proc. Comput. Sci. 2018, 132, 377–384. [Google Scholar] [CrossRef]
  28. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; pp. 6105–6114. [Google Scholar]
  29. Atila, Ü.; Uçar, M.; Akyol, K.; Uçar, E. Plant leaf disease classification using EfficientNet deep learning model. Ecol. Inform. 2021, 61, 101182. [Google Scholar] [CrossRef]
  30. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  31. Zhao, G.; Ge, W.; Yu, Y. GraphFPN: Graph feature pyramid network for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 2763–2772. [Google Scholar]
  32. Li, X.; Lai, T.; Wang, S.; Chen, Q.; Yang, C.; Chen, R.; Lin, J.; Zheng, F. Weighted feature pyramid networks for object detection. In Proceedings of the 2019 IEEE International Conference on Parallel & Distributed Processing with Applications, Big Data & Cloud Computing, Sustainable Computing & Communications, Social Computing & Networking (ISPA/BDCloud/SocialCom/SustainCom), Xiamen, China, 16–18 December 2019; pp. 1500–1504. [Google Scholar]
  33. Duman, S.; Yılmaz, E.F.; Eşer, G.; Çelik, Ö.; Bayrakdar, I.S.; Bilgir, E.; Costa, A.L.F.; Jagtap, R.; Orhan, K. Detecting the presence of taurodont teeth on panoramic radiographs using a deep learning-based convolutional neural network algorithm. Oral Radiol. 2022, 39, 207–214. [Google Scholar] [CrossRef] [PubMed]
  34. Bilgir, E.; Bayrakdar, İ.Ş.; Çelik, Ö.; Orhan, K.; Akkoca, F.; Sağlam, H.; Odabaş, A.; Aslan, A.F.; Ozcetin, C.; Kıllı, M. An artifıcial intelligence approach to automatic tooth detection and numbering in panoramic radiographs. BMC Med. Imaging 2021, 21, 124. [Google Scholar] [CrossRef]
  35. Kılıc, M.C.; Bayrakdar, I.S.; Çelik, Ö.; Bilgir, E.; Orhan, K.; Aydın, O.B.; Kaplan, F.A.; Sağlam, H.; Odabaş, A.; Aslan, A.F. Artificial intelligence system for automatic deciduous tooth detection and numbering in panoramic radiographs. Dentomaxillofac. Radiol. 2021, 50, 20200172. [Google Scholar] [CrossRef]
  36. Estai, M.; Tennant, M.; Gebauer, D.; Brostek, A.; Vignarajan, J.; Mehdizadeh, M.; Saha, S. Deep learning for automated detection and numbering of permanent teeth on panoramic images. Dentomaxillofac. Radiol. 2022, 51, 20210296. [Google Scholar] [CrossRef]
  37. Lian, L.; Zhu, T.; Zhu, F.; Zhu, H. Deep learning for caries detection and classification. Diagnostics 2021, 11, 1672. [Google Scholar] [CrossRef] [PubMed]
  38. Vinayahalingam, S.; Kempers, S.; Limon, L.; Deibel, D.; Maal, T.; Hanisch, M.; Bergé, S.; Xi, T. Classification of caries in third molars on panoramic radiographs using deep learning. Sci. Rep. 2021, 11, 12609. [Google Scholar] [CrossRef]
  39. Lee, S.; Kim, D.; Jeong, H.-G. Detecting 17 fine-grained dental anomalies from panoramic dental radiography using artificial intelligence. Sci. Rep. 2022, 12, 5172. [Google Scholar] [CrossRef]
  40. Zhu, H.; Cao, Z.; Lian, L.; Ye, G.; Gao, H.; Wu, J. CariesNet: A deep learning approach for segmentation of multi-stage caries lesion from oral panoramic X-ray image. Neural Comput. Appl. 2022, 1–9. [Google Scholar] [CrossRef]
  41. Haghanifar, A.; Majdabadi, M.M.; Ko, S.-B. Paxnet: Dental caries detection in panoramic x-ray using ensemble transfer learning and capsule classifier. arXiv 2020, arXiv:2012.13666. [Google Scholar]
  42. Hellyer, P. AI in radiographic caries detection. Br. Dent. J. 2022, 232, 629. [Google Scholar] [CrossRef] [PubMed]
  43. Li, R.; Zhu, J.; Wang, Y.; Zhao, S.; Peng, C.; Zhou, Q.; Sun, R.; Hao, A.; Li, S.; Wang, Y. Development of a deep learning based prototype artificial intelligence system for the detection of dental caries in children. Zhonghua Kou Qiang Yi Xue Za Zhi 2021, 56, 1253–1260. [Google Scholar]
Figure 1. Example images (a,c,e,g) and labels (b,d,f,h). Diagnostics 13 00202 i001 Type I (Occlusal caries). Diagnostics 13 00202 i002 Type II (Proximal caries). Diagnostics 13 00202 i003 Type III (Cervical caries).
Figure 1. Example images (a,c,e,g) and labels (b,d,f,h). Diagnostics 13 00202 i001 Type I (Occlusal caries). Diagnostics 13 00202 i002 Type II (Proximal caries). Diagnostics 13 00202 i003 Type III (Cervical caries).
Diagnostics 13 00202 g001
Figure 2. Proposed deep architecture.
Figure 2. Proposed deep architecture.
Diagnostics 13 00202 g002
Figure 3. Sample images from the dataset and the prediction outputs of the models.
Figure 3. Sample images from the dataset and the prediction outputs of the models.
Diagnostics 13 00202 g003
Table 1. Backbone networks and skip connections used in the proposed DCDNet network.
Table 1. Backbone networks and skip connections used in the proposed DCDNet network.
BackboneSkip Connection 1Skip Connection 2Skip Connection 3Skip Connection 4
MobilenetV2Block 13 expands relublock 6 expand relublock 3 expand relublock 1 expand relu
Inception-V3Mixed7Mixed2Activation 5Activation 3
EfficientNetblock6a expand activationblock4a expand activationblock3a expand activationblock2a expand activation
ResNet50stage4 unit1 relu1stage3 unit1 relu1stage2 unit1 relu1relu0
VGG16block5 conv3block4 conv3block3 conv3block2 conv2
Table 2. Detailed layer structure and connections (layer inputs) of the DCDNet model (Sc: skip connection, f: filters, c: kernel, s: stride).
Table 2. Detailed layer structure and connections (layer inputs) of the DCDNet model (Sc: skip connection, f: filters, c: kernel, s: stride).
Layer NameSectionLayer InputApplied Layer ProcessDetails
L0EncoderImagePre-training NetworkDetails in Table 1
L1MCF-level 0Sc0Pointwise Convolutionf: 256, c: 3 × 3, s = 1 × 1
L2L1Conv2DTransposef: 256, c: 4 × 4, s = 2 × 2
L3MCF-level 1Sc1Pointwise Convolutionf: 256, c: 3 × 3, s = 1 × 1
L4L2, L3Concat-
L5L4Conv2DTransposef: 256, c: 4 × 4, s = 2 × 2
L6L4Convolution, Batch Normalization, ReLUf: 128, c: 3 × 3, s = 1 × 1
L7L6UpSampling2Ds = 8 × 8
L8MCF-level 2Sc2Pointwise Convolutionf: 256, c: 3 × 3, s = 1 × 1
L9L5, L8Concat-
L10L9Conv2DTransposef: 256, c: 4 × 4, s = 2 × 2
L11L9Convolution, Batch Normalization, ReLUf: 128, c: 3 × 3, s = 1 × 1
L12L11UpSampling2Ds = 4 × 4
L13MCF-level 3Sc3Pointwise Convolutionf: 256, c: 3 × 3, s = 1 × 1
L14L10, L13Concat-
L15L14Conv2DTransposef: 256, c: 4 × 4, s = 2 × 2
L16L14Convolution, Batch Normalization, ReLUf: 128, c: 3 × 3, s = 1 × 1
L17L16UpSampling2Ds = 2 × 2
L18MCF-level 4Sc4Pointwise Convolutionf: 256, c: 3 × 3, s = 1 × 1
L19L15, L18Concat-
L20L19Conv2DTransposef: 256, c: 4 × 4, s = 2 × 2
L21L19Convolution, Batch Normalization, ReLUf: 128, c: 3 × 3, s = 1 × 1
L22MPOL6, L11, L16, L19Concat-
L23L22Convolution, Batch Normalization, ReLUf: 32, c: 3 × 3, s = 1 × 1
L24L22Convolution, Batch Normalization, ReLUf: 32, c: 3 × 3, s = 1 × 1
L25L22Convolution, Batch Normalization, ReLUf: 32, c: 3 × 3, s = 1 × 1
L26Output 1L23Convolution, Sigmoidf: 1, c: 1 × 1, s = 1 × 1
L27Output 2L24Convolution, Sigmoidf: 1, c: 1 × 1, s = 1 × 1
L28Output 3L25Convolution, Sigmoidf: 1, c: 1 × 1, s = 1 × 1
Table 3. F1-score results obtained with different backbone networks of DCDNet architecture (%).
Table 3. F1-score results obtained with different backbone networks of DCDNet architecture (%).
DCDNet ModelsType I (Occlusal Caries)Type II (Proximal Caries)Type III (Cervical Caries)Weighted Averages of F1-Score
Mobilenet V273.4566.1511.2161.86
Inception-V369.2564.8614.2960.20
EfficientNet71.9368.0112.9062.67
ResNet5070.7967.6518.6462.79
VGG1664.8864.327.5557.76
Table 4. Detailed results of the proposed models for Type I, Type II and Type III.
Table 4. Detailed results of the proposed models for Type I, Type II and Type III.
DCDNet ModelsMetricsType I
(Occlusal Caries)
Type II (Proximal Caries)Type III
(Cervical Caries)
MobilenetV2Precision76.0271.4820.00
Recall71.0461.567.79
F1-Score73.4566.1511.21
Inception-V3Precision73.8972.8624.14
Recall65.1758.4510.14
F1-Score69.2564.8614.29
EfficientNetPrecision75.9374.5837.50
Recall68.3362.507.79
F1-Score71.9368.0112.90
ResNet50Precision72.0070.5529.73
Recall69.6164.9713.58
F1-Score70.7967.6518.64
VGG16Precision70.7871.5314.81
Recall59.8958.435.06
F1-Score64.8864.327.55
Table 5. F1-Score Performance comparison of DCDNet and other models.
Table 5. F1-Score Performance comparison of DCDNet and other models.
ModelsType I
(Occlusal Caries)
Type II (Proximal Caries)Type III
(Cervical Caries)
Weighted Averages of F1-Score
ResNet50-DCDNet70.7967.6518.6462.79
Unet12.9716.0801.8813.44
FPNet14.6811.0202.2411.10
Mobile-UNet18.3616.5300.2015.46
Eff-Unet17.6916.9100.4015.69
Table 6. Comparison of deep learning models in terms of time consumption (ms: milliseconds, s: seconds).
Table 6. Comparison of deep learning models in terms of time consumption (ms: milliseconds, s: seconds).
ModelsFPS (ms)FPS (s)
Mobile-UNet18.90.0189
Unet23.00.0230
Eff-Unet33.00.0330
FPNet37.50.0375
Mobilenet-DCDNet80.00.0800
VGG16-DCDNet85.70.0857
InceptionV3-DCDNet97.30.0973
EfficientNet-DCDNet97.90.0979
ResNet50-DCDNet98.30.0983
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dayı, B.; Üzen, H.; Çiçek, İ.B.; Duman, Ş.B. A Novel Deep Learning-Based Approach for Segmentation of Different Type Caries Lesions on Panoramic Radiographs. Diagnostics 2023, 13, 202. https://doi.org/10.3390/diagnostics13020202

AMA Style

Dayı B, Üzen H, Çiçek İB, Duman ŞB. A Novel Deep Learning-Based Approach for Segmentation of Different Type Caries Lesions on Panoramic Radiographs. Diagnostics. 2023; 13(2):202. https://doi.org/10.3390/diagnostics13020202

Chicago/Turabian Style

Dayı, Burak, Hüseyin Üzen, İpek Balıkçı Çiçek, and Şuayip Burak Duman. 2023. "A Novel Deep Learning-Based Approach for Segmentation of Different Type Caries Lesions on Panoramic Radiographs" Diagnostics 13, no. 2: 202. https://doi.org/10.3390/diagnostics13020202

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop