Next Article in Journal
Anatomical, Pathophysiological, and Clinical Aspects of Extra-Pelvic Varicose Veins of Pelvic Origin
Next Article in Special Issue
Exploring AI-Driven Machine Learning Approaches for Optimal Classification of Peri-Implantitis Based on Oral Microbiome Data: A Feasibility Study
Previous Article in Journal
Advancements in Hysteroscopic Diagnosis and Management of Endometritis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid CNN-Transformer Model for Accurate Impacted Tooth Detection in Panoramic Radiographs

by
Deniz Bora Küçük
1,
Andaç Imak
2,
Salih Taha Alperen Özçelik
3,*,
Adalet Çelebi
4,
Muammer Türkoğlu
1,
Abdulkadir Sengur
5 and
Deepika Koundal
6
1
Department of Software Engineering, Faculty of Engineering, Samsun University, 55000 Samsun, Turkey
2
Department of Electrical and Electronic Engineering, Faculty of Engineering, Munzur University, 62000 Tunceli, Turkey
3
Department of Electrical and Electronic Engineering, Faculty of Engineering, Bingol University, 12000 Bingol, Turkey
4
Oral and Maxillofacial Surgery Department, Faculty of Dentistry, Mersin University, 33000 Mersin, Turkey
5
Department of Electrical and Electronic Engineering, Faculty of Technology, Firat University, 23100 Elazig, Turkey
6
A.I. Virtanen Institute for Molecular Sciences, Faculty of Health Sciences, University of Eastern Finland, 70211 Kuopio, Finland
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(3), 244; https://doi.org/10.3390/diagnostics15030244
Submission received: 10 December 2024 / Revised: 16 January 2025 / Accepted: 18 January 2025 / Published: 22 January 2025

Abstract

:
Background/Objectives: The integration of digital imaging technologies in dentistry has revolutionized diagnostic and treatment practices, with panoramic radiographs playing a crucial role in detecting impacted teeth. Manual interpretation of these images is time consuming and error prone, highlighting the need for automated, accurate solutions. This study proposes an artificial intelligence (AI)-based model for detecting impacted teeth in panoramic radiographs, aiming to enhance accuracy and reliability. Methods: The proposed model combines YOLO (You Only Look Once) and RT-DETR (Real-Time Detection Transformer) models to leverage their strengths in real-time object detection and learning long-range dependencies, respectively. The integration is further optimized with the Weighted Boxes Fusion (WBF) algorithm, where WBF parameters are tuned using Bayesian optimization. A dataset of 407 labeled panoramic radiographs was used to evaluate the model’s performance. Results: The model achieved a mean average precision (mAP) of 98.3% and an F1 score of 96%, significantly outperforming individual models and other combinations. The results were expressed through key performance metrics, such as mAP and F1 scores, which highlight the model’s balance between precision and recall. Visual and numerical analyses demonstrated superior performance, with enhanced sensitivity and minimized false positive rates. Conclusions: This study presents a scalable and reliable AI-based solution for detecting impacted teeth in panoramic radiographs, offering substantial improvements in diagnostic accuracy and efficiency. The proposed model has potential for widespread application in clinical dentistry, reducing manual workload and error rates. Future research will focus on expanding the dataset and further refining the model’s generalizability.

1. Introduction

Impacted teeth are generally defined as situations in which the tooth cannot take its place in the mouth even though it is time to erupt, and based on clinical and radiologic evaluations, it is not possible to erupt through the natural process. Teeth may remain impacted in bone or soft tissue [1,2]. Teeth may remain impacted due to local and systemic factors. Local factors include deciduous tooth retention, obstacles in the eruption pathway, insufficient arch length, malposed tooth germs, and cleft lip and palate, while systemic factors include cleidocranial dysplasia, febrile diseases, Down syndrome, and hormonal irregularities. Mandibular third molars have one of the highest rates of impacted teeth among all teeth. They are followed by maxillary third molars, maxillary canines, and mandibular premolars [2].
Today, with the integration of artificial intelligence models in dentistry, as in any other field, there are promising results in the detection of dental conditions other than caries and calculus on panoramic radiographs. Thanks to the development of artificial intelligence models in all areas of dental radiology, it has been observed that they assist the clinician in panoramic diagnosis and treatment planning, as well as digitally based student education, especially during the pandemic period [3]. Artificial intelligence models also facilitate the detection of impacted third molars. In this regard, early diagnosis for the clinician and early treatment for the patient will be convenient. Artificial intelligence applications show high accuracy in detecting impacted third molars and their relationship with anatomical structures [4].
With the development of technology, revolutionary changes have taken place in the field of radiology from the past to the present. Studies on the development of patient care through artificial intelligence-based systems in radiographic images in the field of oral and dental health continue. With a growing population and a proportional increase in the number of patients, the inspection of panoramic dental images is a tiring and time-consuming process for experts. In addition, situations may be missed due to clinicians’ lack of experience [5,6]. To address this problem, research into the integration of computer-aided diagnostic tools in oral radiology, as in many other areas of medical imaging, is ongoing. In this way, artificial intelligence-based diagnostic systems provide an important solution to reduce the workload of experts by assisting in the evaluation of clinical data. Oral health plays an important role in a person’s overall health and quality of life. In addition, early detection of dental and oral health problems is important as they also cause systemic diseases in the body, such as respiratory infections [7]. Many applications of artificial intelligence-based oral radiology to oral and dental health have been proposed in the literature. These have mainly focused on areas such as the pathology of oral lesions [8,9], classification of teeth, materials, and dental implants [10,11,12,13,14,15], detection of periodontally problematic teeth [16], periodontitis [17,18,19,20], maxillary sinus [21,22,23,24,25,26,27,28,29,30], and dental caries [31,32,33,34,35,36,37,38,39,40,41,42]. Few studies have been conducted on impacted teeth in the literature compared to the detection of caries, maxillary sinus, and periodontal problems. Orhan et al. introduced a Convolutional Neural Network-based system for the detection of impacted third molars in cone beam computed tomography (CBCT) images. In their experimental study, they achieved a success rate of 86.2% in detecting impacted teeth [4]. Kuwada et al. used three deep learning architectures, AlexNet, VGG-16, and DetectNet, to detect maxillary sinus lesions and impacted third molars. In their experimental studies with three different deep architectures, they achieved a maximum accuracy of 96% in detecting impacted teeth with the DetectNet architecture [41]. Durmuş et al. proposed a deep architecture based on a ResNet backbone network for impacted tooth detection. They achieved 85.14% intersection over union (IoU) performance with the ResNet18 backbone network in the Pyramid Scene Parsing Network (PSPNet) architecture they developed [43]. Başaran et al. used the R-CNN Inception v2 model to evaluate ten different tooth conditions. The deep architecture they used on panoramic images achieved an F1 score performance of 86.25% [3]. Zhicheng et al. developed a SAM model for the detection of impacted teeth. They achieved an accuracy value of 86.73% with their proposed method [5]. Imak et al. proposed an improved version of U-Net called ResMIBCU-Net for the detection of impacted teeth. An accuracy value of 99.82% was observed in the proposed pixel-based method [44]. Celik investigated the effects of two detectors, Faster R-CNN and YOLOv3, on the detection of impacted third molars. In his experimental studies, Faster R-CNN achieved better performance than the YOLOv3 detector when compared to the results obtained with different backbone networks. A mean accuracy performance (mAP) of 96% was achieved with the proposed YOLOv3 detector [45]. In the current literature on impacted teeth, although the results are informative, the success results obtained with a limited number of applications are not at a sufficient level. More research is needed, and ongoing studies combining different innovative methods are required.
In this research, an artificial intelligence-based real-time learning framework is proposed for the detection of affected teeth in dentistry. The proposed model uses super-resolution methods to improve the resolution of panoramic radiography images. These methods are techniques used to convert low-resolution images into high-resolution images. In this way, the details in the image are improved, and the performance of the model is enhanced with clearer and sharper images.
The main contributions of this study are as follows:
  • Integration of deep learning algorithms, including super-resolution techniques, CNNs, transformer-based models, and the Weighted Boxes Fusion (WBF) component, to improve the accuracy and efficiency of object detection in panoramic radiography images.
  • Optimization of the model for accurate and real-time detection of impacted teeth in panoramic radiographs, assisting dentists in their clinical decision-making processes and contributing to more effective patient care.
  • High detection accuracy validated through experimental studies, demonstrating that the model can reliably detect impacted teeth in panoramic images.
  • Highlighting the potential of AI-based tools in dentistry to automate manual analysis processes, reduce the workload of experts, and minimize human error. These contributions show that the proposed model can have a wide range of applications in both academic and clinical settings, establishing a future reference point for artificial intelligence-based solutions in dentistry.

2. Materials and Methods

In this study, a multi-component deep learning model was developed to enhance the analysis of panoramic radiography images. The proposed model utilizes super-resolution techniques to convert low-resolution images into high-resolution representations, enhancing image clarity and detail. This improvement facilitates better performance in diagnostic applications.
The model integrates Convolutional Neural Networks (CNNs) and transformer architectures for regional object detection. CNNs are employed to identify and classify features in the images, while transformers leverage attention mechanisms to model relationships in sequential data. Additionally, a Weighted Boxes Fusion (WBF) component is incorporated to combine the outputs of multiple object detection models, ensuring more accurate and reliable predictions.
The parameters of the model were rigorously optimized through iterative trials and fine-tuning to maximize performance. These adjustments were designed to achieve the best fit for the training data and to ensure the model operates efficiently. A general representation of the proposed model is provided in Figure 1. In this research, an artificial intelligence-based real-time learning framework is proposed for the detection of affected teeth in dentistry. The proposed model is based on the combination of Convolutional Neural Networks (CNNs) and transformer models based on super-resolution images. The basic architectures that make up the proposed model are further explained in more detail in the following subsections.
The proposed model, shown in Figure 1, consists of four main components. Each component provides an optimized function for the detection of dental structures in panoramic images:
  • The Generative Facial Prior (GFP-GAN) module improves the clarity of the input image by enhancing low-resolution panoramic images with a super-resolution method. This process provides higher quality data for the next steps and contributes significantly to the overall performance of the model. GFP-GAN reconstructs the detail in low-resolution images, producing a sharper and more meaningful input.
  • In the second phase, the RT-DETR module uses a transformer-based approach to regionally detect dental structures in panoramic images. This module extracts meaningful features from complex image structures and accurately labels target regions. In particular, it provides high accuracy in dental radiographs thanks to its ability to model sequential relationships and improve positional accuracy.
  • In the third stage, the YOLOv8 model classifies the detected tooth regions in detail and marks them more precisely. YOLOv8 is one of the most efficient object detection algorithms, delivering fast and accurate results. This component allows detailed analysis of important structures in dental radiographs and optimizes the detected areas, improving overall accuracy.
  • In the final stage, the detection results from the different modules are combined with the WBF method, enhanced with Bayesian optimization. This method combines the strengths of the models and provides a more accurate and integrated output. By combining the predictions of different modules, WBF minimizes the false positive and false negative rates. Finally, the analysis of the dental structures is complete, and the results are presented with high accuracy.
The algorithms and architectures used in these process steps are described in more detail in the subheadings. The function and benefits of each component contribute significantly to the overall success of the model. This integrated structure provides an innovative solution for high accuracy, precision, and reliability in dental radiography.

2.1. Generative Facial Prior

Generative Facial Prior (GFP-GAN), proposed by Wang et al., (2021), is an advanced GAN (Generative Adversarial Network)-based model for producing high-resolution images from low-resolution inputs [46]. Image reconstruction generally includes a structure consisting of U-Net-based encoder and decoder parts. Specifically, StyleGAN2 [47] stands out as a technology with strong production capability when used as a front module. A pre-trained StyleGAN2 module provides a robust ability to restore detailed regions in the image by utilizing the rich features of the semantic information extracted from the encoder. The degradation removal module increases the expressive power of the model by retaining semantic attributes and extracting deep feature information during the convolution process, as shown in Equation (1) and Equation (2), respectively [46]:
W = M L P ( F l a t e n t )
F G A N = S t y l e G A N 2 ( W )
Here, the F l a t e n t coding vector, generated by the degradation removal process, encodes meaningful features from the input image. It is then processed through Multi-Layer Perceptron (MLP) layers to produce the latent code W, which is passed into StyleGAN2 to extract deep, meaningful features. These steps allow the model to focus on fine-grained details necessary for high-quality restoration. In order to remove blurs and enhance the final restoration quality, the model uses a comprehensive set of loss functions. Reconstruction Loss (Lrec) ensures the restored image closely resembles the ground truth, while Adversarial Loss (Ladv) enhances the production of realistic textures. Additionally, Facial Component Loss (Lcomp) improves key facial areas such as the eyes and mouth, and Identity Preserving Loss (Lid) maintains the identity of the restored face. This combination of losses enables better compatibility between the encoder and decoder, improving both realism and fidelity [46,47,48,49]. The comprehensive training process, supported by these loss functions, allows GFP-GAN to excel not only in face restoration but also in applications like panoramic dental imaging. For example, it effectively enhances details such as fractures, cavities, or impacted teeth, demonstrating its versatility and practical potential in medical imaging and other image enhancement tasks.

2.2. CNN (YOLO)

The YOLO (You Only Look Once) series, which offers fast and accurate solutions in the field of object detection, continues to develop constantly. One of these developments is the YOLOv8 model. YOLOv8 stands out as an important step forward in the field of computer vision.
  • Main Network: YOLOv8 uses a modified version of CSPDarknet53 as its main network. CSPDarknet53 is replaced by the C2f Module, which uses gradient shunt connectivity to enrich the information flow and maintains a lightweight structure. Furthermore, the GIS Module performs processing using convolution, group normalization, and SiLU activation. YOLOv8 also uses the SPPF Module to convert the input feature maps into a fixed-size map, which reduces computational effort and lowers latency.
  • Neck: The neck structure of YOLOv8 uses the PAN-FPN structure. Inspired by PANet, this structure improves location information by combining features at different scales and provides feature diversity and completion.
  • Head: YOLOv8 uses two separate branches for segmentation and bounding box regression. For classification purposes, binary cross-entropy loss (BCE Loss) is applied, whereas distribution focal loss (DFL) and CIoU are utilized for tasks involving bounding box regression. The model also uses a non-anchor detection model and improves detection accuracy and robustness by assigning by task.
With these components, YOLOv8 provides improved detection accuracy and speed over previous models in the YOLO series. With its modular structure and innovative components, it sets a high standard in the field of computer vision [50,51,52].

2.3. Transformer (RT-DETR)

The RT-DETR (Real-Time Detection Transformer) model, developed by Baidu, offers an innovative approach to object detection by combining high accuracy and real-time performance [53]. The model includes components that provide significant improvements in both speed and accuracy. These are as follows [53,54]:
  • Backbone Network: RT-DETR uses convolution-based networks such as ResNet or HGNetv2 as the backbone. The last three stages of the backbone (S3, S4, S5) serve as the input for the hybrid encoder and enable efficient processing of multi-scale features. This design allows the model to be built on a strong foundation and provides a suitable structure for modeling multi-scale information.
  • Neck: The neck structure of RT-DETR is not designed as a distinct intermediate layer as in conventional models but instead offers an innovative structure integrated within the Hybrid Encoder. The Neck function is realized by the following two modules:
    • Attention in Feature Interaction (AIFI): This module is specifically designed to improve semantic information in deep feature maps (S5). AIFI enhances information flow by modeling dependencies between features. This approach provides high performance with a more minimal design unlike classical neck structures.
    • Cross-Scale Feature-Fusion Module (CCFM): Combining features at different scales (S4 and S5) enables the model to effectively use multi-scale information. Instead of a classical feature pyramid, CCFM offers a lighter and more flexible fusion mechanism.
This innovative approach allows the model to remain lightweight while avoiding the additional computational burden seen in traditional neck structures. On the other hand, the head part of this architecture is as follows [53,54]:
3.
Head: Unlike conventional object detection models, RT-DETR’s head structure is designed to be more dynamic and integrated with the encoder. This structure works as follows:
  • Decoder: The IoU-aware Query Selection mechanism selects the most relevant object queries from the features received from the encoder and optimizes them to generate class information and bounding boxes. This mechanism reduces redundant computations and improves accuracy by focusing on the most important objects in the scene.
  • Auxiliary Prediction Heads: These heads enable the model to learn faster and more accurately by making intermediate predictions. Unlike the separate classification and regression heads in classical models, it works while integrated into the decoder and does not require an extra ‘head’ layer.
RT-DETR provides a lighter, faster, and scalable solution by offering a hybrid and integrated design instead of the distinct neck and head structures in classical models. This innovative approach reduces the computational burden and is optimized for real-time performance. In this way, RT-DETR represents a significant paradigm shift in the field of object detection [53,54].

2.4. Ensemble Strategy

In this study, an ensemble strategy was used to combine the strengths of the YOLO and RT-DETR models. Ensemble methods aim to improve accuracy, precision, and reliability by combining the predictions of different models. In particular, the fast real-time detection capabilities of YOLO and the long-range correlation learning capability of RT-DETR are combined to achieve superior performance in the detection of impacted teeth. The ensemble strategy used in this study is based on the Weighted Boxes Fusion (WBF) algorithm [54]. The WBF algorithm assigns a weight to each box when fusing bounding box predictions from multiple models. This process optimizes the contribution of each bounding box to the model prediction, resulting in a more accurate result.
The steps of the WBF algorithm are as follows [55,56,57,58]:
  • The combined confidence score C of the fused bounding box is calculated as the average of the confidence scores C i of all participating bounding boxes:
C = i = 1 T C i T
Here, T represents the total number of bounding boxes contributing to the fusion.
  • The x , y , w , h coordinates of the fused bounding box are computed as weighted averages of the corresponding coordinates of the individual bounding boxes:
x = i = 1 T C i · x i i = 1 T C i
y = i = 1 T C i · y i i = 1 T C i
w = i = 1 T C i · w i i = 1 T C i
h = i = 1 T C i · h i i = 1 T C i
Here, C i is the confidence score of the i t h bounding box, and x i , y i , w i , h i represent the center coordinates, width, and height of the i t h bounding box.
I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n
Bounding boxes with IoU values below the threshold are excluded from the results. This step ensures that the fused box retains only the most relevant predictions, eliminating excessive overlap.
In summary, Weighted Boxes Fusion (WBF) is an efficient method that combines bounding boxes estimated by different object detection models by optimizing them according to their confidence values and model weights. This approach aims to emphasize the strengths of the models while minimizing their weaknesses, thereby significantly improving the accuracy, consistency, and reliability of the combined predictions [55,56,57]. In this study, the strengths of the YOLO and RT-DETR models are combined using the WBF algorithm, and a high-performance solution for impacted tooth detection is presented.

2.5. Optimization

In this study, Bayesian optimization is used to determine the optimal parameters of the developed model and to improve the detection performance [58]. Unlike traditional grid search and random search methods, Bayesian optimization makes hyperparameter search more efficient and continuously improves model performance in an iterative process. In this process, performance data from past trials are analyzed using Gaussian processes, and the next hyperparameter combinations are selected probabilistically [58,59,60,61].
Bayesian optimization uses Gaussian processes to model the objective function f ( θ ) . This method estimates a mean μ ( θ ) and a variance σ 2 ( θ ) for each hyperparameter θ . A Gaussian process is defined as follows:
f ( θ ) G P ( μ ( θ ) , k ( θ , θ ) )
Here, k ( θ ,   θ ) , is the covariance function, which indicates the correlation of function values at two different hyperparameter points. In Bayesian optimization, the learning function a ( θ ) determines the next trial point by balancing exploration and exploitation strategies. The Expected Improvement (EI) function is often used:
a ( θ ) = E [ m a x ( f ( θ ) f ( θ + ) , 0 ) ]
Here, f ( θ + ) represents the best available observation. The expected improvement function assesses the potential for improvement at a new point. Using Gaussian processes, the hyperparameter θ * at which this acquisition function is maximized is found using Equation (11).
θ * = a r g max θ a ( θ )
The model is trained with this new hyperparameter combination, its performance is measured, and the value of f ( θ ) is calculated. The Gaussian processes are updated with the new performance results, and the optimization process continues until the stopping criterion is reached. This iterative process provides a more accurate estimate of the objective function, making it possible to find the optimal hyperparameters with fewer trials [57,62].
This study aims to optimize the hyperparameters of the Weighted Boxes Fusion (WBF) algorithm such as Threshold, Skip Box Threshold, YOLO and RT-DETR weights using Bayesian optimization. The mathematical methods applied in the optimization process aim to improve the overall accuracy performance of the model by making the hyperparameter search both time and resource efficient.

2.6. Dataset

In this study, three different datasets consisting of panoramic radiographs are used for impacted tooth detection. The images in the datasets were carefully selected according to high resolution and quality standards and optimized for training the model. The dataset contains a total of 407 images, 304 images in PNG format with 540 × 380 pixels [44], 53 images in JPG format with 2041 × 1024 pixels [63] and 50 images in PNG format with 3100 × 1300 pixels [64]. This variety was specifically designed to provide results applicable to a larger population of teeth and to evaluate model performance at different resolutions.
Labeling procedures were carefully performed by experts in the field and all images were processed according to the YOLO format. The bounding box method was applied using the Roboflow platform and the positions of the impacted teeth were determined with high accuracy. This detailed labeling process provided a reliable basis for the training and testing phases of the model. Example images of the labeled dataset are shown in Figure 2.
This dataset has the potential to improve the overall performance of the model by providing samples at a wide range of resolutions for the detection of impacted teeth. The data selection and labeling process ensured a high degree of accuracy in the training and evaluation of the model, while allowing the results of the study to be generalized to a larger population of teeth. The resulting dataset is an important resource for impacted tooth detection studies, both in terms of its size and quality.

2.7. Implementation Details

The experimental studies were developed using the Python programming language and performed on a high-performance workstation with an Intel Core i9-14900K processor and NVIDIA RTX 4090 GPU, running on a Windows operating system. The dataset used was divided into three parts in order to provide an optimized process in the training, validation and testing phases of the model. In this context, 80% of the data (325 samples) are used for training, 5% (21 samples) for validation, and 15% (61 samples) for testing. The data split was performed only once, and the same datasets were used in all experimental studies to ensure consistency of analyses.
In the training process of the models in the experimental studies, default parameters optimized according to the YOLO and RT-DETR architectures were used. During training, the learning rate was set to 0.01, the momentum to 0.9, the weight decay to 0.0005, the batch size to 16, and the number of epochs to 500. The image size was fixed at 640 × 640 pixels. These hyperparameters were carefully chosen to efficiently train the model and improve its overall performance. To minimize the risk of overfitting during the training process, the EarlyStopping method was applied. This method automatically stops training if no improvement in the validation loss is observed for a certain period of time, preventing the model from overlearning unnecessarily. This optimized the training process of the model, resulting in more reliable and consistent results. This approach allowed the model to perform well on both training and test data.
The Weighted Boxes Fusion (WBF) algorithm was used to fuse the outputs of the YOLO and RT-DETR models. The parameters of the WBF algorithm were optimized using Bayesian optimization. In this process, hyperparameters such as IoU threshold, skip box threshold, YOLO weight and RT-DETR weight were optimized. The search ranges of the hyperparameters used in the optimization are shown in Table 1.
These optimized hyperparameters allowed the WBF algorithm to more effectively combine the strengths of the models and improve the overall detection accuracy. This approach, which is one of the highlights of this study, is one of the innovative methods used to improve model performance.

2.8. Performance Metrics

In this study, the performance of object recognition models is evaluated using key metrics such as precision, recall and mAP50. These metrics provide a comprehensive analysis of the prediction accuracy and reliability of the model. The performance metrics are defined mathematically in Equations (12)–(16):
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
Here, TP, FP and FN refer to true positive predictions, false positive predictions and false negative predictions respectively. The F1 score is a performance metric that represents the harmonic mean of precision and recall, offering a balanced measure of a model’s accuracy. It is particularly useful when dealing with imbalanced datasets, as it accounts for both false positives and false negatives. A higher F1 score indicates better overall performance in correctly identifying positive cases while minimizing errors. On the other hand, IoU measures the overlap between a predicted bounding box and the actual bounding box:
I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n
where the estimate is considered a true positive if the IoU threshold is 0.5 or greater, and this is used in the mAP50 calculations. The average precision (AP) is finding the approximation of the area under the PR curve. The actual area under the curve, where p ( r ) is the precision at recall r , can be defined as follows:
A P = 0 1 p ( r ) d r
mAP50 (mean average precision at IoU 0.5) represents the average of the calculated AP values for all classes:
m A P 50 = 1 C c = 1 C A P c
where C is the number of classes and A P c is the average precision for class c at the IoU threshold of 0.5. These metrics reveal the overall performance of the model by assessing its correct and incorrect predictions. They also provide a comprehensive framework for analyzing the accuracy, coverage and consistency of the predictions.

3. Experimental Results and Analysis

The experiments were conducted in three main phases: evaluation of individual models, evaluation of ensemble models, and evaluation of the optimized ensemble approach. The performance results based on these phases are detailed in the following subsections.

3.1. Individual Results

In this experimental study, different versions of the YOLOv8 and RT-DETR architectures are independently trained and tested. The performance of the models is evaluated both on normal images and on images enhanced with GAN-based super-resolution techniques. The performance results are presented in Table 2.
The results obtained from the analysis (as shown in Table 2) are as follows:
  • The YOLOv8L model achieved the highest F1 values (94.6% and 94.5%) with both normal and GAN-based super-resolution images. It was also observed that mAP@0.5 (96.5%) improved with GAN-based images. These results show that YOLOv8L performs well with both image types. On the other hand, the YOLOv8X model performed similarly on normal and GAN-based images (88.8% mAP@0.5 and 93.7% F1). This suggests that the effect of GAN-based super-resolution on this model is limited.
  • RT-DETR-L achieved the highest IoU score of 95.2% mAP@0.5 for normal images. However, this value decreased slightly to 94.7% for GAN-based images. In contrast, the GAN enhancement resulted in a significant increase in the F1 score (from 88.2% to 92.3%). On the other hand, RT-DETR-X showed a balanced performance on both normal and GAN-based images. However, there was a slight decrease in mAP@0.5 (92.3%) and F1 (91.7%) after GAN enhancement.
These results suggest that the impact of GAN-based super-resolution techniques on model performance may differ between model architectures. While YOLOv8L gives the best overall results for both mAP@0.5 and F1 metrics, the RT-DETR-L model shows a significant improvement after GAN application. This shows that the optimization of model components and the correct application of GAN-based enhancement techniques play a critical role in improving performance.

3.2. Ensemble Results

The second experimental study focused on the fusion of the outputs of the YOLO and RT-DETR architectures using the Weighted Boxes Fusion (WBF) algorithm. At this stage, the WBF algorithm was not subjected to any optimization process, and the fusion process was performed with default WBF parameters. The ensemble approach aims to achieve higher accuracy rates by combining the strengths of both models. In this study, different combinations of YOLOv8 and RT-DETR models were tested on both normal and GAN-enhanced images. The performance results are detailed in Table 3.
The results obtained from the analysis (as shown in Table 3) are as follows:
  • In tests on normal images, the model combinations YOLOv8L + RT-DETR-X and YOLOv8X + RT-DETR-X achieved the highest F1 values (94.9% and 94.8%) and precision values (93.0% and 94.4%). These results indicate that the RT-DETR-X component provides high precision in normal images. On the other hand, the YOLOv8X + RT-DETR-L combination achieved the highest IoU value (95.6%) on normal images, but the recall rate (96.0%) remained similar to the other models.
  • For GAN-enhanced images, the YOLOv8L + RT-DETR-X and YOLOv8L + RT-DETR-L combinations achieved the highest IoU and F1 scores (97.5% and 93.5%, 97.4% and 92.3%, respectively). YOLOv8L + RT-DETR-X performed particularly well in terms of precision (89.1%) and recall (98.4%). The combination YOLOv8X + RT-DETR-X showed a balanced performance on both normal and GAN-enhanced images, but the performance improvement after GAN enhancement was limited compared to the other combinations.
These results suggest that the ensemble strategy is effective in improving performance by combining the strengths of different models. In particular, GAN-based visualization was found to provide a significant improvement in IoU and recall rates in the YOLOv8L + RT-DETR-X combination. These results show that the proposed ensemble approach can make a significant contribution to the reliability and accuracy of clinical decision support systems. The detail enhancement effect of GAN-enhanced images is an important method to achieve better results in dental radiography.

3.3. Optimized Results

In this study, the parameters of the Weighted Boxes Fusion (WBF) algorithm are optimized using Bayesian Optimization to improve model performance. Bayesian optimization is a powerful method that makes the parameter search process more efficient by allowing an effective selection of hyperparameters. The optimization process focused on improving the precision, recall and F1 metrics. This process played a crucial role in improving the performance of the ensemble model. As a result of the optimization, models with mAP@0.5 values of 97% and above were considered and detailed performance results were obtained. The performance results of the optimized models are shown in Table 4.
The results obtained from the analysis (as shown in Table 4) are as follows:
  • YOLOv8L + RT-DETR-L: This model showed the highest accuracy of 98.3% with mAP@0.5. This means that the overall recognition accuracy of the model is quite high. The recall rate reached a remarkable 99.2%, indicating that the model was able to correctly detect almost all target objects. However, the precision remained relatively low compared to the other metrics at 86.7%. This suggests that the model’s false positive prediction rate could be improved. The F1 score reached 92.5% in the balance of precision and recall. The optimized parameters of the WBF algorithm for this model were determined as follows:
    • IoU Threshold (iou_thr): 0.3466;
    • Skip Box Threshold (skip_box_thr): 0.0340;
    • YOLOv8L Weight (weight1): 4.2111;
    • RT-DETR-L Weight (weight2): 2.9897.
  • YOLOv8L + RT-DETR-X: This model achieved a high accuracy of 97.5% mAP@0.5 and a remarkable performance with an F1 score of 96.0%. The precision was 93.8% and the recall 98.4%. This shows that the model provides a balanced and reliable performance. The optimized parameters of the WBF algorithm for this model were determined as follows:
    • IoU Threshold (iou_thr): 0.3597;
    • Skip Box Threshold (skip_box_thr): 0.0843;
    • YOLOv8L Weight (weight1): 7.7882;
    • RT-DETR-L Weight (weight2): 7.6551.
These results obtained with Bayesian optimization show that the WBF algorithm significantly improves its performance and that the correct choice of hyperparameters plays a critical role in the success of the models. The YOLOv8L + RT-DETR-L model showed strong performance in comprehensive object detection with high mAP and recall rates. However, the YOLOv8L + RT-DETR-X model offered a superior result in terms of overall accuracy and precision due to its balanced precision and recall ratios.
These results suggest that parameter refinement using Bayesian optimization is an effective tool to improve model performance, and that the WBF algorithm can achieve higher accuracy rates with optimized parameters.

3.4. Visualization

In this study, the visual results of YOLOv8-L, YOLOv8-X, RT-DETR-L, RT-DETR-X, and the proposed GAN-based optimization-assisted ensemble model are presented in detail in Figure 3. Figure 3 contains a series of panoramic images that visually compare the performance of the models used in the impacted tooth detection task. Each column represents a specific sample panoramic image, while each row shows the predictions of different model combinations. This structure allows a visual evaluation of the detection accuracy of the models. The first row contains the actual detections (ground truth), marked with green boxes, which are used as a reference to assess the accuracy of other models. Between the second and fifth rows, the predictions of the YOLOv8-L, YOLOv8-X, RT-DETR-L, and RT-DETR-X models are shown in blue boxes. The results in these rows show how the base models perform in different situations. The last row shows the predictions of the GAN-assisted RT-DETR-L and the proposed GAN-based optimization-assisted ensemble models. These results are included to highlight the effectiveness of the proposed method compared to other models. The visualizations provide a visual verification of the numerical results in Table 4 and clearly show the performance differences of the models.
The visualizations in Figure 3 clearly show how the embedded tooth detection performance of the proposed models responds to different scenarios. In particular, the optimized super-resolution GAN-based YOLOv8L + RT-DETR-L model showed an overall superior accuracy. The model correctly detected the vast majority of impacted teeth in all cases; however, in case 3, it incorrectly classified a non-impacted region as impacted. While this does not detract from the overall success of the model, it does suggest that further improvements are needed to reduce false positive predictions. In addition, the GAN-based YOLOv8L + RT-DETR-X model showed balanced performance, but was only able to correctly detect two of the three impacted teeth in Sample-4. This suggests that the model may occasionally have difficulty detecting impacted teeth in low-contrast regions or small sizes. However, considering the overall accuracy and precision of the model, it can be said that it provides a highly effective alternative for detecting impacted teeth.
On the other hand, the individual models YOLOv8-L, YOLOv8-X, RT-DETR-L and RT-DETR-X generally lagged behind the proposed GAN-based models. These models were found to under-detect, mis-locate or make false positive predictions in many cases. For example, these models failed to accurately localize target objects, particularly in low contrast areas or in complex panoramic images where impacted teeth overlap with other teeth. It was also found that in some cases, non-impacted teeth were incorrectly identified as impacted due to the model’s limited contextual understanding. This highlights the potential limitations of individual models in deep learning-based detection tasks and emphasizes the need for more advanced algorithms in complex image scenarios. Nevertheless, the fast processing capacity of individual models and their accuracy in certain situations suggest that they may be useful in lower cost and specific applications. However, in terms of overall performance, the integration of these models with the proposed GAN-based approaches provides significant improvements in terms of accuracy and reliability.
In conclusion, the visualizations presented in Figure 3 show that the proposed GAN-based models provide a significant advantage in terms of overall accuracy, precision and reliability compared to the individual models. However, addressing and improving the limitations observed in these models will contribute to obtaining more comprehensive and reliable results in clinical applications such as impacted tooth detection.

4. Discussion

4.1. Comparison of Previous Studies with Proposed Model

Studies on impacted tooth detection are limited in the literature and are generally based on pixel-based segmentation methods or regional detection approaches. Segmentation-based studies have generally achieved accuracies of 85% and higher [4,5,42,43,44]. However, regional detection methods have achieved higher accuracy and sensitivity rates, and the potential of these methods has been confirmed by studies using different deep learning models. The datasets, models used, and performance results of some important studies in the literature are summarized in Table 5.
In the studies presented in Table 5, although it is difficult to directly compare the performance of the models since they are based on different datasets, it is clearly seen that the proposed model shows a superior performance in terms of mAP and F1 score. According to the results obtained in this study, it is observed that the proposed model has high accuracy rates. It provides a significant performance improvement compared to similar studies in the literature, which emphasizes the superiority of the model. Looking at the comparative results in the literature, the YOLOv3 model used in [45] stands out with a mAP value of 96%. However, the proposed GAN-assisted YOLOv8L + RT-DETR-L model provided higher accuracy and precision than the existing methods in the literature with 98.3% mAP and a 92.5% F1 score. In addition, the GAN-assisted YOLOv8L + RT-DETR-X model provides a balanced and reliable result with 97.5% mAP and a 96% F1 score. These results show that the proposed model has made significant progress in the field of impacted tooth detection and outperforms the existing methods in the literature.
The main advantage of the proposed model is that it provides higher accuracy and consistency thanks to the GAN-based super-resolution techniques and ensemble approach. These methods significantly improved the overall performance by reducing false positive and negative predictions while increasing the clarity of the detected regions. Other studies in the literature usually use only a single model or limited combinations of models. In contrast, the proposed model not only integrates GAN-assisted super-resolution methods, but also achieves superior performance by combining powerful models such as YOLO and RT-DETR. Moreover, the model fusion process with the WBF algorithm emphasizes the strengths of each model while minimizing its weaknesses, resulting in more reliable results. In conclusion, this study has presented an innovative modeling strategy that is different from the existing work on impacted tooth detection and has made a significant contribution to the literature. The proposed model is considered to have a wide application potential, not only for impacted tooth detection, but also for various dental detection and analysis problems

4.2. Limitations

The proposed model is based on the integration of Convolutional Neural Networks (CNN) and transformer-based models on panoramic images enhanced with super-resolution techniques. The ability of the model to accurately detect the location of impacted teeth in all test images clearly demonstrates its overall success and reliability. However, it was observed that the model made errors in some cases, such as misidentifying non-impacted teeth as impacted teeth. Such limitations are critical to better understand the model’s performance in specific scenarios and to identify opportunities for future improvement. To better illustrate the limitations and challenges faced by the model, example images are shown in Figure 4.
The limitations of the model can be clearly seen in images Example-1 and Example-2 in Figure 4. In Example-1, the model incorrectly identified a fractured tooth as an impacted tooth, resulting in a false positive. Similarly, in Example-2, a region with no teeth was incorrectly labeled as an impacted tooth. These cases show that the model can misinterpret some complex structures and produce incorrect results in certain situations. These limitations provide important guidance by highlighting areas where the model needs to be improved. In particular, strategies such as larger and more diverse datasets, integration of more advanced feature extraction methods, and optimization of attention mechanisms could be considered in future studies. Furthermore, techniques to reduce false positive and false negative rates can be applied to make the model more robust in real-world applications.
In conclusion, this study highlights the strengths of the proposed model as well as its limitations, providing important insights that can guide future research. Optimizing the model to address these limitations will provide more effective and comprehensive solutions, not only for impacted tooth detection, but also for other dental detection and analysis problems.

4.3. Conclusions

This study presents an artificial intelligence-based solution for the detection of impacted teeth in panoramic radiographic images, combining YOLO and RT-DETR models with the Weighted Boxes Fusion (WBF) algorithm and Bayesian optimization. The proposed model achieved impressive results, including 98.3% mAP and a 96% F1 score, demonstrating its reliability and accuracy.
The model minimizes false positives and negatives while improving identification rates, making it a practical tool for clinical applications. It also shows potential for broader applications in dental and medical imaging tasks.
Future research could focus on validating the model on larger, diverse datasets and adapting it for other dental challenges, such as jaw anomalies, caries detection, or implant analysis.

Author Contributions

Conceptualization, D.B.K., M.T. and A.I.; methodology, A.I., D.B.K. and M.T.; software, D.B.K. and A.I.; validation, S.T.A.Ö., A.Ç. and D.K.; formal analysis, S.T.A.Ö., M.T. and A.Ç.; investigation, M.T.; resources, S.T.A.Ö.; data curation, A.Ç.; writing—original draft preparation, M.T., A.I. and D.B.K.; writing—review and editing, D.K., S.T.A.Ö. and A.Ç.; visualization, D.B.K. and A.I.; supervision, A.S.; project administration, A.S.; funding acquisition, A.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by Firat University, Scientific Research Project Committee, under grant no: TEKF.24.46.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is available on request from the corresponding author. The data are not publicly available due to privacy concerns.

Conflicts of Interest

None of the authors have potential conflicts of interest to be disclosed.

References

  1. Göksu, V.C.; Ersoy, H.E.; Eberliköse, H.; Yücel, E. Gömülü Mandibular Üçüncü Molar Diş Pozisyonlarının Demografik Olarak İncelenmesi: Retrospektif Çalışma. Ado Klin. Bilim. Derg. 2021, 10, 165–171. [Google Scholar]
  2. Kaczor-Urbanowicz, K.; Zadurska, M.; Czochrowska, E. Impacted Teeth: An Interdisciplinary Perspective. Adv. Clin. Exp. Med. Off. Organ Wroclaw Med. Univ. 2016, 25, 575–585. [Google Scholar] [CrossRef]
  3. Başaran, M.; Çelik, Ö.; Bayrakdar, I.S.; Bilgir, E.; Orhan, K.; Odabaş, A.; Jagtap, R. Diagnostic charting of panoramic radiography using deep-learning artificial intelligence system. Oral Radiol. 2022, 38, 363–369. [Google Scholar] [CrossRef] [PubMed]
  4. Orhan, K.; Bilgir, E.; Bayrakdar, I.S.; Ezhov, M.; Gusarev, M.; Shumilov, E. Evaluation of artificial intelligence for detecting impacted third molars on cone-beam computed tomography scans. J. Stomatol. Oral Maxillofac. Surg. 2021, 122, 333–337. [Google Scholar] [CrossRef] [PubMed]
  5. Zhicheng, H.; Yipeng, W.; Xiao, L. Deep Learning-Based Detection of Impacted Teeth on Panoramic Radiographs. Biomed. Eng. Comput. Biol. 2024, 15, 11795972241288319. [Google Scholar] [CrossRef] [PubMed]
  6. Faure, J.; Engelbrecht, A. Impacted tooth detection in panoramic radiographs. In Proceedings of the Advances in Computational Intelligence: 16th International Work-Conference on Artificial Neural Networks, IWANN 2021, Virtual Event, 16–18 June 2021; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 525–536, Part I 16. [Google Scholar]
  7. Sheiham, A. Editorials Oral Health, General Health and Quality of Life. Bull. World Health Organ. 2005, 83, 644. [Google Scholar]
  8. Sultan, A.S.; Elgharib, M.A.; Tavares, T.; Jessri, M.; Basile, J.R. The use of artificial intelligence, machine learning and deep learning in oncologic histopathology. J. Oral Pathol. Med. 2020, 49, 849–856. [Google Scholar] [CrossRef]
  9. Alhazmi, A.; Alhazmi, Y.; Makrami, A.; Masmali, A.; Salawi, N.; Masmali, K.; Patil, S. Application of artificial intelligence and machine learning for prediction of oral cancer risk. J. Oral Pathol. Med. 2021, 50, 444–450. [Google Scholar] [CrossRef]
  10. Takahashi, T.; Nozaki, K.; Gonda, T.; Mameno, T.; Wada, M.; Ikebe, K. Identification of dental implants using deep learning—Pilot study. Int. J. Implant. Dent. 2020, 6, 1–6. [Google Scholar] [CrossRef] [PubMed]
  11. Lee, D.W.; Kim, S.Y.; Jeong, S.N.; Lee, J.H. Artificial intelligence in fractured dental implant detection and classification: Evaluation using dataset from two dental hospitals. Diagnostics 2021, 11, 233. [Google Scholar] [CrossRef]
  12. Imak, A.; Çelebi, A.; Türkoğlu, M.; Şengür, A. Dental material detection based on faster regional convolutional neural networks and shape features. Neural Process. Lett. 2022, 54, 2107–2126. [Google Scholar] [CrossRef]
  13. Tuzoff, D.V.; Tuzova, L.N.; Bornstein, M.M.; Krasnov, A.S.; Kharchenko, M.A.; Nikolenko, S.I.; Bednenko, G.B. Tooth detection and numbering in panoramic radiographs using convolutional neural networks. Dentomaxillofac. Radiol. 2019, 48, 20180051. [Google Scholar] [CrossRef] [PubMed]
  14. Putra, R.H.; Astuti, E.R.; Putri, D.K.; Widiasri, M.; Laksanti, P.A.M.; Majidah, H.; Yoda, N. Automated permanent tooth detection and numbering on panoramic radiograph using a deep learning approach. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2024, 137, 537–544. [Google Scholar] [CrossRef] [PubMed]
  15. Brahmi, W.; Jdey, I. Automatic tooth instance segmentation and identification from panoramic X-Ray images using deep CNN. Multimed. Tools Appl. 2024, 83, 55565–55585. [Google Scholar] [CrossRef]
  16. Sadr, S.; Mohammad-Rahimi, H.; Motamedian, S.R.; Zahedrozegar, S.; Motie, P.; Vinayahalingam, S.; Dianat, O.; Nosrat, A. Deep Learning for Detection of Periapical Radiolucent Lesions: A Systematic Review and Meta-analysis of Diagnostic Test Accuracy. J. Endod. 2023, 49, 248–261.e3. [Google Scholar] [CrossRef]
  17. Zhu, X.; Wang, F.; Xie, Y.; Li, W.; Wang, X.; Liu, J. Periodontitis auxiliary diagnosis based on deep learning and oral dental x-ray images. Russ. J. Nondestruct. Test. 2023, 59, 487–500. [Google Scholar] [CrossRef]
  18. Bayrakdar, I.S.; Orhan, K.; Çelik, Ö.; Bilgir, E.; Sağlam, H.; Kaplan, F.A.; Różyło-Kalinowska, I. AU-net approach to apical lesion segmentation on panoramic radiographs. BioMed Res. Int. 2022, 2022, 7035367. [Google Scholar] [CrossRef] [PubMed]
  19. Shon, H.S.; Kong, V.; Park, J.S.; Jang, W.; Cha, E.J.; Kim, S.Y.; Kim, K.A. Deep learning model for classifying periodontitis stages on dental panoramic radiography. Appl. Sci. 2022, 12, 8500. [Google Scholar] [CrossRef]
  20. Jiang, L.; Chen, D.; Cao, Z.; Wu, F.; Zhu, H.; Zhu, F. A two-stage deep learning architecture for radiographic staging of periodontal bone loss. BMC Oral Health 2022, 22, 106. [Google Scholar] [CrossRef] [PubMed]
  21. Mori, M.; Ariji, Y.; Katsumata, A.; Kawai, T.; Araki, K.; Kobayashi, K.; Ariji, E. A deep transfer learning approach for the detection and diagnosis of maxillary sinusitis on panoramic radiographs. Odontology 2021, 109, 941–948. [Google Scholar] [CrossRef] [PubMed]
  22. Murata, M.; Ariji, Y.; Ohashi, Y.; Kawai, T.; Fukuda, M.; Funakoshi, T.; Kise, Y.; Nozawa, M.; Katsumata, A.; Fujita, H.; et al. Deep-learning classification using convolutional neural network for evaluation of maxillary sinusitis on panoramic radiography. Oral Radiol. 2019, 35, 301–307. [Google Scholar] [CrossRef]
  23. Çelebi, A.; Imak, A.; Üzen, H.; Budak, Ü.; Türkoğlu, M.; Hanbay, D.; Şengür, A. Maxillary sinus detection on cone beam computed tomography images using ResNet and Swin Transformer-based UNet. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2024, 138, 149–161. [Google Scholar] [CrossRef] [PubMed]
  24. Kim, H.G.; Lee, K.M.; Kim, E.J.; Lee, J.S. Improvement diagnostic accuracy of sinusitis recognition in paranasal sinus X-ray using multiple deep learning models. Quant Imaging Med. Surg. 2019, 9, 942. [Google Scholar] [CrossRef]
  25. Kuwana, R.; Ariji, Y.; Fukuda, M.; Kise, Y.; Nozawa, M.; Kuwada, C.; Muramatsu, C.; Katsumata, A.; Fujita, H.; Ariji, E. Performance of deep learning object detection technology in the detection and diagnosis of maxillary sinus lesions on panoramic radiographs. Dentomaxillofac. Radiol. 2020, 50, 20200171. [Google Scholar] [CrossRef]
  26. Kim, Y.; Lee, K.J.; Sunwoo, L.; Choi, D.; Nam, C.M.; Cho, J.; Kim, J.; Bae, Y.J.; Yoo, R.E.; Choi, B.S.; et al. Deep Learning in Diagnosis of Maxillary Sinusitis Using Conventional Radiography. Investig. Radiol. 2019, 54, 7–15. [Google Scholar] [CrossRef] [PubMed]
  27. Ohashi, Y.; Ariji, Y.; Katsumata, A.; Fujita, H.; Nakayama, M.; Fukuda, M.; Nozawa, M.; Ariji, E. Utilization of computer-aided detection system in diagnosing unilateral maxillary sinusitis on panoramic radiographs. Dentomaxillofac. Radiol. 2016, 45, 20150419. [Google Scholar] [CrossRef] [PubMed]
  28. Xu, X.; Wang, K.; Wang, C.; Chen, R.; Zhu, F.; Long, H.; Guan, Q. Iterative learning for maxillary sinus segmentation based on bounding box annotations. Multimed. Tools Appl. 2024, 83, 33263–33293. [Google Scholar] [CrossRef]
  29. Yoo, Y.S.; Kim, D.; Yang, S.; Kang, S.R.; Kim, J.E.; Huh, K.H.; Yi, W.J. Comparison of 2D, 2.5 D, and 3D segmentation networks for maxillary sinuses and lesions in CBCT images. BMC Oral Health 2023, 23, 866. [Google Scholar] [CrossRef] [PubMed]
  30. Zeng, P.; Song, R.; Lin, Y.; Li, H.; Chen, S.; Shi, M.; Chen, Z. Abnormal maxillary sinus diagnosing on CBCT images via object detection and ‘straight-forward’classification deep learning strategy. J. Oral Rehabil. 2023, 50, 1465–1480. [Google Scholar] [CrossRef]
  31. Chen, I.D.S.; Yang, C.M.; Chen, M.J.; Chen, M.C.; Weng, R.M.; Yeh, C.H. Deep learning-based recognition of periodontitis and dental caries in dental x-ray images. Bioengineering 2023, 10, 911. [Google Scholar] [CrossRef]
  32. Khan, M.H.; Giri, P.S.; Jothi, J.A.A. Detection of cavities from oral Images using convolutional neural networks. In Proceedings of the 2022 International Conference on Electrical, Computer and Energy Technologies (ICECET), Prague, Czech Republic, 20–22 July 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  33. ForouzeshFar, P.; Safaei, A.A.; Ghaderi, F.; Hashemikamangar, S.S. Dental Caries diagnosis from bitewing images using convolutional neural networks. BMC Oral Health 2024, 24, 211. [Google Scholar] [CrossRef] [PubMed]
  34. Esmaeilyfard, R.; Bonyadifard, H.; Paknahad, M. Dental Caries Detection and Classification in CBCT Images Using Deep Learning. Int. Dent. J. 2024, 74, 328–334. [Google Scholar] [CrossRef] [PubMed]
  35. Liu, Y.; Xia, K.; Cen, Y.; Ying, S.; Zhao, Z. Artificial intelligence for caries detection: A novel diagnostic tool using deep learning algorithms. Oral Radiol. 2024, 40, 375–384. [Google Scholar] [CrossRef]
  36. Szabó, V.; Szabo, B.T.; Orhan, K.; Veres, D.S.; Manulis, D.; Ezhov, M.; Sanders, A. Validation of Artificial Intelligence Application for Dental Caries Diagnosis on Intraoral Bitewing and Periapical Radiographs. J. Dent. 2024, 147, 105105. [Google Scholar] [CrossRef] [PubMed]
  37. Haghanifar, A.; Majdabadi, M.M.; Haghanifar, S.; Choi, Y.; Ko, S.B. PaXNet: Tooth segmentation and dental caries detection in panoramic X-ray using ensemble transfer learning and capsule classifier. Multimed. Tools Appl. 2023, 82, 27659–27679. [Google Scholar] [CrossRef]
  38. Pérez de Frutos, J.; Holden Helland, R.; Desai, S.; Nymoen, L.C.; Langø, T.; Remman, T.; Sen, A. AI-Dentify: Deep learning for proximal caries detection on bitewing x-ray-HUNT4 Oral Health Study. BMC Oral Health 2024, 24, 344. [Google Scholar] [CrossRef]
  39. Imak, A.; Celebi, A.; Siddique, K.; Turkoglu, M.; Sengur, A.; Salam, I. Dental caries detection using score-based multi-input deep convolutional neural network. IEEE Access 2022, 10, 18320–18329. [Google Scholar] [CrossRef]
  40. Basri, K.N.; Yazid, F.; Zain, M.N.M.; Yusof, Z.M.; Rani, R.A.; Zoolfakar, A.S. Artificial neural network and convolutional neural network for prediction of dental caries. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2024, 312, 124063. [Google Scholar] [CrossRef] [PubMed]
  41. Chaves, E.T.; Vinayahalingam, S.; van Nistelrooij, N.; Xi, T.; Romero, V.H.D.; Flügge, T.; Cenci, M.S. Detection of caries around restorations on bitewings using deep learning. J. Dent. 2024, 143, 104886. [Google Scholar] [CrossRef]
  42. Kuwada, C.; Ariji, Y.; Fukuda, M.; Kise, Y.; Fujita, H.; Katsumata, A.; Ariji, E. Deep learning systems for detecting and classifying the presence of impacted supernumerary teeth in the maxillary incisor region on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2020, 130, 464–469. [Google Scholar] [CrossRef]
  43. Durmuş, M.; Ergen, B.; Çelebi, A.; Türkoğlu, M. Panoramik diş görüntülerinde derin evrişimsel sinir ağına dayalı gömülü diş tespiti ve segmentasyonu. Çukurova Üniv. Mühendis. Fak. Derg. 2023, 38, 713–724. [Google Scholar] [CrossRef]
  44. Imak, A.; Çelebi, A.; Polat, O.; Türkoğlu, M.; Şengür, A. ResMIBCU-Net: An encoder–decoder network with residual blocks, modified inverted residual block, and bi-directional ConvLSTM for impacted tooth segmentation in panoramic X-ray images. Oral Radiol. 2023, 39, 614–628. [Google Scholar] [CrossRef] [PubMed]
  45. Celik, M.E. Deep learning based detection tool for impacted mandibular third molar teeth. Diagnostics 2022, 12, 942. [Google Scholar] [CrossRef] [PubMed]
  46. Wang, X.; Li, Y.; Zhang, H.; Shan, Y. Towards real-world blind face restoration with generative facial prior. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9168–9178. [Google Scholar]
  47. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8110–8119. [Google Scholar]
  48. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 694–711, Part II 14. [Google Scholar]
  49. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  50. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  51. Widayani, A.; Putra, A.M.; Maghriebi, A.R.; Adi, D.Z.C.; Ridho, M.H.F. Review of Application YOLOv8 in Medical Imaging. Indones. Appl. Phys. Lett. 2024, 5, 23–33. [Google Scholar] [CrossRef]
  52. Ju, R.Y.; Cai, W. Fracture Detection in Pediatric Wrist Trauma X-Ray Images Using YOLOv8 Algorithm. Sci. Rep. 2023, 13, 20077. [Google Scholar] [CrossRef] [PubMed]
  53. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 16965–16974. [Google Scholar]
  54. Yu, C.; Shin, Y. An Enhanced RT-DETR with Dual Convolutional Kernels for SAR Ship Detection. In Proceedings of the 2024 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Osaka, Japan, 19–22 February 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 425–428. [Google Scholar]
  55. Solovyev, R.; Wang, W.; Gabruseva, T. Weighted boxes fusion: Ensembling boxes from different object detection models. Image Vis. Comput. 2021, 107, 104117. [Google Scholar] [CrossRef]
  56. Wu, J.; Zhang, L.; Zhang, D. An Ensemble Learning and Slice Fusion Strategy for Three-Dimensional Nuclei Instance Segmentation in Medical Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 18–24 June 2022; pp. 1–10. [Google Scholar]
  57. Xiong, Y.; Deng, L.; Wang, Y. Pulmonary nodule detection based on model fusion and adaptive false positive reduction. Expert Syst. Appl. 2024, 238, 121890. [Google Scholar] [CrossRef]
  58. Pelikan, M.; Goldberg, D.E.; Cantú-Paz, E. BOA: The Bayesian optimization algorithm. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), Orlando, FL, USA, 13–17 July 1999; pp. 525–532. [Google Scholar]
  59. Frazier, P.I. A tutorial on Bayesian optimization. arXiv 2018, arXiv:1807.02811. [Google Scholar]
  60. Victoria, A.H.; Maragatham, G. Automatic tuning of hyperparameters using Bayesian optimization. Evol. Syst. 2021, 12, 217–223. [Google Scholar] [CrossRef]
  61. Wu, J.; Chen, X.Y.; Zhang, H.; Xiong, L.D.; Lei, H.; Deng, S.H. Hyperparameter optimization for machine learning models based on Bayesian optimization. J. Electron. Sci. Technol. 2019, 17, 26–40. [Google Scholar]
  62. Klein, A.; Falkner, S.; Bartels, S.; Hennig, P.; Hutter, F. Fast bayesian optimization of machine learning hyperparameters on large datasets. In Proceedings of the Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 528–536. [Google Scholar]
  63. Román, J.C.M.; Fretes, V.R.; Adorno, C.G.; Silva, R.G.; Noguera, J.L.V.; Legal-Ayala, H.; Mello-Román, J.D.; Torres, R.D.E.; Facon, J. Panoramic dental radiography image enhancement using multiscale mathematical morphology. Sensors 2021, 21, 3110. [Google Scholar] [CrossRef] [PubMed]
  64. Abdi, A.H.; Kasaei, S.; Mehdizadeh, M. Automatic segmentation of mandible in panoramic X-ray. J. Med. Imaging 2015, 2, 044003. [Google Scholar] [CrossRef]
Figure 1. Schema of the proposed architecture.
Figure 1. Schema of the proposed architecture.
Diagnostics 15 00244 g001
Figure 2. Sample labeled images.
Figure 2. Sample labeled images.
Diagnostics 15 00244 g002
Figure 3. Visual estimation results obtained from experimental studies.
Figure 3. Visual estimation results obtained from experimental studies.
Diagnostics 15 00244 g003aDiagnostics 15 00244 g003b
Figure 4. Sample images illustrating the limitations of the proposed model.
Figure 4. Sample images illustrating the limitations of the proposed model.
Diagnostics 15 00244 g004
Table 1. The hyperparameters and search ranges.
Table 1. The hyperparameters and search ranges.
HyperparametersSearch Ranges
IoU threshold0.1–0.9
Skip box threshold0.001–0.1
Weight1 (YOLO weight)1–10
Weight2 (RT-DETR weight)1–10
Table 2. Normal and GAN-based individual performance results.
Table 2. Normal and GAN-based individual performance results.
ModelsNormal ImagesGAN Images
mAP@0.5F1mAP@0.5F1
Yolov8L91.394.696.594.5
Yolov8X88.893.788.893.7
RT-DETR-L95.288.294.792.3
RT-DETR-X94.292.292.391.7
Table 3. Ensemble performance results with default WBF algorithm.
Table 3. Ensemble performance results with default WBF algorithm.
InputHybrid ModelsmAP@0.5F1PrecisionRecall
Normal ImagesYolov8L + RT-DETR-L95.488.882.796
Yolov8X + RT-DETR-L95.689.583.996
Yolov8L + RT-DETR-X95.394.99396.8
Yolov8X + RT-DETR-X94.694.894.495.2
GAN ImagesYolov8L + RT-DETR-L97.492.386.698.4
Yolov8X + RT-DETR-L95.59389.696.8
Yolov8L + RT-DETR-X97.593.589.198.4
Yolov8X + RT-DETR-X94.694.894.495.2
Table 4. Performance results of the optimized models.
Table 4. Performance results of the optimized models.
InputHybrid ModelsmAP@0.5F1PrecisionRecall
GAN ImagesYolov8L + RT-DETR-L 98.392.586.799.2
Yolov8L + RT-DETR-X 97.59693.898.4
Table 5. Comparison of studies on impacted tooth detection.
Table 5. Comparison of studies on impacted tooth detection.
ReferenceDatasetModelsPerformance Results
[3]1084 panoramic radiographsFaster R-CNN Inception v2F1 score: 86.25%
[44]440 panoramic radiographsAlexNet-Faster R-CNNmAP: 86%
VGG16-Faster R-CNNmAP: 87%
ResNet50-Faster R-CNNmAP: 91%
YOLO v3mAP: 96%
Our model407 panoramic radiographsGAN based Yolov8L + RT-DETR-LmAP: 98.3%
F1 score: 92.5%
GAN based Yolov8L + RT-DETR-XmAP: 97.5%
F1 score: 96%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Küçük, D.B.; Imak, A.; Özçelik, S.T.A.; Çelebi, A.; Türkoğlu, M.; Sengur, A.; Koundal, D. Hybrid CNN-Transformer Model for Accurate Impacted Tooth Detection in Panoramic Radiographs. Diagnostics 2025, 15, 244. https://doi.org/10.3390/diagnostics15030244

AMA Style

Küçük DB, Imak A, Özçelik STA, Çelebi A, Türkoğlu M, Sengur A, Koundal D. Hybrid CNN-Transformer Model for Accurate Impacted Tooth Detection in Panoramic Radiographs. Diagnostics. 2025; 15(3):244. https://doi.org/10.3390/diagnostics15030244

Chicago/Turabian Style

Küçük, Deniz Bora, Andaç Imak, Salih Taha Alperen Özçelik, Adalet Çelebi, Muammer Türkoğlu, Abdulkadir Sengur, and Deepika Koundal. 2025. "Hybrid CNN-Transformer Model for Accurate Impacted Tooth Detection in Panoramic Radiographs" Diagnostics 15, no. 3: 244. https://doi.org/10.3390/diagnostics15030244

APA Style

Küçük, D. B., Imak, A., Özçelik, S. T. A., Çelebi, A., Türkoğlu, M., Sengur, A., & Koundal, D. (2025). Hybrid CNN-Transformer Model for Accurate Impacted Tooth Detection in Panoramic Radiographs. Diagnostics, 15(3), 244. https://doi.org/10.3390/diagnostics15030244

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop