Next Article in Journal
Regional Variations in the Intra- and Intervertebral Trabecular Microarchitecture of the Osteoporotic Axial Skeleton with Reference to the Direction of Puncture
Previous Article in Journal
Dual-Energy CT Iodine Uptake of Head and Neck: Definition of Reference Values in a Big Data Cohort
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

STSN-Net: Simultaneous Tooth Segmentation and Numbering Method in Crowded Environments with Deep Learning

1
Department of Orthodontics, Beijing Stomatological Hospital, Capital Medical University, Beijing 100050, China
2
School of Biomedical Engineering, Capital Medical University, Beijing 100069, China
3
Laboratory for Clinical Medicine, Capital Medical University, Beijing 100069, China
4
Beijing Key Laboratory of Fundamicationental Research on Biomechanics in Clinical Application, Capital Medical University, Beijing 100069, China
5
LargeV Instrument Corp., Ltd., Beijing 100084, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Diagnostics 2024, 14(5), 497; https://doi.org/10.3390/diagnostics14050497
Submission received: 14 December 2023 / Revised: 25 January 2024 / Accepted: 1 February 2024 / Published: 26 February 2024
(This article belongs to the Special Issue Impact of Deep Learning in Biomedical Engineering)

Abstract

:
Accurate tooth segmentation and numbering are the cornerstones of efficient automatic dental diagnosis and treatment. In this paper, a multitask learning architecture has been proposed for accurate tooth segmentation and numbering in panoramic X-ray images. A graph convolution network was applied for the automatic annotation of the target region, a modified convolutional neural network-based detection subnetwork (DSN) was used for tooth recognition and boundary regression, and an effective region segmentation subnetwork (RSSN) was used for region segmentation. The features extracted using RSSN and DSN were fused to optimize the quality of boundary regression, which provided impressive results for multiple evaluation metrics. Specifically, the proposed framework achieved a top F1 score of 0.9849, a top Dice metric score of 0.9629, and an mAP (IOU = 0.5) score of 0.9810. This framework holds great promise for enhancing the clinical efficiency of dentists in tooth segmentation and numbering tasks.

1. Introduction

Tooth segmentation and numbering are of immense significance in dental care, owing to their critical role in assessing oral health and facilitating precise diagnosis [1]. Panoramic radiograph (PR), a two-dimensional (2D) X-ray that captures the entire mouth in a single image, is commonly used for the clinical diagnosis of tooth segmentation and numbering [2]. Cone beam computed tomography (CBCT) is an imaging technology that provides three-dimensional (3D) images of the teeth, soft tissues, nerve pathways, and bone in a single scan [3]. While CBCT offers superior details and a 3D perspective, panoramic radiographs exhibit advantages such as lower radiation exposure and reduced imaging costs [4]. Furthermore, the latter is more commonly used, owing to its simplicity and effectiveness in general dental evaluations, routine screenings, and initial diagnostic tasks [5]. Panoramic radiography is particularly useful in identifying dental caries and periodontal diseases and assessing tooth orientations, making it a versatile tool in a range of common clinical scenarios [6,7]. Obtaining tooth segmentation and numbering is a critical aspect of dental care and is typically performed by dental professionals to ensure accurate assessment and treatment [8]. Normally, the accurate segmentation and numbering of teeth is difficult and labor-intensive. The task becomes even more complicated when using panoramic dental X-ray images as overlapping boundaries between the teeth pose difficulties in the annotation [9]. In addition, this procedure is often performed manually, which can be quite time-consuming. The segmentation and numbering are not objective, and there is no guarantee of the accuracy of the results as it relies heavily on the subjective judgment of the dentist [10].
With the advancements in image processing technology and artificial intelligence, automatic tooth segmentation and numbering have become feasible and have significantly alleviated the burden of annotation for dentists [11]. Nevertheless, panoramic X-ray images of different patients’ teeth exhibit considerable variations, encompassing diverse pathological presentations and treatment indications [12]. In addition, the background of panoramic X-ray images can be quite complex, incorporating redundant background information, which may lead to incorrect segmentation [13]. The accurate segmentation and numbering of teeth in panoramic X-ray images become quite challenging because of these issues, thus impeding the practical utility of automated intelligent segmentation and numbering methods in clinical settings. To overcome these challenges, there is a critical need to develop advanced image processing techniques that can distinguish teeth from their complex backgrounds and accommodate variations in dental conditions, ultimately enhancing the reliability and applicability of automated dental image analysis tools in real-world clinical scenarios.
The accurate segmentation of objects (including tooth segmentation) from images is a key application of computer vision technology [14] and is the cornerstone of quantitative analysis (for example, numbering) that aims to cluster the pixels of a specific category [15]. Since the development of CNN, the use of conventional segmentation methods has declined [16]. Various CNN-based structures have been proposed and applied in semantic segmentation, of which fully convolutional networks (FCNs) that have an encoder–decoder structure are a milestone [17]. Since then, several types of CNN architectures, such as ResNet [18], ResNext [19], VGG [20], and GoogleNet [21], have been used as encoders and decoders. U-net [22] is based on FCN [23] and exhibits the advantages of good performance, low data requirement, and high speed. Various segmentation methods based on deep convolutional neural networks (CNNs), including crowd counting [24,25], text recognition [26,27], and medical image analysis [3,28,29,30,31], have been proposed and are widely used to obtain statistics on target objects [32,33,34].
To further enhance the clinical utility of panoramic X-rays, the development of automated tooth segmentation and numbering methods has gained considerable attention. Segmentation of individual teeth and their subsequent numbering can streamline the diagnostic process, allowing dentists to rapidly and accurately assess tooth morphology and spatial relationships. This approach is particularly valuable in identifying tooth loss, a critical concern, especially in the transitional dentition phase, where early intervention can mitigate the severity of permanent tooth alignment issues. Moreover, the integration of automated assistance in dental diagnostics substantially reduces the likelihood of diagnostic errors by oral health practitioners. Such automation aids in the meticulous documentation of patient records, thereby enhancing the overall quality of dental care. Elif Bilgir et al. proposed the Faster R-CNN Inception v2 model to predict the label and bounding box of the 2482 panoramic radiographs, with an average accuracy of 0.9652 [8]. Münevver Coruh Kılıc et al. devised a method using Faster R-CNN Inception v2 (COCO) for the automatic detection and numbering of deciduous teeth on pediatric panoramic radiographs based on 421 panoramic images to improve the accuracy of deciduous tooth position recognition [35]. André Ferreira Leite et al. improved the accuracy of tooth segmentation and recognition in curved panoramic radiographs of permanent dentition based on 153 radiographs using two deep CNNs [11]. Sheng improved the accuracy of the detection and localization of tooth tissue on panoramic radiographs using the PLAGH-BH tooth segmentation [36]. However, the above methods treated each tooth as a uniform class and failed to incorporate information on distinct tooth shapes and positions in the analysis framework. Moreover, the features were extracted from entire panoramic radiographs without any prior constraints; thus, the ratio of false positives may be high.
To overcome these problems, the proposed framework presents a two-stage network architecture (named STSN-Net) to simultaneously segment the teeth and number them, innovatively delineating the processes into bone window segmentation and tooth segmentation with numbering. The proposed approach can considerably improve the accuracy of feature learning and efficiently eliminate redundant data. The initial stage employed a graph convolution neural network (GCN) to obtain the structure information hidden in the PRs for precise bone window segmentation. The second stage incorporated a shared 2D CNN, augmented by a dual-stage detection subnetwork (DSN) and a sophisticated refinement segmentation subnetwork (RSSN), adding robustness to the framework. The following are the key contributions of this work:
  • A GCN for automatic bone window segmentation was proposed to generate masks of regions to train the RSNN. This process leveraged dental segmentation annotations for initial calibration, followed by feature extraction and mask refinement via GCN.
  • A modified DSN was proposed, which made use of a dilated convolution module in certain specific layers to enlarge the receptive field and employed a deformable convolution block to compensate for the diversity in tooth morphology and distribution among individuals.
  • The proposed RSSN aimed to accurately segment the tooth region within the images, simultaneously suppressing the background noise and resulting in refined predictions that substantially reduced analytical errors.
  • The high precision and strong practicability of the proposed STSN-Net were demonstrated. The proposed two-stage structure initially deployed a GCN for bone window extraction, followed by a multinetwork structure that performed object detection and refinement segmentation, respectively, and simultaneously. Furthermore, STSN-Net was adeptly optimized and adapted for real-world clinical applications in embedded devices.
The rest of the paper comprises four parts. The materials and methods are described in Section 2. Section 3 summarizes the experimental results. The discussion is presented in Section 4, and the conclusion is presented in Section 5.

2. Materials and Methods

2.1. Dataset TSNDS

This paper introduces a new dataset, the tooth segmentation and numbering dataset (TSNDS), which comprises 2116 oral panoramic X-ray images from patients with permanent dentition, encompassing 2D imaging data points of all teeth, the mandible, and parts of the maxilla. The initial 116 images were obtained from the publicly available dataset, and the subsequent 2000 images were retrospectively selected from panoramic radiographs acquired at the Beijing Stomatological Hospital between March 2019 and April 2022. These 2000 PRs were produced using a Sirona digital machine (ORTHOPHOS XG 5 DS Ceph, Bensheim, Germany) with standard parameters, operating with tube voltages of 60–90 kV and tube operating currents of 1–10 mA. A default program of the device with a predetermined magnification of 1.9× and a rotation time of 14.1 s was used for X-ray exposures with a resolution of 2440 × 1280, and the PRs were in portable network graphics (PNG) format. During the data processing phase, the original DICOM files inherently contained sensitive patient information, such as name and age. To address privacy issues, a method was adopted for converting the image data into the PNG file format for storage. The dataset covers a wide range of dental conditions as observed in dental radiographs, which can be broadly categorized into two main types: permanent dentition and primary dentition. Within these categories, various conditions may coexist, including missing teeth, dental crowding, impacted teeth, and teeth with fillings or restorations. Annotation was performed using the online tool “Label Studio”, ensuring a consistent and precise approach to data labeling. Two dental practitioners, each with over 15 years of clinical experience, independently annotated the radiographs. To ensure that the annotations were accurate and reliable, a senior dentist with more than 25 years of clinical experience reviewed and revised the data. The finalized annotated dataset was then used in the study. The annotations adhered to the FDI tooth notation system established by the International Dental Federation in 1970. Each image within the dataset included segmentation annotations for individual teeth and bone windows, along with eight types of positional information on the teeth. The dataset was categorized into three parts: a training part (Train), a validation part (Validation), and a test part (Test) using the train–test–split function (TTSF) of the scikit-learn library. The distribution and details of the dataset split are presented in Table 1 and Table 2.

2.2. Data Augmentation

The panoramic X-ray images were subjected to a multitude of enhancement processes to augment the dataset while preserving relevant dental features. Via this augmentation pipeline (as shown in Figure 1), diverse transformations such as rotation, scaling, flipping, and contrast adjustments were applied, which facilitated the extraction of robust features and enhanced the ability of the model to generalize. The effectiveness of these augmentation strategies was evaluated and demonstrated via comprehensive experimentation with the section results, showcasing notable improvements in the learning capacity and generalization of the model across diverse dental conditions and variations.

2.3. The Pipeline of the Proposed STSN-Net

The proposed STSN-Net for simultaneous tooth segmentation and numbering comprised three parts: an automatic annotation method for bone window segmentation, a modified DSN that predicted the bounding boxes and number of teeth, and an RSSN that extracted refinement segmentation information from the regions of interest (as shown in Figure 2). The details of these three parts of the proposed method are described in the following subsections.

2.4. Automatic Segmentation of Mandibular Morphology

To obtain an accurate boundary annotation of the target area, a novel network architecture was proposed. The annotation of the bounding boxes was used as an initialization. The spatial and gray-level information were subsequently propagated via a multilayer GCN. In the initialization phase, N points were selected randomly to form a graph G. The propagation process for a node g i was defined as follows:
f i ( n + 1 ) = w 0 n f i n + g i N ( g i ) w 1 n w j n
where w 0 n and w 1 n represent the weight matrices and N ( g i ) is a set that contains all the nodes connected to node g i . The parameters of the pretrained GCN model were fed into the model for rapid convergence. After the initialization, the nodes extracted from the edges of the segmentation annotations were used as target values for mandibular morphology segmentation. The network structure and processing pipeline are illustrated in Figure 3.

2.5. STSN-Net Structure

The proposed STSN-Net architecture comprised three blocks (shared convolution layers, a modified DSN, and an RSSN), as shown in Figure 2. Block I represents the two-layer shared convolution layers; 64 convolution filters with a size of 3 × 3 were used in this subnetwork to generate feature maps with 64 channels. The inputs to the shared convolution layers were the PRs, and the output was the feature map named shared features. The extracted shared features, usually identified as shallow characteristics, were concurrently inputted into Block II and III to enhance the efficiency of feature reuse and considerably reduce the overall model complexity and size. Block II was the proposed DSN, which had an improved convolution filter capable of adapting to perspective changes and obtaining a large receptive field. Block III was the proposed RSSN, which distinguished between the foreground and background while generating the instance mask of each tooth from the previous mandibular morphology segmentation generated by the GCN block described in Section 2.4 and shared spatial information on the target region with the DSN via an element-wise sum operation before the RPN in the DSN.

2.6. Detection Sub-Network

The proposed DSN is a modified two-stage detection network. The input of the subnetwork was generated using a two-layer shared convolution block in the STSN-Net structure. The DSN is also a multitask network that has certain shared convolution layers and two parts that perform recognition and bounding box regression. In the shared convolution layers, the normal 3 × 3 convolution filter was replaced with a deformable convolution filter in the second layer to handle the different perspectives. The deformable convolution filter comprised two parts: the convolution and the offset. It is a variant of the normal convolution filter. The formulation of the deformable convolution filter was defined as follows:
y ( p 0 ) = p n R w ( p n ) · x ( p 0 + p n + Δ p n )
where p n is a location on the feature map y, and Δ p n is an extra offset added to the input feature map. In this way, the convolution filter extracts features in a deformable manner to handle the changes in the perspective. A dilated convolution filter, which was originally used for semantic segmentation, was employed in the fifth layer to obtain a large receptive field. This filter inserted empty cells into the convolution output. This procedure was defined as follows:
f n + 1 = D ( f n )
where f n represents the input feature map of the dilated convolution layer, f n + 1 represents the output feature map of the dilated convolution layer, and D represents the dilated filter that filters the input feature map by inserting cells with a value of 0 in the convolution output. The filter size was increased from 3 × 3 to 5 × 5 so as to expand the receptive field without adding extra computational burden. The architecture of the proposed DSN is shown in Figure 4.

2.7. Refinement Segmentation Sub-Network

The proposed RSSN is a multiscale CNN-based network that comprised three components: a skip connection component that fused the shallow feature map with the deep feature map, a max pool component that performed downsampling of the high-resolution feature map to obtain global features, and an up-convolution component that restored scale information while generating the segmentation map. As depicted in Figure 5, a max pooling filter with a size of 2 × 2 was adopted for downsampling, a convolution filter with a size of 3 × 3 was employed to extract the features in each subnet, an up-convolution filter was applied to perform upsampling, and the skip connection operation was performed using the copy and paste operation. In the last phase, a 1 × 1 convolution filter was used to reduce the dimension of the feature map and generate the final segmentation map.

2.8. Loss Function

The entire framework was trained end-to-end using a batch training method. The loss function of the network consisted of two main parts, defined as follows:
L all = α L det + β L seg
where α and β are the weight parameters of the two tasks. In this study, the values are 0.8 and 0.2, respectively.
The loss of the DSN was defined as follows:
L det = L classification + L reg
The classification loss of the DSN is a logarithmic loss, which was defined as follows:
L classification = log ( p i )
where p i represents the probability of the output belonging to class i. The regression loss of the DSN was defined by a smooth L1 function, as follows:
L reg = i { x , y , w , h } smooth L 1 ( t i k g i )
where t i k is a set x i k , y i k , w i k , h i k that forms a bounding box of class k, and g i represents the ground truth of the predicted value.
The smooth L1 function was defined as follows:
smooth L 1 ( x ) = 0.5 x 2 if | x | < 1 | x | 0.5 otherwise
The segmentation loss is a cross-entropy (CE) loss function, defined as follows:
L seg = CE ( m p , m g )
where m p is the predicted segmentation map and m g is the ground truth of the mask map. The CE function for each pixel in the two-class segmentation map was defined as follows:
CE ( p , p ^ ) = p log ( p ^ ) + ( 1 p ) log ( 1 p ^ )
where p is the ground truth probability of the target belonging to class 0, and p ^ is the prediction probability of the target belonging to class 1. The optimization function was obtained by the stochastic gradient descent (SGD) method.

3. Results

3.1. Evaluation Metrics

The F1 score was used as the evaluation metric because it considers the accuracy and recall of the model and can be regarded as a harmonic average of the model accuracy and recall for tooth numbering. R was used to represent recall, and P was used to represent accuracy. R was defined as follows:
R = T P + T N N
P is defined as follows:
P = T P + T N T P + T N + F P + F N
where N represents the number of all targets. TP represents the true positive, TN represents the true negative, FP represents the false positive, and FN represents the false negative. True and false were determined by calculating the numbering results and the intersection over union (IOU), which is a metric that compares the similarity and diversity of the detection and segmentation results. The IOU was defined as follows:
IOU = Area of overlap Area of Union
where the area of overlap is the overlap between the predicted bounding box and the actual bounding box, and the area of union is the union of the predicted bounding box and the ground truth. A threshold parameter was set to determine whether the result was true. If the IOU score was >0.5, it was considered a true detection, otherwise, it was a false detection. The DICE coefficient is a common evaluation metric used in segmentation tasks to measure the accuracy of segmentation results. It quantifies the degree of overlap between the predicted segmentation and the ground truth. The DICE coefficient was defined as follows:
D I C E = 2 × | A B | | A | + | B |
where A is the prediction and B is the ground truth. The F1 score was defined as follows:
F 1 - s c o r e = 2 P R P + R

3.2. Experiment Results

Multiple experiments were performed to verify the contribution of each part of the proposed method. In our experiments, a two-stage architecture network was adopted, with a backbone network of convolutional neural networks as the baseline. Furthermore, the performance achieved using the multitask learning instance segmentation framework named Mask RCNN [37] and the single-stage detection architecture named YOLOv8 [38] on the dataset was determined for a comprehensive comparison.
Each part of the proposed method was evaluated separately to obtain a better understanding of the proposed method. Common image augmentation operations, including rotation, horizontal flipping, and vertical flipping, were performed to enlarge the training dataset. Network training was performed with two NVIDIA GeForce RTX 4090 GPUs (NVIDIA Corporation: Santa Clara, CA, USA) for 30 epochs. The learning rate was initialized as 0.01, and the factor was decreased by 0.1 after 24 and 29 epochs, respectively. A batch training scheme utilizing the stochastic gradient descent (SGD) optimizer was adopted for training, with a batch size of 16. The base width and height of the images were altered to 1333 and 800, respectively, without changing the aspect ratio. A multiscale training scheme was adopted in the training procedure, with scaling factors of 1.0×, 1.2×, 1.4×, and 1.6×. In the test phase, each scale was evaluated separately, and the same scaling factors were used (Table 3). F1 represents the F1 score as defined in Section 3.1. The following test results were obtained using a single GPU (4090). The experimental results are presented in Table 3.
The F1 score measures the harmonic mean of precision and recall, reflecting the accuracy and robustness of the model in detecting instances (as shown in Figure 6). Furthermore, it is imperative to focus on instance segmentation metrics, including mAP (mean average precision) and DICE (Dice similarity coefficient). The mAP metric evaluates the overall detection accuracy across multiple object categories, whereas DICE primarily concerns pixel-level similarity in image segmentation tasks. The experimental results of mAP metrics (IOU = 0.5) and DICE scores are presented in Table 4.
The mAP (object detection) results are illustrated in Figure 7 for clear representation.
The mAP (instance segmentation) results are illustrated in Figure 8 for clear representation.
The DICE score results are illustrated in Figure 9 for clear representation.
The inference times in milliseconds (ms) are depicted in Table 5.

4. Discussion

This study demonstrated the use of an STSN-Net framework comprising a modified DSN and an RSSN to obtain accurate segmentation and numbering of teeth in a crowded environment. The performances of the deformable convolution filter and dilated convolution filter in the DSN were analyzed. These filters ensured that different perspectives could be handled and that the receptive field was sufficiently large for a better representation of the panoramic X-ray images. The RSSN aided in determining the region with the highest probability of occurrence of the target while simultaneously reducing the rate of false positives. The STSN-Net architecture improved the performance of the detection task using information provided by the RSSN with an element-wise sum operation. Block I was a shared convolution block that extracted primary features, such as edges, lines, and corners. The extracted shared features were characterized as shallow features, which were generic in different networks and universally applicable for downstream tasks. These extracted features served as the building blocks for more complex patterns in the subsequent DSN and RSSN, which enhanced the feature reusability and reduced the model size. The inclusion of Block I in the network architecture facilitated initial data processing and ensured that downstream blocks received preprocessed, feature-rich inputs. This strategy not only streamlined the overall computational process but also enhanced the ability of the network to generalize across different tasks by providing a versatile feature set. Block II was the modified DSN that employed the deformable convolution filter and dilated convolution filter to handle various perspectives of the target and provided a large receptive field to understand the contextual information around the target. Block III was the RSSN that simultaneously processed the shared feature maps. Multiscale features were extracted and fused in a specific manner. A 3 × 3 convolution filter was used to extract the features while maintaining a high resolution. Features of different scales were fused using max-pooling, and up-convolution was performed depending on the source and target resolution. A skip connection was employed to make full use of the features at each resolution while reducing the number of network parameters. The RSSN provided useful information on the spatial location of the target and not only optimized the quality of the proposed regions obtained from the DSN but also minimized the background noise. Overall, the STSN-Net framework provided excellent performance for multiple tasks, and the F1 score of the detection task was 98.49% (best F1 score), the mAP (object detection) score was 98.90%, the mAP (instance segmentation) score was 98.10%, and the DICE score was 96.29%, which outperformed the two state-of-the-art models (Mask RCNN with Swin Transformer backbone network and the YOLOV8 for instance segmentation).
In addition to the aforementioned features of the STSN-Net framework, it is crucial to address the challenge of class imbalance in dental imaging, notably, the less frequent occurrence of certain tooth types. Class imbalance is a significant concern in medical imaging, as it can skew the performance of the model, leading to suboptimal detection and segmentation of less-represented classes. The proposed STSN-Net framework tackled this challenge in multiple ways. First, the modified DSN, with its deformable and dilated convolution filters, was adept at handling varying perspectives and tooth morphologies. This flexibility ensured that even less common tooth types were accurately captured and represented, thus mitigating the impact of class imbalance on the performance of the model.
Furthermore, the ability of RSSN to focus on regions with the highest probability of target occurrence played a pivotal role in addressing class imbalance. By prioritizing the regions likely to contain the target, the RSSN effectively enhanced the detection of underrepresented tooth types. This selective focus, combined with the reduction of false positives, ensured a balanced treatment of all tooth classes in the segmentation and numbering process.
Additionally, the multiscale feature extraction and fusion in the RSSN contributed to this balance. By extracting and combining features at different scales, the network ensured high-resolution representations of all tooth types regardless of their frequency in the dataset. The use of skip connections further guaranteed that features from each resolution were fully utilized, thereby augmenting the sensitivity of the network to less frequent tooth types.
Furthermore, most existing studies in dental arch classification rely on bounding box annotations without segmenting teeth based on their morphology. Additionally, samples with significant dental crowding are often excluded, which leads to reduced practicality in clinical application. In contrast, our study utilized a dataset comprising panoramic radiographs across various age groups. This dataset encompassed common pathological imaging features encountered in clinical practice, such as deciduous teeth, residual crowns, roots, restorations, root canal treatments, periapical lesions, impacted teeth, implants, brackets, and titanium plates. Because of the inclusion of these features, our model adeptly handled various complex clinical scenarios, thus enhancing its effectiveness in segmentation and tooth position recognition.
The application of artificial intelligence in automatic segmentation and identification of teeth in panoramic radiographs has considerably boosted diagnostic efficiency across various dental specialties. Rapid tooth position identification has assisted clinicians in diagnosing the number and location of missing teeth, thus preventing diagnostic errors, aiding in medical decision-making, and facilitating the documentation of case histories. In large-scale epidemiological surveys, automated tooth position recognition has permitted the analysis of extensive panoramic radiograph data, providing valuable insights into the oral health status of specific populations at a reduced cost.
Moreover, rapid automatic tooth segmentation serves as the foundation for various AI-assisted diagnostic models. By accurately identifying the tooth morphology, the crown–root ratio can be calculated, the completeness of tooth crowns can be ascertained, the type of impaction in wisdom teeth can be determined, and alveolar bone height can be assessed. In instances of radiographic overlap between teeth, segmentation aids in distinguishing pathological issues or restorations specific to each tooth. Moreover, segmenting the alveolar ridge line enables the quick assessment of alveolar bone height and, combined with tooth segmentation results, facilitates the diagnosis of periodontal conditions.

5. Conclusions

This study has described a unique STSN-Net architecture for the accurate segmentation and numbering of teeth with a best F1 score of 98.49%, an mAP score of 98.10%, and a DICE score of 96.29%, which is well suited for use in oral clinical practice and has the potential to assist doctors. This method can help doctors obtain an objective and accurate segmentation and numbering of the teeth and standardize the oral clinical process while simultaneously reducing their workload. The proposed STSN-Net framework for automatic segmentation and tooth numbering in panoramic radiographs marks a significant advancement in augmenting diagnostic and therapeutic efficiency across various dental specialties. Furthermore, in the context of large-scale epidemiological studies, the proposed STSN-Net offers a promising avenue to analyze extensive datasets of panoramic radiographs. Such automation aids in acquiring comprehensive oral health data for specific populations at a reduced cost. In the future, the performance of this method should be evaluated in more clinical practices to further validate the application potential of the proposed framework.

Author Contributions

Conceptualization, X.X. and Y.G.; methodology, S.W. and S.L.; investigation, S.W. and S.L.; resources, X.X. and Y.G.; data curation, S.W., X.X. and Y.B.; visualization, Q.C., L.Z., B.G., F.Z. and Y.W.; writing—original draft preparation, S.W. and S.L.; writing—review and editing, S.W., S.L., Y.B., X.X. and Y.G.; supervision, X.X. and Y.G.; project administration, X.X. and Y.G.; funding acquisition, Y.B., X.X. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number: 62276174; the Beijing Municipal Natural Science Foundation, grant number: Z220015; the Natural Science Foundation of Guangdong Province, grant number: 2021A1515011970; the Beijing Hospitals Authority Clinical medicine Development of special funding support, grant number: ZLRK202330; the Beijing Stomatological Hospital, grant number: CXTD202203; and the Beijing Hospitals Authority, grant number: PG2022029.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Ethics Committee of the Beijing Stomatological Hospital (protocol code: CMUSH-IRB-KJ-PJ-2022-49 and 7 December 2022).

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to thank the National Natural Science Foundation of China (Grant No. 62276174), the Beijing Municipal Natural Science Foundation (Grant No. Z220015), Natural Science Foundation of Guangdong Province (Grant No. 2021A1515011970), the Beijing Stomatological Hospital (Grant No. CXTD202203), the Beijing Hospitals Authority Clinical medicine Development of special funding support (Grant No.ZLRK202330), the Beijing Hospitals Authority (Grant No. PG2022029), and the Laboratory for Clinical Medicine, Capital Medical University (Grant No. 1300-12301210) for their support.

Conflicts of Interest

Author Feifei Zuo and Yajie Wang were employed by the company LargeV Instrument Corp. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Chandrashekar, G.; AlQarni, S.; Bumann, E.E.; Lee, Y. Collaborative deep learning model for tooth segmentation and identification using panoramic radiographs. Comput. Biol. Med. 2022, 148, 105829. [Google Scholar] [CrossRef]
  2. Kato, C.N.; Barra, S.G.; Tavares, N.P.; Amaral, T.M.; Brasileiro, C.B.; Mesquita, R.A.; Abreu, L.G. Use of fractal analysis in dental images: A systematic review. Dentomaxillofac. Radiol. 2020, 49, 20180457. [Google Scholar] [CrossRef]
  3. Mureșanu, S.; Almășan, O.; Hedeșiu, M.; Dioșan, L.; Dinu, C.; Jacobs, R. Artificial intelligence models for clinical usage in dentistry with a focus on dentomaxillofacial CBCT: A systematic review. Oral Radiol. 2023, 39, 18–40. [Google Scholar] [CrossRef]
  4. Schroder, Â.G.D.; Gonçalves, F.M.; Germiniani, J.d.S.; Schroder, L.D.; Porporatti, A.L.; Zeigelboim, B.S.; de Araujo, C.M.; Santos, R.S.; Stechman-Neto, J. Diagnosis of TMJ degenerative diseases by panoramic radiography: Is it possible? A systematic review and meta-analysis. Clin. Oral Investig. 2023, 27, 6395–6412. [Google Scholar] [CrossRef]
  5. Rozylo-Kalinowska, I.; Rozylo-Kalinowska, I. Panoramic Radiography in Dentistry. In Imaging Techniques in Dental Radiology: Acquisition, Anatomic Analysis and Interpretation of Radiographic Images; Springer: Cham, Switzerland, 2020; pp. 43–56. [Google Scholar]
  6. Hlongwa, P.; Moshaoa, M.A.L.; Musemwa, C.; Khammissa, R.A.G. Incidental Pathologic Findings from orthodontic pretreatment panoramic radiographs. Int. J. Environ. Res. Public Health 2023, 20, 3479. [Google Scholar] [CrossRef]
  7. Heller, Z.A.; Hogge, M.; Ragan, M.R.; Portnof, J.E. Applications of Cone Beam Computed Tomography Scans in Dental Medicine and Potential Medicolegal Issues. Dent. Clin. 2024, 68, 55–65. [Google Scholar] [CrossRef]
  8. Bilgir, E.; Bayrakdar, İ.Ş.; Çelik, Ö.; Orhan, K.; Akkoca, F.; Sağlam, H.; Odabaş, A.; Aslan, A.F.; Ozcetin, C.; Kıllı, M.; et al. An artifıcial ıntelligence approach to automatic tooth detection and numbering in panoramic radiographs. BMC Med. Imaging 2021, 21, 124. [Google Scholar] [CrossRef] [PubMed]
  9. Çelik, B.; Çelik, M.E. Automated detection of dental restorations using deep learning on panoramic radiographs. Dentomaxillofac. Radiol. 2022, 51, 20220244. [Google Scholar] [CrossRef] [PubMed]
  10. AlQarni, S.; Chandrashekar, G.; Bumann, E.E.; Lee, Y. Incremental Learning for Panoramic Radiograph Segmentation. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, UK, 11–15 July 2022; pp. 557–561. [Google Scholar]
  11. Leite, A.F.; Gerven, A.V.; Willems, H.; Beznik, T.; Lahoud, P.; Gaêta-Araujo, H.; Vranckx, M.; Jacobs, R. Artificial intelligence-driven novel tool for tooth detection and segmentation on panoramic radiographs. Clin. Oral Investig. 2021, 25, 2257–2267. [Google Scholar] [CrossRef] [PubMed]
  12. Yüksel, A.E.; Gültekin, S.; Simsar, E.; Özdemir, Ş.D.; Gündoğar, M.; Tokgöz, S.B.; Hamamcı, İ.E. Dental enumeration and multiple treatment detection on panoramic X-rays using deep learning. Sci. Rep. 2021, 11, 12342. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, C.; Jeong, H.; Park, W.; Kim, D. Tooth-Related Disease Detection System Based on Panoramic Images and Optimization Through Automation: Development Study. JMIR Med. Inform. 2022, 10, e38640. [Google Scholar] [CrossRef] [PubMed]
  14. Yin, X.X.; Sun, L.; Fu, Y.; Lu, R.; Zhang, Y. U-Net-Based medical image segmentation. J. Healthc. Eng. 2022, 2022, 4189781. [Google Scholar] [CrossRef]
  15. Estai, M.; Tennant, M.; Gebauer, D.; Brostek, A.; Vignarajan, J.; Mehdizadeh, M.; Saha, S. Deep learning for automated detection and numbering of permanent teeth on panoramic images. Dentomaxillofac. Radiol. 2022, 51, 20210296. [Google Scholar] [CrossRef] [PubMed]
  16. Tuzoff, D.V.; Tuzova, L.N.; Bornstein, M.M.; Krasnov, A.S.; Kharchenko, M.A.; Nikolenko, S.I.; Sveshnikov, M.M.; Bednenko, G.B. Tooth detection and numbering in panoramic radiographs using convolutional neural networks. Dentomaxillofac. Radiol. 2019, 48, 20180051. [Google Scholar] [CrossRef] [PubMed]
  17. Lee, J.H.; Han, S.S.; Kim, Y.H.; Lee, C.; Kim, I. Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2020, 129, 635–642. [Google Scholar] [CrossRef] [PubMed]
  18. Ahn, Y.; Hwang, J.J.; Jung, Y.H.; Jeong, T.; Shin, J. Automated mesiodens classification system using deep learning on panoramic radiographs of children. Diagnostics 2021, 11, 1477. [Google Scholar] [CrossRef]
  19. Chen, M.; Zhao, C.; Tian, X.; Liu, Y.; Wang, T.; Lei, B. Placental super micro-vessels segmentation based on resnext with convolutional block attention and U-Net. In Proceedings of the 2021 43rd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Virtual, 1–5 November 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 4015–4018. [Google Scholar]
  20. Bui, T.H.; Hamamoto, K.; Paing, M.P. Automated Caries Screening Using Ensemble Deep Learning on Panoramic Radiographs. Entropy 2022, 24, 1358. [Google Scholar] [CrossRef]
  21. Siuly, S.; Li, Y.; Wen, P.; Alcin, O.F. SchizoGoogLeNet: The googlenet-based deep feature extraction design for automatic detection of schizophrenia. Comput. Intell. Neurosci. 2022, 2022, 1992596. [Google Scholar] [CrossRef]
  22. Beeche, C.; Singh, J.P.; Leader, J.K.; Gezer, N.S.; Oruwari, A.P.; Dansingani, K.K.; Chhablani, J.; Pu, J. Super U-Net: A modularized generalizable architecture. Pattern Recognit. 2022, 128, 108669. [Google Scholar] [CrossRef]
  23. Agarwal, M.; Gupta, S.K.; Biswas, K. Development of a compressed FCN architecture for semantic segmentation using Particle Swarm Optimization. Neural Comput. Appl. 2023, 35, 11833–11846. [Google Scholar] [CrossRef]
  24. Hu, Z.; Wang, B.; Li, X. Densitytoken: Weakly-Supervised Crowd Counting with Density Classification. In Proceedings of the ICASSP 2023–2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–5. [Google Scholar]
  25. Sivaprakash, P.; Sankar, M.; Chithambaramani, R.; Marichamy, D. A Convolutional Neural Network Approach for Crowd Counting. In Proceedings of the 2023 4th International Conference on Smart Electronics and Communication (ICOSEC), Trichy, India, 20–22 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1515–1520. [Google Scholar]
  26. Chandio, A.A.; Asikuzzaman, M.; Pickering, M.R.; Leghari, M. Cursive text recognition in natural scene images using deep convolutional recurrent neural network. IEEE Access 2022, 10, 10062–10078. [Google Scholar] [CrossRef]
  27. Xue, C.; Huang, J.; Zhang, W.; Lu, S.; Wang, C.; Bai, S. Image-to-Character-to-Word Transformers for Accurate Scene Text Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 12908–12921. [Google Scholar] [CrossRef] [PubMed]
  28. Islam, S.M.; Mondal, H.S. Image enhancement based medical image analysis. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  29. Subramanian, A.K.; Chen, Y.; Almalki, A.; Sivamurthy, G.; Kafle, D. Cephalometric Analysis in Orthodontics Using Artificial Intelligence—A Comprehensive Review. BioMed Res. Int. 2022, 2022, 1880113. [Google Scholar] [CrossRef] [PubMed]
  30. Alshagathrh, F.M.; Househ, M.S. Artificial Intelligence for Detecting and Quantifying Fatty Liver in Ultrasound Images: A Systematic Review. Bioengineering 2022, 9, 748. [Google Scholar] [CrossRef] [PubMed]
  31. Huang, S.; Su, Z.; Liu, S.; Chen, J.; Su, Q.; Su, H.; Shang, Y.; Jiao, Y. Combined assisted bone age assessment and adult height prediction methods in Chinese girls with early puberty: Analysis of three artificial intelligence systems. Pediatr. Radiol. 2023, 53, 1108–1116. [Google Scholar] [CrossRef] [PubMed]
  32. Coşkun, M.F.; Akar, G.B. Crowd Multi Prediction: Single Network for Crowd Counting, Localization and Anomaly Detection. In Proceedings of the 2023 IEEE International Conference on Consumer Electronics (ICCE), Las Vegas, NV, USA, 6–8 January 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–3. [Google Scholar]
  33. Sheng, J.; Zhang, Z. Summary of Pedestrian Traffic Statistics Research. In Proceedings of the 2020 15th IEEE Conference on Industrial Electronics and Applications (ICIEA), Kristiansand, Norway, 9–13 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 43–47. [Google Scholar]
  34. Zhi, X.; Zhao, B.; Wang, Y. A Hybrid Framework for Text Recognition Used in Commodity Futures Document Verification. In Proceedings of the 2021 6th International Conference on Computational Intelligence and Applications (ICCIA), Xiamen, China, 11–13 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 143–148. [Google Scholar]
  35. Kılıc, M.C.; Bayrakdar, I.S.; Çelik, Ö.; Bilgir, E.; Orhan, K.; Aydın, O.B.; Kaplan, F.A.; Sağlam, H.; Odabaş, A.; Aslan, A.F.; et al. Artificial intelligence system for automatic deciduous tooth detection and numbering in panoramic radiographs. Dentomaxillofac. Radiol. 2021, 50, 20200172. [Google Scholar] [CrossRef]
  36. Sheng, C.; Wang, L.; Huang, Z.; Wang, T.; Guo, Y.; Hou, W.; Xu, L.; Wang, J.; Yan, X. Transformer-Based Deep Learning Network for Tooth Segmentation on Panoramic Radiographs. J. Syst. Sci. Complex. 2023, 36, 257–272. [Google Scholar] [CrossRef]
  37. Wang, J.; Wu, Y.; Mei, Y. Insulator instance segmentation based on deep learning network Mask RCNN. In Proceedings of the 2022 China International Conference on Electricity Distribution (CICED), Changsha, China, 7–8 September 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1424–1427. [Google Scholar]
  38. Gunawan, F.; Hwang, C.L.; Cheng, Z.E. ROI-YOLOv8-Based Far-Distance Face-Recognition. In Proceedings of the 2023 International Conference on Advanced Robotics and Intelligent Systems (ARIS), Taipei, Taiwan, 30 August–1 September 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
Figure 1. Examples of data augmentation. (a) Original image; (b) image after horizon flipped augmentation; (c) image after contrast adjustment; (d) image after rotation adjustment.
Figure 1. Examples of data augmentation. (a) Original image; (b) image after horizon flipped augmentation; (c) image after contrast adjustment; (d) image after rotation adjustment.
Diagnostics 14 00497 g001
Figure 2. The pipeline of the proposed STSN-Net.
Figure 2. The pipeline of the proposed STSN-Net.
Diagnostics 14 00497 g002
Figure 3. The pipeline of the GCN.
Figure 3. The pipeline of the GCN.
Diagnostics 14 00497 g003
Figure 4. The architecture of the DSN.
Figure 4. The architecture of the DSN.
Diagnostics 14 00497 g004
Figure 5. The architecture of the RSSN.
Figure 5. The architecture of the RSSN.
Diagnostics 14 00497 g005
Figure 6. The F1 score comparison.
Figure 6. The F1 score comparison.
Diagnostics 14 00497 g006
Figure 7. Comparison of mAP (object detection) results.
Figure 7. Comparison of mAP (object detection) results.
Diagnostics 14 00497 g007
Figure 8. Comparison of mAP (instance segmentation) results.
Figure 8. Comparison of mAP (instance segmentation) results.
Diagnostics 14 00497 g008
Figure 9. Comparison of DICE score.
Figure 9. Comparison of DICE score.
Diagnostics 14 00497 g009
Table 1. Details of the dataset split.
Table 1. Details of the dataset split.
CategoryTrainValidationTestTotal
Permanent Dentition11152282281571
Primary Dentition4228687595
Total15003003162116
Table 2. Distribution of the dataset.
Table 2. Distribution of the dataset.
SubcategoryTrainValidationTestTotal
Missing Teeth4278788602
Dental Crowding8371711721180
Impacted Teeth528108108744
Teeth with Fillings3847879541
Teeth with Restorations3286767462
Table 3. F1 score obtained by the baseline, Mask R-CNN (using Swin Transformer as backbone network), YOLOv8 and the proposed network (contributions of different steps of the proposed network were also demonstrated).
Table 3. F1 score obtained by the baseline, Mask R-CNN (using Swin Transformer as backbone network), YOLOv8 and the proposed network (contributions of different steps of the proposed network were also demonstrated).
Base Size (1333 × 800)F1 (1.0×)F1 (1.2×)F1 (1.4×)F1 (1.6×)
Baseline97.36%97.53%97.64%97.88%
+ Deformable convolution98.09%98.14%98.19%98.23%
+ Dilated convolution98.08%98.13%98.14%98.16%
+ Region segmentation98.11%98.12%98.15%98.22%
Mask R-CNN97.49%97.62%97.82%97.89%
YOLOv896.49%96.78%96.89%97.02%
The proposed network98.30%98.32%98.38%98.49%
Table 4. mAP and DICE scores obtained by the Mask R-CNN (using Swin Transformer as backbone network), YOLOv8, and the proposed network.
Table 4. mAP and DICE scores obtained by the Mask R-CNN (using Swin Transformer as backbone network), YOLOv8, and the proposed network.
MetricsMask R-CNNYOLOv8 STSN-Net
mAP (object detection)96.58%95.61%98.90%
mAP (instance segmentation)95.88%94.60%98.10%
DICE94.98%94.02%96.29%
Table 5. Inference times of the baseline and different steps of the proposed method (the numbers represent the average inference time per image).
Table 5. Inference times of the baseline and different steps of the proposed method (the numbers represent the average inference time per image).
Base Size (1333 × 800)F1 (1.0×)F1 (1.2×)F1 (1.4×)F1 (1.6×)
Baseline858994100
+ Deformable convolution100110115120
+ Dilated convolution9598100110
+ Region segmentation909599104
The proposed network110115118121
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, S.; Liang, S.; Chang, Q.; Zhang, L.; Gong, B.; Bai, Y.; Zuo, F.; Wang, Y.; Xie, X.; Gu, Y. STSN-Net: Simultaneous Tooth Segmentation and Numbering Method in Crowded Environments with Deep Learning. Diagnostics 2024, 14, 497. https://doi.org/10.3390/diagnostics14050497

AMA Style

Wang S, Liang S, Chang Q, Zhang L, Gong B, Bai Y, Zuo F, Wang Y, Xie X, Gu Y. STSN-Net: Simultaneous Tooth Segmentation and Numbering Method in Crowded Environments with Deep Learning. Diagnostics. 2024; 14(5):497. https://doi.org/10.3390/diagnostics14050497

Chicago/Turabian Style

Wang, Shaofeng, Shuang Liang, Qiao Chang, Li Zhang, Beiwen Gong, Yuxing Bai, Feifei Zuo, Yajie Wang, Xianju Xie, and Yu Gu. 2024. "STSN-Net: Simultaneous Tooth Segmentation and Numbering Method in Crowded Environments with Deep Learning" Diagnostics 14, no. 5: 497. https://doi.org/10.3390/diagnostics14050497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop