Next Article in Journal
Back to the Metrics: Exploration of Distance Metrics in Anomaly Detection
Previous Article in Journal
Kinematic IMU-Based Assessment of Postural Transitions: A Preliminary Application in Clinical Context
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Age and Sex Estimation in Children and Young Adults Using Panoramic Radiographs with Convolutional Neural Networks

by
Tuğçe Nur Şahin
1,* and
Türkay Kölüş
2
1
Department of Pedodontics, Ahmet Keleşoğlu Faculty of Dentistry, Karamanoğlu Mehmetbey University, Karaman 70200, Turkey
2
Department of Restorative Dentistry, Ahmet Keleşoğlu Faculty of Dentistry, Karamanoğlu Mehmetbey University, Karaman 70200, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(16), 7014; https://doi.org/10.3390/app14167014
Submission received: 10 July 2024 / Revised: 30 July 2024 / Accepted: 8 August 2024 / Published: 9 August 2024
(This article belongs to the Special Issue Oral Diseases: Diagnosis and Therapy)

Abstract

:
Image processing with artificial intelligence has shown significant promise in various medical imaging applications. The present study aims to evaluate the performance of 16 different convolutional neural networks (CNNs) in predicting age and gender from panoramic radiographs in children and young adults. The networks tested included DarkNet-19, DarkNet-53, Inception-ResNet-v2, VGG-19, DenseNet-201, ResNet-50, GoogLeNet, VGG-16, SqueezeNet, ResNet-101, ResNet-18, ShuffleNet, MobileNet-v2, NasNet-Mobile, AlexNet, and Xception. These networks were trained on a dataset of 7336 radiographs from individuals aged between 5 and 21. Age and gender estimation accuracy and mean absolute age prediction errors were evaluated on 340 radiographs. Statistical analyses were conducted using Shapiro–Wilk, one-way ANOVA, and Tukey tests (p < 0.05). The gender prediction accuracy and the mean absolute age prediction error were, respectively, 87.94% and 0.582 for DarkNet-53, 86.18% and 0.427 for DarkNet-19, 84.71% and 0.703 for GoogLeNet, 81.76% and 0.756 for DenseNet-201, 81.76% and 1.115 for ResNet-18, 80.88% and 0.650 for VGG-19, 79.41% and 0.988 for SqueezeNet, 79.12% and 0.682 for Inception-Resnet-v2, 78.24% and 0.747 for ResNet-50, 77.35% and 1.047 for VGG-16, 76.47% and 1.109 for Xception, 75.88% and 0.977 for ResNet-101, 73.24% and 0.894 for ShuffleNet, 72.35% and 1.206 for AlexNet, 71.18% and 1.094 for NasNet-Mobile, and 62.94% and 1.327 for MobileNet-v2. No statistical difference in age prediction performance was found between DarkNet-19 and DarkNet-53, which demonstrated the most successful age estimation results. Despite these promising results, all tested CNNs performed below 90% accuracy and were not deemed suitable for clinical use. Future studies should continue with more-advanced networks and larger datasets.

1. Introduction

When estimating age, criteria such as weight, height, signs of puberty, mental development, and teeth and bone development are considered [1]. Estimated age is essential in some dental treatment approaches. It can also be used in anthropology to determine the age of death of hominid fossils. In forensic cases, it can be used when chronological age information is unavailable or the existing information is questionable [2].
Hard tissues in the body can be examined to determine age. For bone-based estimation, the hand–wrist area, pubic symphysis, and medial clavicular epiphysis cartilage can be examined, but teeth-based methods are generally less susceptible to error [3]. Teeth are the most durable part of the human body and can provide valuable information about a person’s ethnic origin, occupation, sex, and age [4]. Cameriere’s, Demirjian’s, and Willems’s methods and other methods, such as measuring pulp volume or root transparency, have been developed to estimate age from teeth [5,6]. Each method has different levels of accuracy and is generally limited to specific populations [7]. Performing age estimation based on manual measurement methods such as Cameriere’s open apices method or using dental development tables like Demirjian’s method can increase the likelihood of errors and result in subjective outcomes that may vary between operators. Moreover, these conventional methods are time-consuming and require the teeth used for age estimation to be free of cavities or to have not undergone root canal treatment [8]. Indeed, it is essential to note that sex can influence dental development patterns. It has been demonstrated that girls reach certain stages of dental development earlier than boys, particularly in permanent teeth [9]. Therefore, age estimation methods such as those of Demirjian and Willems have separate dental development tables for girls and boys to account for these sex-related differences [10].
CBCT and cephalometric and panoramic radiography can be used to estimate radiographic sex based on teeth and jawbones [11,12,13]. Although panoramic radiographs have disadvantages, such as ghost images, geometric distortion, and superimpositions, they are frequently preferred for diagnosis and follow-up because they are cheaper and more accessible, can be taken quickly, and have a lower radiation dose [14].
Parallel to technological advancements and the integration of deep learning (DL) into various industries, traditional techniques are being replaced by artificial intelligence (AI) methods. DL is a subset of machine learning that uses networks with multiple layers of computation to analyze input data, thereby significantly enhancing the performance of AI [15]. Convolutional neural networks (CNNs) are a type of DL algorithm particularly well suited for image processing [16]. AlexNet is a classic CNN architecture with an eight-layer depth that won the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) in 2012 by significantly surpassing previous methods. It became popular and inspired the development of other similar methods. VGG-16 is another popular method with a straightforward design, but it has many parameters that take a long time to learn. The numbers at the end of the network names usually indicate the number of layers, in other words, their depth. VGG-19 is very similar to VGG-16 but contains more layers. Google implemented inception building blocks into CNNs and called the network GoogLeNet (Inceptionv1). These blocks work well together, resulting in a model that is easy to generalize. This model is regularly updated with new versions, including Xception, which uses depthwise separable convolutions. ResNet, short for Residual Network, is a specific type of neural network. It comes in various sizes and numbers of layers, such as ResNet-18, ResNet-50, and ResNet-101. DenseNet is very similar to ResNet but concatenates the previous layer’s output with the next layer’s output. DarkNet-19 and DarkNet-53 are the networks used in YOLO, a popular object recognition method. CNNs generally require very high processing power. As alternatives to other models, SqueezeNet, ShuffleNet, and NasNet-Mobile were developed for use in less-powerful systems [17].
The study hypothesizes that at least one of the tested CNNs will have an accuracy greater than 90%, which is considered the success threshold. Our aim in this exploratory study is to compare the performance of 16 popular CNNs in predicting age and gender from panoramic radiographs in children and young adults.

2. Materials and Methods

This study was conducted according to the Declaration of Helsinki and approved by the Karaman Provincial Health Directorate and the Karamanoğlu Mehmetbey University Faculty of Medicine Clinical Research Ethics Committee with decision number 07-2022/19. Patient consent was obtained before the radiographs were taken and the consent forms have been archived at institutions.
In machine learning processes, the prediction accuracy of trained models is closely related to the sample size, with accuracy tending to improve as the sample size increases. However, after reaching a certain point, often referred to as the “sweet spot”, further increases in sample size do not significantly enhance the model’s performance. Identifying this point is a state-of-the-art research topic in biostatistics, and unfortunately, there is currently no generally accepted rule. Since having a large sample size does not harm the artificial intelligence training process aside from extending the required training time, the approach of using “the largest possible sample size” was preferred in this research. The dataset consisted of 7667 panoramic radiographs obtained from patients aged 5–21, who did not have any systemic diseases; these radiographs were previously taken for diagnostic or treatment purposes at Karamanoğlu Mehmetbey University’s Ahmet Keleşoğlu Faculty of Dentistry and Karaman Oral and Dental Health Center between 2016 and 2023. Out of a total of 8000 panoramic radiographs obtained using PCH-2500 (60–75 kVp, 5–10 mA, Vatech, Gyeonggi-do, Republic of Korea), Veraview IC5 HD (60–70 kV, 1–7.5 mA, Morita Corporation, Osaka, Japan), Hyperion X9 (60–75 kVp, 5–10 mA, My Ray, Imola, Italy), and Hyperion X5 (60–75 kVp, 5–10 mA, My Ray, Imola, Italy) panoramic radiography devices, 333 radiographs with imaging errors such as significant distortion, imaging errors due to patient positioning, or large ghost images were excluded from the study. To reflect reality more accurately, radiographs from patients with trans-palatal arches, braces, root canal treatments, caries, implants, bridges, and all types of restorations, hypodontia, or oligodontia were not excluded from the study. If present, metadata such as the name, date of birth, and radiograph capture date were cropped from specific radiographs.
When building a DL system, a significant focus must be on avoiding overfitting. The learning capacity must be appropriate, considering the amount of data and the difficulty of the task. Training, validation, and testing must be used to correctly verify which accuracy a DL system will achieve for future unseen data. This means that the acquired dataset must be split into three parts. The training set (1) is used to learn which patterns in the images correlate with the labels. For the validation set (2), once these patterns are learned, a different set of images is used to verify whether the model is under- or overfitting. Based on this assessment, the learning capacity of the network can be adapted to better fit the data. Regarding test set (3), since design choices are made based on the validation set’s performance, we need a third set to avoid selection bias and to report an unbiased estimate of the model performance. No perfect ratio can be determined a priori, but it should be determined based on the amount of available data. The basic rule is that the validation and test sets should be large enough to represent the whole range of variability found in real life. The remainder of the data can then be used for training. For a small amount of data, the ratio used for validation and testing will be relatively high compared to that used for training. However, as the amount of data grows, this ratio will become smaller [18]. Considering that we had a relatively medium-sized dataset, this study allocated 70% of our dataset for training, 25% for validation, and 5% (10 radiographs for each class) for testing. The numerical distribution of the radiographs according to age/sex and purpose of use is provided in Table 1.
The chronological age predictions were conducted using CNNs. In this study, several popular CNN architectures were employed: DarkNet-19, DarkNet-53, Inception-ResNet-v2, VGG-19, DenseNet-201, ResNet-50, GoogLeNet, VGG-16, SqueezeNet, ResNet-101, ResNet-18, ShuffleNet, MobileNet-v2, NasNet-Mobile, AlexNet, and Xception. Different radiography devices produce images at varying resolutions, necessitating the standardization of radiographs for training each CNN model. Therefore, all radiographs were automatically resized by adhering to the input resolution specifications outlined in Table 2.
The training parameters, such as batch size, learning rate, and optimization method, significantly impact the accuracy of CNN networks. The optimal combination of these parameters can vary depending on the dataset’s structure. Therefore, experiments were conducted using three randomly selected networks (SqueezeNet, GoogLeNet, and DarkNet-19) with 36 different parameter combinations to identify the optimal training parameters. These combinations included batch sizes of 8, 16, and 32; learning rates of 0.01, 0.001, and 0.0001; and optimization methods Adam, SGDM, and RMSProp. Each network was trained for 50 epochs.
It was observed that as the number of epochs increased during AI training, the validation loss also increased, indicating overfitting of the model. To prevent overfitting, the number of epochs was set to 10. For learning curve optimization, a learning rate of 0.001 was used for the first five epochs, followed by a learning rate of 0.0001 for the subsequent five epochs. The batch size was set to 16, and the optimization method chosen was SGDM. After determining these parameters, all networks were trained using these specified values on MatLab (R2023a) software.
The performance of the models in age determination was evaluated on 10 radiographs from each age–sex combination, totaling 340 radiographs. The test set comprised approximately 5% of our entire radiograph dataset. The predictive performance of the trained models was evaluated based on age, sex, and a combination of both. The accuracy percentage was calculated by dividing the correctly predicted radiographs by the total number of radiographs in the specified test group. We determined each network model’s mean absolute error (MAE) by calculating the absolute difference between predicted and chronological age for a more detailed age prediction performance evaluation.

Statistics

The MAE data were analyzed using Shapiro–Wilk’s normality test and Levene’s homogeneity test with the IBM SPSS Statistics (v. 21, IBM, Chicago, IL, USA) software package. For the normally and homogeneously distributed dataset, one-way analysis of variance (ANOVA) and post hoc Tukey tests were performed. A p-value of less than 0.05 was considered statistically significant.

3. Results

According to Table 2, DarkNet-19 achieved the highest accuracy in age prediction alone, with an accuracy of 80.59% and a mean absolute error (MAE) of 0.4265. On the other hand, DarkNet-53 achieved the highest accuracy in sex prediction alone, with an accuracy of 87.94%. The detailed prediction matrix for DarkNet-53 is shown in Figure 1.
Table S1 (Supplementary Materials) provides a detailed statistical comparison of the MAE of all networks in the study. Accordingly, the differences between DarkNet-19 and DarkNet-53, the networks that provide the closest results to chronological age, and AlexNet and Xception, which give the most inaccurate results, are not statistically significant. In contrast, the difference between DarkNet-19 and Xception is significant.
Additionally, when age and sex predictions were combined, the accuracy was lower than when they were studied separately. DarkNet-53 achieved the highest accuracy in combined prediction, 70.88%.
When examining the prediction success distribution graphs in Figure 1 and Figure 2, it can be observed that CNNs mostly made predictions close to the actual sex and age but sometimes made significantly incorrect predictions. For example, DarkNet-53 incorrectly predicted that one in ten radiographs of a 19-year-old female was that of a 5-year-old male (Figure 1). There were also cases where certain age groups were collectively mispredicted, such as GoogLeNet failing to make any correct predictions for the 17-year-old female group (Figure 2).

4. Discussion

AI is increasingly used in complex decision-making, problem-solving, and object recognition tasks. Studies evaluating the use of AI in age estimation have shown that AI can be more successful than manual techniques [19]. Thus, this study focused on AI instead of manual methods and aimed to evaluate the performance of 16 different convolutional neural networks (CNNs) in predicting age and gender from panoramic radiographs in children and young adults. The networks tested included DarkNet-19, DarkNet-53, Inception-ResNet-v2, VGG-19, DenseNet-201, ResNet-50, GoogLeNet, VGG-16, SqueezeNet, ResNet-101, ResNet-18, ShuffleNet, MobileNet-v2, NasNet-Mobile, AlexNet, and Xception. The hypothesis of the study that at least one artificial intelligence network would achieve an accuracy greater than 90% in predicting age and gender was rejected. The highest accuracy rate observed for age prediction was 80.59%, while the highest accuracy rate for gender prediction was 87.94%.
The use of AI in forensic medicine and dentistry has been explored using various CNNs, dental X-rays, MRIs, and photographs [20,21,22,23,24,25,26]. Panoramic radiographs are commonly used in routine diagnosis, treatment planning, and follow-up of treatments in the maxillofacial region. They are easily accessible imaging methods and are frequently preferred in age estimation studies using AI [20,23,24,27,28,29]. For these reasons, panoramic radiographs obtained from patients in the past were also used for this study.
An additional advantage of age estimation using AI, in addition to the shorter completion time compared to traditional methods [30], is that it can be used for age estimation even in cases where conventional methods are not applicable, such as when dealing with individuals who have dental caries, missing teeth, restorations, or root canal treatments. AI-based age estimation has no limitations in these cases [20]. As a result, the radiographs of individuals who had undergone dental treatments were not excluded from our study, and a considerable number of panoramic radiographs could be used in our analysis. This enabled a more comprehensive and inclusive approach to age estimation using AI, encompassing a diverse range of individuals, including those who had received dental treatments. In addition to the chosen network, the most crucial factors for the success of artificial intelligence models are the data sample size, data heterogeneity, and training epochs [31]. Considering that the replacement of primary teeth with permanent teeth and the development of jawbones occur dramatically between the ages of 5 and 21, patients within this age range were selected for the study to provide precise data for the AI. Furthermore, some laws differ depending on age, with ages 14, 18, and 21 generally considered to be legal thresholds [32].
Ataş et al. [24] used 1332 panoramic radiographs obtained from patients who were aged from 8 to 68 for their study, and they reported the mean absolute error (MAE) values as follows: 3.47 for EfficientNetB4, 4.35 for ResNet50V2, 4.22 for DenseNet-201, 3.44 for InceptionV3, 3.68 for MobileNetV2, and 10.46 for VGG16. They also modified the InceptionV3 architecture and achieved an MAE of 3.13. We obtained lower MAE values in our study group aged 5 to 21. The MAE values were 0.75 for DenseNet201, 1.32 for MobileNetV2, and 1.04 for VGG16. It is well known that input size has a significant impact on the success of machine learning. In our study, the use of a larger input size, along with our focus on a younger age range, may explain the observed difference.
According to the study conducted by Mualla et al. [23] using 1429 panoramic radiographs and employing AlexNet and ResNet-101, they reported an accuracy of 95.80% for AlexNet and 92.60% for ResNet. In our study, however, we obtained lower accuracies of 60.29% for ResNet-101 and 47.35% for AlexNet. It is important to note that, unlike in their research, we did not apply any additional pre-processing to the images, which may have contributed to the differences in the results.
In the study conducted by Kim et al. [27], they used cropped images of teeth from panoramic radiographs, specifically teeth numbers 16, 26, 36, and 46, and employed the ResNet-152 network for dental age estimation. They reported an accuracy range between 89.05% and 90.27% in their findings. Similar studies can be found in the literature where cropped panoramic images were used [28]. Using a deeper network and providing a smaller region of interest instead of a broad panoramic view for AI training may have contributed to higher performance.
Ko et al. [13] used 15,000 panoramic radiographs and Darknet-19, and they reported an accuracy of approximately 84% and 96% when the acceptable range of dental estimation was ±5 and ±10 years, respectively. They also reported a mean error between actual age and predicted age of 2.74 and 3.49, respectively. In our study, using DarkNet-19, which showed the highest accuracy in age estimation, we achieved an accuracy of 80.59% and an MAE of 0.42 ± 1.18641. We believe that the more significant number of panoramic radiographs used in Ko et al.’s study may have contributed to a higher accuracy than in ours.
Although there is no definitive consensus in the literature, it is generally reported that at least one thousand examples per class are necessary for AI image classification. The most significant limitation of this study is that it was conducted with a dataset that cannot be considered large due to the inherent constraints of the healthcare field. Another limitation worth mentioning is that the images needed to be pre-processed. Similar studies that included pre-processing have observed higher success rates. Future studies could utilize larger datasets with pre-processed images. Additionally, instead of the conventional CNNs we tested, networks specifically designed for medical imaging could be developed.

5. Conclusions

Although artificial intelligence offers promising results in estimating age and gender from dental panoramic radiographs, it has been concluded that none of the examined networks are suitable for clinical use without any pre-processing on panoramic radiographs or any additional algorithms on networks, since their accuracy remains below 90%. Artificial intelligence technology is developing very rapidly, and in the future, studies should be continued with more-advanced networks using a larger dataset.

Supplementary Materials

The following supporting information can be downloaded at this location: https://www.mdpi.com/article/10.3390/app14167014/s1, Table S1. Post hoc Tukey test results for the CNNs’ Mean Absolute Error (MAE) data.

Author Contributions

Conceptualization, T.N.Ş.; Methodology, T.K.; Software, T.K.; Formal analysis, T.K.; Investigation, T.N.Ş.; Data curation, T.N.Ş.; Writing—original draft, T.N.Ş.; Writing—review & editing, T.K.; Visualization, T.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The ethics approval was obtained from the Karamanoğlu Mehmetbey University Faculty of Medicine Clinical Research Ethics Committee on 26 July 2022 (Approval code: 07-2022/19). The study was conducted following the Declaration of Helsinki and approved by the Karaman Provincial Health Directorate.

Informed Consent Statement

Written informed consent has been obtained from the patients/legal guardians.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to Personal Data Protection Law in our country.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, J.H.; Kim, Y.J.; Kim, K.G. Bone age estimation using deep learning and hand X-ray images. Biomed. Eng. Lett. 2020, 10, 323–331. [Google Scholar] [CrossRef] [PubMed]
  2. Manjunatha, B.; Soni, N.K. Estimation of age from development and eruption of teeth. Forensic Dent. Sci. 2014, 6, 73–76. [Google Scholar] [CrossRef] [PubMed]
  3. Marconi, V.; Iommi, M.; Monachesi, C.; Faragalli, A.; Skrami, E.; Gesuita, R.; Ferrante, L.; Carle, F. Validity of age estimation methods and reproducibility of bone/dental maturity indices for chronological age estimation: A systematic review and meta-analysis of validation studies. Sci. Rep. 2022, 12, 15607. [Google Scholar] [CrossRef] [PubMed]
  4. Krishan, K.; Kanchan, T.; Garg, A.K. Dental Evidence in Forensic Identification—An Overview, Methodology and Present Status. Open Dent. J. 2015, 9, 250–256. [Google Scholar] [CrossRef] [PubMed]
  5. Yousefi, F.; Mohammadi, Y.; Ahmadvand, M.; Razaghi, P. Dental age estimation using cone-beam computed tomography: A systematic review and meta-analysis. Imaging Sci. Dent. 2023, 53, 91–99. [Google Scholar] [CrossRef] [PubMed]
  6. Singhal, A.; Ramesh, V.; Balamurali, P. A comparative analysis of root dentin transparency with known age. J. Forensic Dent. Sci. 2010, 2, 18–21. [Google Scholar] [CrossRef] [PubMed]
  7. Kurniawan, A.; Chusida, A.n.; Atika, N.; Gianosa, T.K.; Solikhin, M.D.; Margaretha, M.S.; Utomo, H.; Marini, M.I.; Rizky, B.N.; Prakoeswa, B.F.W.R. The applicable dental age estimation methods for children and adolescents in Indonesia. Int. J. Dent. 2022, 2022, 6761476. [Google Scholar] [CrossRef] [PubMed]
  8. Cameriere, R.; Ferrante, L.; Cingolani, M. Age estimation in children by measurement of open apices in teeth. Int. J. Leg. Med. 2006, 120, 49–52. [Google Scholar] [CrossRef] [PubMed]
  9. Liversidge, H. Variation in modern human dental development. In Patterns of Growth and Development in the Genus Homo, 1st ed.; Thompson, J.L., Krovitz, G.E., Nelson, A.J., Eds.; Cambridge University Press: Cambridge, UK, 2003; pp. 73–113. [Google Scholar]
  10. Apaydin, B.; Yaşar, F. Accuracy of the Demirjian, Willems and Cameriere methods of estimating dental age on Turkish children. Niger. J. Clin. Pract. 2018, 21, 257–263. [Google Scholar] [CrossRef]
  11. Khazaei, M.; Mollabashi, V.; Khotanlou, H.; Farhadian, M. Sex determination from lateral cephalometric radiographs using an automated deep learning convolutional neural network. Imaging Sci. Dent. 2022, 52, 239–244. [Google Scholar] [CrossRef]
  12. Saric, R.; Kevric, J.; Hadziabdic, N.; Osmanovic, A.; Kadic, M.; Saracevic, M.; Jokic, D.; Rajs, V. Dental age assessment based on CBCT images using machine learning algorithms. Forensic Sci. Int. 2022, 334, 111245. [Google Scholar] [CrossRef] [PubMed]
  13. Ko, J.; Kim, Y.-R.; Hwang, H.-J.; Chang, W.-D.; Han, M.-S.; Nam, S.-H.; Jung, Y.-J. Dental Panoramic Radiography in Age Estimation for Dental Care using Dark-Net 19. J. Magn. 2022, 27, 485–491. [Google Scholar] [CrossRef]
  14. Guo, Y.; Wang, D.; Wang, Y.; Peng, X.; Guo, C. Imaging features of medicine-related osteonecrosis of the jaws: Comparison between panoramic radiography and computed tomography. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2016, 122, 69–76. [Google Scholar] [CrossRef] [PubMed]
  15. Silva, G.; Oliveira, L.; Pithon, M. Automatic segmenting teeth in X-ray images: Trends, a novel data set, benchmarking and future perspectives. Expert Syst. Appl. 2018, 107, 15–31. [Google Scholar] [CrossRef]
  16. Khanagar, S.B.; Al-Ehaideb, A.; Maganur, P.C.; Vishwanathaiah, S.; Patil, S.; Baeshen, H.A.; Sarode, S.C.; Bhandi, S. Developments, application, and performance of artificial intelligence in dentistry–A systematic review. J. Dent. Sci. 2021, 16, 508–522. [Google Scholar] [CrossRef] [PubMed]
  17. Alzahrani, S.; Al-Bander, B.; Al-Nuaimy, W. A Comprehensive Evaluation and Benchmarking of Convolutional Neural Networks for Melanoma Diagnosis. Cancers 2021, 13, 4494. [Google Scholar] [CrossRef] [PubMed]
  18. Eelbode, T.; Sinonquel, P.; Maes, F.; Bisschops, R. Pitfalls in training and validation of deep learning systems. Best Pract. Res. Clin. Gastroenterol. 2021, 52, 101712. [Google Scholar] [CrossRef] [PubMed]
  19. Guo, Y.-C.; Han, M.; Chi, Y.; Long, H.; Zhang, D.; Yang, J.; Yang, Y.; Chen, T.; Du, S. Accurate age classification using manual method and deep convolutional neural network based on orthopantomogram images. Int. J. Leg. Med. 2021, 135, 1589–1597. [Google Scholar] [CrossRef]
  20. Vila-Blanco, N.; Carreira, M.J.; Varas-Quintana, P.; Balsa-Castro, C.; Tomas, I. Deep neural networks for chronological age estimation from OPG images. IEEE Trans. Med. Imaging 2020, 39, 2374–2384. [Google Scholar] [CrossRef]
  21. Rajee, M.; Mythili, C. Gender classification on digital dental X-ray images using deep convolutional neural network. Biomed. Signal Process Control 2021, 69, 102939. [Google Scholar] [CrossRef]
  22. Ataş, I. Human gender prediction based on deep transfer learning from panoramic dental radiograph images. Trait. Du. Signal 2022, 39, 1585. [Google Scholar] [CrossRef]
  23. Mualla, N.; Houssein, E.H.; Hassan, M. Dental Age Estimation Based on X-ray Images. Comput. Mater. Contin. 2020, 62, 591–605. [Google Scholar] [CrossRef]
  24. Ataş, İ.; Özdemir, C.; Ataş, M.; Doğan, Y. Forensic dental age estimation using modified deep learning neural network. arXiv 2022, arXiv:2208.09799. [Google Scholar] [CrossRef]
  25. Štern, D.; Payer, C.; Urschler, M. Automated age estimation from MRI volumes of the hand. Med. Image Anal. 2019, 58, 101538. [Google Scholar] [CrossRef] [PubMed]
  26. Shanthi, N.; Yuvasri, P.; Vaishnavi, S.; Vidhya, P. Gender and Age Detection using Deep Convolutional Neural Networks. In Proceedings of the 4th International Conference on Smart Systems and Inventive Technology (ICSSIT), Tirunelveli, India, 20–22 January 2022; pp. 951–956. [Google Scholar]
  27. Kim, S.; Lee, Y.-H.; Noh, Y.-K.; Park, F.C.; Auh, Q.-S. Age-group determination of living individuals using first molar images based on artificial intelligence. Sci. Rep. 2021, 11, 1073. [Google Scholar] [CrossRef] [PubMed]
  28. Avuçlu, E.; Başçiftçi, F. The determination of age and gender by implementing new image processing methods and measurements to dental X-ray images. Med. Image Anal. 2020, 149, 106985. [Google Scholar] [CrossRef]
  29. Capitaneanu, C.; Willems, G.; Jacobs, R.; Fieuws, S.; Thevissen, P. Sex estimation based on tooth measurements using panoramic radiographs. Int. J. Leg. Med. 2017, 131, 813–821. [Google Scholar] [CrossRef] [PubMed]
  30. Banjšak, L.; Milošević, D.; Subašić, M. Implementation of artificial intelligence in chronological age estimation from orthopantomographic X-ray images of archaeological skull remains. Bull. Int. Assoc. Paleodont. 2020, 14, 122–129. [Google Scholar]
  31. Ali, L.; Alnajjar, F.; Jassmi, H.A.; Gocho, M.; Khan, W.; Serhani, M.A. Performance evaluation of deep CNN-based crack detection and localization techniques for concrete structures. Sensors 2021, 21, 1688. [Google Scholar] [CrossRef]
  32. Schmeling, A.; Olze, A.; Reisinger, W.; Geserick, G. Forensic age diagnostics of living people undergoing criminal proceedings. Forensic Sci. Int. 2004, 144, 243–245. [Google Scholar] [CrossRef]
Figure 1. Detailed prediction distribution of test radiographs by age and sex groups for the DarkNet-53 network. Correct predictions are shown in shades of blue, while incorrect predictions are displayed in shades of red. For example, all 10 radiographs of 5-year-old girls (05F) were correctly predicted by DarkNet-53. However, among the ten radiographs of 6-year-old females (06F), only one was correctly predicted, while one was predicted as a 5-year-old male (05M), five as 6-year-old males (06M), one as a 7-year-old female (07F), and two as 8-year-old females (08F).
Figure 1. Detailed prediction distribution of test radiographs by age and sex groups for the DarkNet-53 network. Correct predictions are shown in shades of blue, while incorrect predictions are displayed in shades of red. For example, all 10 radiographs of 5-year-old girls (05F) were correctly predicted by DarkNet-53. However, among the ten radiographs of 6-year-old females (06F), only one was correctly predicted, while one was predicted as a 5-year-old male (05M), five as 6-year-old males (06M), one as a 7-year-old female (07F), and two as 8-year-old females (08F).
Applsci 14 07014 g001
Figure 2. Distribution of predictions for the GoogLeNet network. Correct predictions are shown in shades of blue, while incorrect predictions are displayed in shades of red.
Figure 2. Distribution of predictions for the GoogLeNet network. Correct predictions are shown in shades of blue, while incorrect predictions are displayed in shades of red.
Applsci 14 07014 g002
Table 1. The numerical distribution of the radiographs according to age/sex and purpose of use.
Table 1. The numerical distribution of the radiographs according to age/sex and purpose of use.
AgeSexTrainingValidationTestTotal
5Male1787510263
Female1646910243
6Male1657010245
Female1727310255
7Male28512210417
Female27711810405
8Male1767510261
Female26011110381
9Male23710110348
Female1908110281
10Male1576710234
Female1667110247
11Male1476210219
Female1255310188
12Male1104710167
Female1195010179
13Male964110147
Female1034410157
14Male903810138
Female1064510161
15Male753210117
Female1154910174
16Male773310120
Female1295510194
17Male1084610164
Female1787610264
18Male1084610164
Female1807610266
19Male1014310154
Female1596810237
20Male1134810171
Female1807710267
21Male1164910175
Female1857810273
Total514721893407676
Table 2. Age, sex, and combined prediction accuracy with mean absolute error (MAE) and its descriptives.
Table 2. Age, sex, and combined prediction accuracy with mean absolute error (MAE) and its descriptives.
Input ResolutionAge Prediction AccuracySex Prediction AccuracyAge and Sex Prediction AccuracynMean Absolute Age ErrorStd. Dev.Std. Err.95% Confidence IntervalMin.Max.
LowerUpper
DarkNet-19256 × 25680.59%86.18%70.59%3400.42651.186410.064340.29990.5530010
DarkNet-53256 × 25677.35%87.94%70.88%3400.58241.557830.084490.41620.7485014
Inception-Resnet-v2299 × 29968.53%79.12%59.41%3400.68241.491070.080860.52330.8414012
VGG-19224 × 22467.94%80.88%59.71%3400.65001.184150.064220.52370.776308
DenseNet-201224 × 22467.06%81.76%59.71%3400.75591.304340.070740.61670.895007
ResNet-50224 × 22466.76%78.24%58.82%3400.74711.397690.075800.59800.896207
GoogLeNet224 × 22464.12%84.71%58.24%3400.70291.150860.062410.58020.825705
VGG-16224 × 22463.82%77.35%55.00%3401.04711.922730.104270.84201.252208
SqueezeNet227 × 22762.65%79.41%54.41%3400.98821.763240.095620.80011.1763010
ResNet-101224 × 22460.29%75.88%51.18%3400.97651.717350.093140.79331.159709
ResNet-18224 × 22458.53%81.76%50.59%3401.11471.866110.101200.91561.3138015
ShuffleNet224 × 22457.65%73.24%47.94%3400.89411.389160.075340.74591.042307
MobileNet-v2224 × 22451.47%62.94%42.06%3401.32652.052960.111341.10751.5455012
NasNet-Mobile224 × 22449.12%71.18%40.29%3401.09411.669600.090550.91601.2722012
AlexNet227 × 22747.35%72.35%37.06%3401.20591.666600.090381.02811.383709
Xception299 × 29944.12%76.47%35.29%3401.10881.540960.083570.94441.273209
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Şahin, T.N.; Kölüş, T. Age and Sex Estimation in Children and Young Adults Using Panoramic Radiographs with Convolutional Neural Networks. Appl. Sci. 2024, 14, 7014. https://doi.org/10.3390/app14167014

AMA Style

Şahin TN, Kölüş T. Age and Sex Estimation in Children and Young Adults Using Panoramic Radiographs with Convolutional Neural Networks. Applied Sciences. 2024; 14(16):7014. https://doi.org/10.3390/app14167014

Chicago/Turabian Style

Şahin, Tuğçe Nur, and Türkay Kölüş. 2024. "Age and Sex Estimation in Children and Young Adults Using Panoramic Radiographs with Convolutional Neural Networks" Applied Sciences 14, no. 16: 7014. https://doi.org/10.3390/app14167014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop