Next Article in Journal
Compensation Strategies for Bioelectric Signal Changes in Chronic Selective Nerve Cuff Recordings: A Simulation Study
Next Article in Special Issue
CAFD: Context-Aware Fault Diagnostic Scheme towards Sensor Faults Utilizing Machine Learning
Previous Article in Journal
Enhanced Routing Algorithm Based on Reinforcement Machine Learning—A Case of VoIP Service
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images

1
Department of Orthodontics, Graduate School, Kyung Hee University, Seoul 02447, Korea
2
Department of Orthodontics, Peking University School of Stomatology, Beijing 100081, China
3
Department of Oral and Maxillofacial Radiology, Graduate School, Kyung Hee University, Seoul 02447, Korea
4
Division of Orthodontics, Department of Orofacial Science, University of California San Francisco, San Francisco, CA 94143, USA
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(2), 505; https://doi.org/10.3390/s21020505
Submission received: 12 December 2020 / Revised: 7 January 2021 / Accepted: 8 January 2021 / Published: 12 January 2021

Abstract

:
This study was designed to develop and verify a fully automated cephalometry landmark identification system, based on multi-stage convolutional neural networks (CNNs) architecture, using a combination dataset. In this research, we trained and tested multi-stage CNNs with 430 lateral and 430 MIP lateral cephalograms synthesized by cone-beam computed tomography (CBCT) to make a combination dataset. Fifteen landmarks were manually and respectively identified by experienced examiner, at the preprocessing phase. The intra-examiner reliability was high (ICC = 0.99) in manual identification. The results of prediction of the system for average mean radial error (MRE) and standard deviation (SD) were 1.03 mm and 1.29 mm, respectively. In conclusion, different types of image data might be the one of factors that affect the prediction accuracy of a fully-automated landmark identification system, based on multi-stage CNNs.

1. Introduction

As a prerequisite for diagnosis in orthodontic treatment, cephalometric analysis is examined with a goal to achieve a much higher accuracy [1]. The conventional cephalogram is most used in orthodontics, however, it only provides plane information from a three-dimensional (3D) craniofacial structure [2]. Emergence of the cone-beam computed tomography (CBCT) provides high-quality diagnostic images to develop an appropriate treatment plan and facilitate successful orthodontic and orthognathic treatment results [3,4]. Typical software programs enable the clinician to synthesize two-dimensional (2D) digital radiographs in multi-angles with various algorithms, from the CBCT image [3,4]. Several images can be synthesized from one CBCT scan and used for 2D or 3D cephalometric analysis [5,6,7].
A 2D radiograph synthesized from the CBCT has the advantages of maintaining the object size without magnification, and the ability to adjust the head position to reduce distortion (reorientation) [8,9,10,11]. Maximum intensity projection (MIP) is one of the image modalities available in software used to synthesize 2D from CBCT. The visualization effect made with maximum intensity issues voxels parallel on the projection plane. The CBCT image is displayed in a matrix of isotropic voxels or volume elements. The concept of the pixel is 2D cross-sections of voxels that represent image density, or the absorption features of a specific structure in the CBCT image. The highest value of the pixel column corresponds to the projection displaying the anatomic structure with clearer contours. Therefore, the 3D craniofacial image orthographically projected onto the plane to synthesize the 2D MIP lateral cephalogram, has the advantage of a clearer facial tissue profile, an important analytic feature of orthodontic cephalometry [6,12].
Several studies propose systems of fully automated cephalometry landmark identification, based on a machine learning technique [13,14,15,16,17]. Deep learning is subset of the machine learning concept that gives outstanding abilities to recognize features of a complicated image [18,19,20,21]. Several factors influence the accuracy of landmark identification with deep learning—the type of deep learning architecture, the number of datasets, the image quality, or the number of landmarks and identification pattern. Some studies exist used conventional lateral cephalograms and a deep learning architecture based on a fully automated cephalometry landmark identification system. They report a precision range of 2 mm [22,23,24,25].
Accordingly, lateral cephalograms synthesized by CBCT are able to minimize intervention in the layered structure images and enable us to easily identify landmarks. It also has less distortion images than 2D conventional lateral cephalograms. As mentioned above, MIP is able to intensively and clearly address the skeletal structure. Therefore, we assumed that the accuracy of automatic landmark identification would increase if both kinds of lateral cephalograms are used to take advantages of.
This study was designed to make a combination dataset that consisted of 2D lateral and MIP lateral cephalograms synthesized by CBCT, to perform efficient landmark identification. It was made for developing and verifying a fully automated cephalometry landmark identification system, based on the multi-stage convolutional neural networks (CNNs) architecture.

2. Materials and Methods

This retrospective study was performed under approval from the Institutional Review Board of Kyung Hee University Dental Hospital (IRB Number: IRB-KH DT19013). Informed consent was waived due to the retrospective nature of this study. All experiments were carried out in accordance with the approved guidelines.

2.1. Subjects

Subjects were randomly selected from the PACS (picture archiving and communication systems) database at Kyung Hee University Dental Hospital. Subjects that interfered with landmarks identification with missing upper and lower permanent incisors, missing permanent upper and lower first molars, craniofacial syndromes, or dento-facial traumas, were excluded. All age, gender, and skeletal discrepancy were included. A total of 430 CBCT scans were selected for CBCT-LC and MIP-LC.

2.2. Image Acquisition from CBCT

2.2.1. CBCT Protocol

The CBCT scans were taken with a 0.39-mm3 voxel size level, 16 × 13 cm field of view, 10 mA, 80 kV, and 30 s scan time (Alphad vega, Asahi Roentgen Inc., Kyoto, Japan). The obtained data were imported as DICOM (Digital Imaging and Communications in Medicine) files to the Dolphin software 11.95 Premium (Dolphin Imaging & Management Solutions, Chatsworth, CA, USA).

2.2.2. Reorientation

The 430 CBCT images were oriented according to the anatomic structures of reference [26]. The horizontal plane was established using the skeletal orbitale and right porion. The sagittal plane passing through the nasion was perpendicular to the horizontal plane. The coronal plane passing through nasion, perpendicular to the horizontal plane and sagittal plane, was utilized to finish the reorientation. Simultaneously, we aligned the bilateral fronto-zygomatic point to the same coordinate of X-axis in the lateral view. The consistent coordination was used for all CBCT images.

2.2.3. Synthesizing the Cephalogram

A total of 430 reoriented CBCT images were used to automatically synthesize the CBCT-LC and MIP-LC by the Dolphin software, to prepare the dataset (Figure 1). A total of 860 images were prepared for multi-stage CNN training and testing. Based on previous findings [27,28], 80% of the data were prepared for deep CNNs training and 20% of data were used for testing, or 345 and 85 for each dataset. The synthesized images were saved with a range of pixel size width at 2048 pix, and height at 1755–1890 pix in JPG format.

2.3. Reproducibility of Intra-Examiner

We randomly selected 50 CBCT-LC and 50 MIP-LC to verify the inter-examiner reproducibility. All landmarks were identified twice at intervals of two weeks, by a single examiner (MJK). The See-through Ceph (See-through Tech Inc., Seoul, Korea) software was implemented to accomplish landmark identification. Pixel values of each landmark were extracted in an Excel file (version 2010; Microsoft, Redmond, Washington). The intraclass-correlation-coefficient (ICC) was calculated to indicate reproducibility in intra-examiner repetitive identification with 95% confidence intervals. Statistical analyses were performed using the SPSS software (version-22.0, SPSS Inc., Chicago, IL, USA).

2.4. Multi-Stage CNNs Architecture

The multi-stage CNNs used in this study was developed with customized open-source Keras library and TensorFlow (Google, Mountain View, CA, USA), backend on Python language. The computer components used in this study were—CPU: AMD Ryzen 7 1700; Mainboard: MSI Z370-A PRO; Ram: SAMSUNG DDR4 16G x 2; GPU: Geforce, GTX 1080Ti; POWER: Master Watt Lite 600 80PLUS Standard 230V EU; SSD: SAMSUNG 850 EVO Series 256 GB; and Linux: Ubuntu 14.04 (Canonical, London, UK).
One of the prominent abilities in deep CNNs learning is that it disseminates salient feature information based on a hierarchy to the subsequent layers. CNNs’ architecture is composed of convolution layers, pooling, and dense layers (fully connected layer). The target features of image were extracted from the convolution layers and pooling, during training. At the convolution layer, two-dimensional simple addition and multiplication were performed using suitable filters. Convolution operation outputs the intuitive measurement of the spatial similarity of the input. The CNNs learns from a filter that observes a specific image pattern in some spatial location of the previous layer’s output. The pooling layer is to sort out sampling and to prevent the number of parameters from increasing gradually. Commonly used pooling operation in CNNs is maximum pooling, which takes the maximum value pixels. Last layers of CNNs are dense layers, where a set of neurons are fully connected and the feature of input data to the network is classified to make a final decision [18,19].
At the image preprocessing stage, the first input model was developed with 400 widths, 400 height, and three-color channels—R (red), G (green), B (blue). In the model architecture, six convolutional layers extracted target features from the input image.
Mathematically, a convolution of two functions ‘f’ and ‘g’ was defined as:
f g i = j = 1 m g j · f i j + m 2
Therefore, the algorithm could only be expressed by the dot products in the input function and a kernel function, which we used. A unit conversion used 1 mm = 10 pixels. To increase the learning effect, data augmentation techniques were applied as follows—up to left, right, up and down 50 pixels shift, up to 10° rotation left and right each side. To input the training dataset, the deep CNNs must first learn the full image. At a second learning phase, each 15 landmarks on the lateral cephalograms was cropped and trained with different sizes—250, 200, 150, 100, and 50, so that a total of five stages of multiple convolutional layers were arranged in parallel. Fifteen landmarks used for the Tweemac cephalometric analysis were to be intensively trained by CNNs [29]. The definition of 15 landmarks are described in Table 1. Schematic diagram of our proposed multi-stage CNNs architecture is in Figure 2, and the visualization is in Figure 3.

2.5. System Evaluation

The accuracy of AI prediction was evaluated by mean radial errors (MRE) and a successful detection rate (SDR). MRE (mm) was the absolute distance differences between the manual identification (truth ground) and the AI prediction position. Definition:
M R E   =   i = 1 n R 1 n   ( m m )
Standard deviation (SD) = i = 1 n R i M R E n 1 2 , R = Δ x 2 + Δ y 2 .
SDR (%) represents percentages of the absolute landmark distance difference between manual identification and the AI prediction position, if the prediction range was less than 2 mm, it was considered to be the clinical acceptance level.
S D R   ( % )   =   n u m b e r   o f   a c c u r a t e   i d e n t i f i c a t i o n n u m b e r   o f   i d e n t i f i c a t i o n × 100 %
Common ranges of ≤2 mm, 2.5 mm, 3 mm, 4 mm were used to divide the groups for the number of accurate identification in SDR.

2.6. AI Prediction on Different Lateral Cephalograms

The paired t-test was used to compare the AI prediction on CBCT-LC and MIP-LC. To verify whether AI could make better prediction on MIP-LC or not, statistical analysis was performed using the SPSS software (version 22.0; IBM, Armonk, NY, USA). Measurements were calculated and statistically analyzed at the 0.05 level of significance.

3. Results

The ICC was 0.99, which had high rate of reproducibility in intra-examiner repetitive identification.
The results of AI prediction with the combination data showed an average MRE of 1.03 ± 1.29 mm, and SDR of 2.0 mm, 2.5 mm, 3.0 mm, and 4.0 mm precision ranges achieved 87.13%, 91.19%, 93.52%, 96.59%, respectively. The details between manual identification and AI prediction for each landmark are described in Table 2 and Figure 4.
The MRE for each landmark revealed that nasion showed the highest accuracy, and gonion showed the lowest. Eight landmarks out of the nineteen yielded distance errors ranging within 1 mm. Six landmarks showed accurate MRE range within 2 mm. Only gonion showed 2.04 mm in MRE measurements.
The paired t-test showed that there were no significant differences between truth ground and AI prediction on CBCT-LC. No significant differences were observed between the truth ground and AI prediction on MIP-LC. The details are described in Table 3 and Table 4.

4. Discussion

A fully automated landmark identification system was presented as an alternative option for consistent cephalometric landmark identification in repetitive tasks [30]. Several factors might affect deep learning prediction—the data size, the number of layers in the architecture, computer components, and the image resolution [31,32]. In this study, we intended to train the multi-stage CNNs with two modalities of lateral cephalograms synthesized from CBCT to increase the training samples, to broaden the training ranges, and to enhance the ability of image recognition without additional radiation to the patient.
The purpose of creating and using combination dataset in this study was to perform effective and efficient landmark identification both in manual and AI prediction. Although we used lateral cephalograms synthesized by CBCT to minimize the superimposed and layered bilateral structures, layered images still remained as confounding variables. For instance, the superimposed mandible prominently made two lines, especially in patients who had asymmetry that made it difficult to decipher gonion. Therefore, as a compensation, we tried to use the MIP image to reinforce these defects. However, the paired t-test showed that there were no significant differences between truth ground and AI prediction on CBCT-LC, MIP-LC. The AI learned the examiner’s identification pattern from preprocessed image data in deep learning, which explained that there were no significant differences in AI prediction on the CBCT-LC or MIP-LC, in our results. Whether the dataset was combined or not, AI would prediction on the same corresponding position with the same pattern. Details are explained below.
Deep learning is conceptually similar to supervised machine learning, but there are some differences. Supervised machine learning requires that the researcher process the image to extract target features. Deep learning works directly on the data image; and automatically reduces the burden of work. Traditional machine learning should have created a new algorithm for each new datum. On the other hand, retraining in the deep learning model is possible with new data. Substantial number data were used to train the deep learning architecture. Among the deep learning options, CNNs architecture had the outstanding ability to recognize particular appearance patterns that were widely used. This is an essential aspect of cephalometric landmark identification [33,34]. Classic deep CNNs are composed of a convolutional layer, pooling layer, and a fully connected layer. Although CNNs have the outstanding ability to recognize images that are used in medical image science, robustness is limited for geometrical transformations, other than parallel movements, such as scaling and rotation. Therefore, image features with scaling and rotation would present recognition errors in CNNs, but it steadily improved in CNNs architectures.
In this study, we proposed the multi-stage CNNs architecture that was constructed by stacking the convolution layers. The multi-stage architecture consisted of multiple convolutional layers and the number of layers was arranged in increasing parallel order. Since these CNNs structures weights of each convolution layer were connected to share features with corresponding layers of other stages, features of the original input image transferred to each stage. The features of preprocessed image extracted at all stages were concatenated and fulfilled to the integration layer, which showed strengths to improve the accuracy [35,36,37].
Most recent systems were developed using a conventional lateral cephalogram. However, the lateral cephalogram synthesized by CBCT had advantages—customization of resolution, ability to re-orientate the image to enhance the image quality, and use of an orthogonal projection to reducing the interference caused by superimposition of bilateral structures. Due to difficulties in the segmentation of internal organs, the MIP algorithm was invented in 1988 by the nuclear medicine doctor Jerold Wallis [38,39]. The MIP image helped to find the relative 3D position of anatomic structures. Emergence of a three-dimensional image and this display technique offered the information on depth and volume. However, since the three-dimensional image required a rendering process, inevitably noise and blurring occurred, inhibiting accurate detection of the anatomic structure. The MIP algorithm provided better detection of the relationship between objects and the surface contours. Structures of interest could be highlighted by selecting the objects, and the surface contour could be indirectly inferred by the depth information. One of the matters with the lateral cephalogram was superimposed bilateral structures, which could be eliminated by the enhanced image contrast in MIP. While MIP was visualized in planar space, it still included a part of the three-dimensional structural information, without using the full three-dimension rendering [39].
Previous studies compared reproducibility between conventional lateral cephalogram and different types of CBCT synthesized 2D lateral cephalograms (MIP, Ray-sum or usual). Conclusion of these studies was that linear and angular measurement derived from CBCT synthesized cephalogram resulted in higher consistency than conventional cephalogram [1,2,6,40,41]. We used CBCT synthesized combination data to perform efficient landmark identification for training. We used 15 landmarks positioned on the anatomic structure contours that were displayed more clearly in the MIP image. Comparison of AI prediction results between CBCT-LC and MIP-LC are shown in Figure 5.
Although the CBCT was currently used to provide 3D information, further studies would lead to better standardization in cephalometry. The ability of direct landmark identification and measurement was lacking. Even though we obtained the 3D images from patients to use current research-based analyses, it had to be transformed to a 2D aspect [42,43].
In 2017, the first fully-automated cephalometric landmark identification system was applied to a deep CNNs learning model [22]. They trained 400 conventional lateral cephalogram datasets that included 19 landmarks. The result showed a 75.58% SDR range of 0 to 2 mm. In 2019, 1311 trained conventional lateral cephalograms using the You-Only-Look-Once version 3 (YOLOv3) with 80 landmarks achieved an 80.4% SDR range of 0 to 2 mm. The prediction result achieved approximately 5% higher in all ranges of SDR, than a previous study [24]. Another study used a personal computer to develop the CNNs architecture training with 153 lateral cephalograms and tested 66 images. Ten landmarks frequently used in cephalometry were included. There were no significant differences between the manual and automatic prediction in the cephalometric analysis, but the average prediction errors recorded 17.02 in pixel (approximately 4.50 mm) [23]. In 2020, many studies introduced new CNNs algorithms or methods [44,45,46]. Kunz et al. [44] used customized Keras and Tensorflow, similar to us. In their study, MRE was not described but AI prediction showed similar results with examiners. Kim et al. [45] Reported a web-based deep learning method, which had advantage in accessibilities. They evaluated four different dataset groups. The highest SDR range of 0 to 2 mm achieved 84.53 % with an acceptable MRE, 1.37 ± 1.79 mm. Comparison of SDR in different CNNs architecture, based on fully automatic landmark identification systems, is described in Table 5 and Figure 6. However, the deep learning model also applied automatic landmark identification of the CBCT image. Some important factors that affected accuracy for the CNNs-based fully automatic landmark identification system were—CNNs architecture’s structure, sufficient learning data, and the supervising method. Comparison to the above studies, we employed improvements—(i) to achieve better accuracy we increased the number of layers and only extracted the necessary features for learning; (ii) we used a sufficient dataset; and (iii) used combination data to perform efficient supervising (setting the truth ground). The number of landmarks did not significantly affect the system accuracy. However, an increased number of landmarks might enhance clinical procedures.
The highlights of this study were proposal of customized CNNs architecture and we reported the AI prediction results of CBCT-LC and MIP-LC as a transitional study of 3D CBCT data. However, a limitation of this study, first of all, was that the amount of data required to achieve the expected accuracy could not be explained. A previous study reported that greater the number of data, higher the accuracy [47]. They used 2200 training data as the maximum number to prove it and results of MRE showed more than 1.5 mm. Although we did not try to discover the suitable number of data with a gradual dataset, this was possible to presume through the results, which achieved our expectations, that the means and the amount of data, was used properly in this study. Second, we could not compare if our CNNs architecture was better than that of similar studies as different computer components and dataset were used in other studies. In order to identify whether the combination dataset had any benefit, we needed to compare with a unified dataset in the further study.

5. Conclusions

Recently, artificial intelligence technology is rapidly being commercialized. In this study, we introduced a method to customize the CNNs architecture using open-source. A new learning concept in the aspect of the CNNs architecture and dataset achieved superior results compared to previous studies that obtained 87.10% SDR range of 0 to 2 mm with 1.03 mm average MRE. However, no control group was established in this study. Hence, a comparison of CBCT-LC combined with MIP-LC training and CBCT-LC training is needed in future studies, to verify if a combination dataset has any benefit or not.

Author Contributions

Conceptualization, M.-J.K. and S.-H.K.; methodology, M.-J.K.; software, S.H.O.; validation, M.-J.K. and S.H.O.; formal analysis, Y.L. and H.-W.A.; investigation, S.H.O.; resources, Y.L. and S.-H.K.; data curation, M.-J.K. and S.-H.K.; writing—original draft preparation, M.-J.K.; writing—review and editing, Y.L., H.-W.A. and S.-H.K.; visualization, S.-H.K. and G.N.; supervision, S.-H.K. and G.N.; project administration, S.-H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Kyung Hee University Dental Hospital (protocol code IRB-KH DT19013 and date of approval 24 July 2019).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data sharing not applicable.

Acknowledgments

This research article was partly from one the authors (M.J.K.) PhD thesis. The authors want show special thanks to Kyung-Jae Hong, CEO of See-Through Tech incorporation, Seoul, Korea for the support of the final form of the machine-learning system developed by computer engineers of See-through Tech incorporation.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Smektala, T.; Jedrzejewski, M.; Szyndel, J.; Sporniak-Tutak, K.; Olszewski, R. Experimental and clinical assessment of three-dimensional cephalometry: A systematic review. J. Craniomaxillofac. Surg. 2014, 42, 795–801. [Google Scholar] [CrossRef] [PubMed]
  2. Gribel, B.F.; Gribel, M.N.; Frazao, D.C.; McNamara, J.A., Jr.; Manzi, F.R. Accuracy and reliability of craniometric measurements on lateral cephalometry and 3D measurements on CBCT scans. Angle Orthod. 2011, 81, 26–35. [Google Scholar] [CrossRef] [PubMed]
  3. Hong, M.H.; Kim, M.J.; Shin, H.J.; Cho, H.J.; Baek, S.H. Three-dimensional surgical accuracy between virtually planned and actual surgical movements of the maxilla in two-jaw orthognathic surgery. Korean J. Orthod. 2020, 50, 293–303. [Google Scholar] [CrossRef] [PubMed]
  4. Kapila, S.D.; Nervina, J.M. CBCT in orthodontics: Assessment of treatment outcomes and indications for its use. Dentomaxillofac. Radiol. 2015, 44, 20140282. [Google Scholar] [CrossRef] [Green Version]
  5. Na, E.R.; Aljawad, H.; Lee, K.M. A comparative study of the reproducibility of landmark identification on posteroanterior and anteroposterior cephalograms generated from cone-beam computed tomography scans. Korean J. Orthod. 2019, 49, 41–48. [Google Scholar] [CrossRef]
  6. Clendenen, L.H. Intra- and Inter-Operator Reliability of Cephalometric Landmark Identification: A Comparison Amongst CBCT Ray-Sum, Maximum Intensity Projection Techniques, and Conventional Digital Cephalometric Radiographs; University of Missouri—Kansas City: Kansas City, MO, USA, 2010. [Google Scholar]
  7. Lee, J.G.; Jun, S.; Cho, Y.W.; Lee, H.; Kim, G.B.; Seo, J.B. Kim, N. Deep Learning in Medical Imaging: General Overview. Korean J. Radiol. 2017, 18, 570–584. [Google Scholar] [CrossRef] [Green Version]
  8. Dorgham, O.M.; Laycock, S.D.; Fisher, M.H. GPU accelerated generation of digitally reconstructed radiographs for 2-D/3-D image registration. IEEE Trans. Biomed. Eng. 2012, 59, 594–603. [Google Scholar] [CrossRef]
  9. Ruellas, A.C.; Tonello, C.; Gomes, L.R.; Yatabe, M.S.; Macron, L.; Lopinto, J.; Goncalves, J.R.; Garib Carreira, D.G.; Alonso, N.; Souki, B.Q.; et al. Common 3-dimensional coordinate system for assessment of directional changes. Am. J. Orthod. Dentofacial. Orthop. 2016, 149, 645–656. [Google Scholar] [CrossRef]
  10. Jung, P.G.; Lee, G.C.; Moon, C.H. Comparison of cone-beam computed tomography cephalometric measurements using a midsagittal projection and conventional two-dimensional cephalometric measurements. Korean J. Orthod. 2015, 45, 282–288. [Google Scholar] [CrossRef] [Green Version]
  11. Lee, E.H.; Yu, H.S.; Lee, K.J.; Han, S.S.; Jung, H.D. Comparison of three midsagittal planes for three-dimensional cone beam computed tomography head reorientation. Korean J. Orthod. 2020, 50, 3–12. [Google Scholar] [CrossRef]
  12. Cattaneo, P.M.; Melsen, B. The use of cone-beam computed tomography in an orthodontic department in between research and daily clinic. World J. Orthod. 2008, 9, 269–282. [Google Scholar] [PubMed]
  13. Levy-Mandel, A.D.; Venetsamopolus, A.N.; Tsosos, J.K. Knowledge based landmarking of cephalograms. Comput. Biomed. Res. 1986, 19, 282–309. [Google Scholar] [CrossRef]
  14. Ibragimov, B.; Likar, B.; Pernuš, F.; Tomaž, V. Automatic Cephalometric X-Ray Landmark Detection by Applying Game Theory and Random Forests. Proc. ISBI Int. Symp. Biomedical. Imaging 2014. [Google Scholar]
  15. Leonardi, R.; Giordano, D.; Maiorana, F.; Spampinato, C. Automatic Cephalometric Analysis: A Systematic Review. Angle Orthod. 2008, 78, 145–151. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Lindner, C.; Wang, C.W.; Huang, C.T.; Li, C.H.; Chang, S.W.; Cootes, T.F. Fully Automatic System for Accurate Localisation and Analysis of Cephalometric Landmarks in Lateral Cephalograms. Sci. Rep. 2016, 6, 33581. [Google Scholar] [CrossRef]
  17. Wang, S.; Li, H.; Li, J.; Zhang, Y.; Zou, B. Automatic Analysis of Lateral Cephalograms Based on Multiresolution Decision Tree Regression Voting. J. Healthc. Eng. 2018, 2018, 1797502. [Google Scholar] [CrossRef]
  18. Schmidhuber, J. Deep learning in neural networks: An overview. Neural. Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  19. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  20. Hamet, P.; Tremblay, J. Artificial intelligence in medicine. Metabolism 2017, 69S, S36–S40. [Google Scholar] [CrossRef]
  21. Zhang, Y.; An, M. Deep Learning- and Transfer Learning-Based Super Resolution Reconstruction from Single Medical Image. J. Healthc. Eng. 2017, 2017, 5859727. [Google Scholar] [CrossRef] [Green Version]
  22. Arık, S.Ö.; Ibragimov, B.; Xing, L. Fully automated quantitative cephalometry using convolutional neural networks. J. Med Imaging 2017, 4, 014501. [Google Scholar] [CrossRef] [PubMed]
  23. Nishimoto, S.; Sotsuka, Y.; Kawai, K.; Ishise, H.; Kakibuchi, M. Personal Computer-Based Cephalometric Landmark Detection With Deep Learning, Using Cephalograms on the Internet. J. Craniofac. Surg. 2019, 30, 91–95. [Google Scholar] [CrossRef] [PubMed]
  24. Park, J.H.; Hwang, H.W.; Moon, J.H.; Yu, Y.S.; Kim, H.S.; Her, S.B.; Srinivasan, G.; Aljanabi, M.N.A.; Donatelli, R.E.; Lee, S.J. Automated identification of cephalometric landmarks: Part 1—Comparisons between the latest deep-learning methods YOLOV3 and SSD. Angle Orthod. 2019, 89, 903–909. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Kim, M.J.; Liu, Y.; Oh, S.H.; Ahn, H.W.; Kim, S.H.; Nelson, G. Evaluation of multi-stage convolutional neural network based fully automatic landmark identification system using CBCT synthesized posteroanterior cephalometric images. Korean J. Orthod. 2020; in press. [Google Scholar]
  26. Lee, Y.J.; Park, J.H.; Chang, N.Y.; Lee, M.Y.; Kim, B.C.; Seo, H.Y.; Mangal, U.; Chae, J.M. Assessment of bone density changes following two-jaw surgery using multidetector computed tomography: A pilot study. Korean J. Orthod. 2020, 50, 157–169. [Google Scholar] [CrossRef]
  27. Cho, J.H.; Lee, K.W.; Shin, E. How Much Data Is Needed to Train A Medical Image Deep Learning System to Achieve Necessary High Accuracy. arXiv 2015, 1511, 06348. [Google Scholar]
  28. Steyerberg, E.W.; Eijkemans, M.J.C.; Harrell, F.E.; Habbema, J.; Dik, F. Prognostic Modeling with Logistic Regression Analysis: In Search of a Sensible Strategy in Small Data Sets. Med Decis. Mak. 2001, 21, 45–56. [Google Scholar] [CrossRef]
  29. Kim, S.H.; Ahn, H.W.; Chung, K.R.; Gerald, N. The biocreative strategy. Part 2: The Tweemac analysis. J. Clin. Orthod. 2018, 52, 359–371. [Google Scholar]
  30. Hwang, H.W.; Park, J.H.; Moon, J.H.; Yu, Y.S.; Kim, H.S.; Her, S.B.; Srinivasa, G.; Aljanabi, M.N.A.; Donatelli, R.E.; Lee, S.J. Automated Identification of Cephalometric Landmarks: Part 2-Might It Be Better Than human? Angle Orthod. 2019, 90, 69–76. [Google Scholar] [CrossRef] [Green Version]
  31. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J.M. Convolutional Neural Networks for Medical Image Analysis: Full Training or Fine Tuning? IEEE Trans. Med. Imaging. 2016, 35, 1299–1312. [Google Scholar] [CrossRef] [Green Version]
  32. Anwar, S.H.; Majid, M.; Qayyum, A.; Awais, M.; Alnowam, M.; Khan, M.K. Medical Image Analysis using Convolutional Neural Networks: A Review. J. Med. Syst. 2018, 42, 226. [Google Scholar] [CrossRef] [Green Version]
  33. Schwendicke, F.; Golla, T.; Dreher, M.; Krois, J. Convolutional neural networks for dental image diagnostics: A scoping review. J. Dent. 2019, 91, 103226. [Google Scholar] [CrossRef] [PubMed]
  34. Yadav, S.S.; Jadhav, S.M. Deep convolutional neural network based medical image classification for disease diagnosis. J. Big. Data 2019, 6, 113. [Google Scholar] [CrossRef] [Green Version]
  35. Takahashiy, R.; Takashi, M.; Kuniaki. Multi-Stage Convolutional Neural Networks for Robustness to Scale Transformation. Int. Symp. Nonlinear Theory Appl. 2017. [Google Scholar] [CrossRef]
  36. Takahashi, R.; Matsubara, T.; Uehara, K. A Novel Weight-Shared Multi-Stage CNN for Scale Robustness. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1090–1101. [Google Scholar] [CrossRef]
  37. Sun, M.L.; Song, Z.J.; Jiang, X.H.; Pan, J.; Pang, Y.W. Learning Pooling for Convolutional Neural Network. Neurocomputing 2017, 224, 96–104. [Google Scholar] [CrossRef]
  38. Wallis, J.W.; Miller, T.R.; Lerner, C.A.; Kleerup, E.C. Three-dimensional display in nuclear medicine. IEEE Trans. Med. Imaging 1989, 8, 297–300. [Google Scholar] [CrossRef]
  39. Wallis, J.W.; Miller, T.R. Three-dimensional display in nuclear medicine and radiology. J. Nucl. Med. 1991, 32, 534–546. [Google Scholar]
  40. Rossini, G.; Cavallini, C.; Cassetta, M. 3D cephalometric analysis obtained from computed tomography. Review of the literature. Ann. Stomatol. 2011, 2, 31–39. [Google Scholar]
  41. Navarro, R.L.; Oltramari-Navarro, P.V.P.; Fernandes, T.M.F.; Oliveira, G.F.; Conti, A.C.C.; Almeida, M.R.; Almeida, R.R. Comparison of manual, digital and lateral CBCT cephalometric analyses. J. Appl. Oral Sci. 2013, 21, 167–176. [Google Scholar] [CrossRef] [Green Version]
  42. Gribel, B.F.; Gribel, M.N.; Manzi, F.R.; Brooks, S.L.; McNamara, J.A., Jr. From 2D to 3D: An algorithm to derive normal values for 3-dimensional computerized assessment. Angle Orthod. 2011, 81, 3–10. [Google Scholar] [CrossRef] [Green Version]
  43. Park, C.S.; Park, J.K.; Kim, H.; Han, S.S.; Jeong, H.G.; Park, H. Comparison of conventional lateral cephalograms with corresponding CBCT radiographs. Imaging Sci. Dent. 2012, 42, 201–205. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Kunz, F.; Stellzig-Eisenhauer, A.; Zeman, F.; Boldt, J. Artificial intelligence in orthodontics. J. Orofac. Orthop. 2020, 81, 52–68. [Google Scholar] [CrossRef] [PubMed]
  45. Kim, H.N.; Shim, E.J.; Park, J.E.; Kim, Y.J.; Lee, U.Y.; Kim, Y.J. Web-based fully automated cephalometric analysis by deep learning. Comput. Methods Programs Biomed. 2020, 194, 105513. [Google Scholar] [CrossRef] [PubMed]
  46. Lee, J.H.; Yu, H.J.; Kim, M.J.; Kim, J.W.; Choi, J.E. Automated cephalometric landmark detection with confidence regions using Bayesian convolutional neural networks. BMC Oral Health 2020, 20, 270. [Google Scholar] [CrossRef]
  47. Moon, J.H.; Hwang, H.W.; Yu, Y.S.; Kim, M.G.; Donatelli, R.E.; Lee, S.J. How much deep learning is enough for automatic identification to be reliable?: A cephalometric example. Angle Orthod. 2020, 90, 823–830. [Google Scholar] [CrossRef]
Figure 1. Image acquisition from CBCT: Import the CBCT data to the Dolphin software and reorient the head image. Using the ‘Build X-ray’ function in the software to synthesize the CBCT lateral cephalogram (CBCT-LC) and CBCT MIP lateral cephalogram (MIP-LC) with orthogonal X-ray. All synthesized image data were saved with a range of pixel size width at 2048 pix and the height at 1755–1890 pix in the JPG format.
Figure 1. Image acquisition from CBCT: Import the CBCT data to the Dolphin software and reorient the head image. Using the ‘Build X-ray’ function in the software to synthesize the CBCT lateral cephalogram (CBCT-LC) and CBCT MIP lateral cephalogram (MIP-LC) with orthogonal X-ray. All synthesized image data were saved with a range of pixel size width at 2048 pix and the height at 1755–1890 pix in the JPG format.
Sensors 21 00505 g001
Figure 2. Schematic diagram of our proposed multi-stage CNNs architecture. Put the preprocessed image data. Input data passing through the deep CNNs model of composition with 6 convolution layers, with 32 × 2, 64 × 2, 128, 256 nodes, and max pooling for the features to be extracted. Subsequently, the two dense layers classified the features to make final decision.
Figure 2. Schematic diagram of our proposed multi-stage CNNs architecture. Put the preprocessed image data. Input data passing through the deep CNNs model of composition with 6 convolution layers, with 32 × 2, 64 × 2, 128, 256 nodes, and max pooling for the features to be extracted. Subsequently, the two dense layers classified the features to make final decision.
Sensors 21 00505 g002
Figure 3. The visualized effect of input data at convolution layers during training.
Figure 3. The visualized effect of input data at convolution layers during training.
Sensors 21 00505 g003
Figure 4. Mean Radial Error of each landmark and average MRE.
Figure 4. Mean Radial Error of each landmark and average MRE.
Sensors 21 00505 g004
Figure 5. Comparison of AI prediction between CBCT-LC (a,c,e,g,i) and MIP-LC (b,d,f,h,j); Orbitale c and d; Maxilla e and f; Porion g and h; and Mandible i and j. The red dot is manual identification (truth ground), and the blue dot is AI prediction.
Figure 5. Comparison of AI prediction between CBCT-LC (a,c,e,g,i) and MIP-LC (b,d,f,h,j); Orbitale c and d; Maxilla e and f; Porion g and h; and Mandible i and j. The red dot is manual identification (truth ground), and the blue dot is AI prediction.
Sensors 21 00505 g005aSensors 21 00505 g005b
Figure 6. Comparison of successful detection rate of different CNNs architecture based on fully automatic landmark identification system.
Figure 6. Comparison of successful detection rate of different CNNs architecture based on fully automatic landmark identification system.
Sensors 21 00505 g006
Table 1. Landmark definitions.
Table 1. Landmark definitions.
Landmarks Definition
Nasion (N)The most anterior point of the frontonasal suture
Orbitale (Or)Inferior margin of the orbit
Porion (Po)Superior margin of the external auditory meatus
Anterior nasal spine (ANS)Tip of anterior nasal spine
A point Point at the deepest concavity on the maxilla between the anterior nasal spine and prosthion
Maxillary central incisor root (Mx1r)Maxillary central incisor root
Maxillary incisal edge (Is)Maxillary incisal edge
Mandibular incisal edge (Ii)Mandibular incisal edge
Infradentale (Id)The highest and most anterior point on the alveolar process in the median plane between the mandibular central incisors
Mandibular central incisor root (Md1r)Mandibular central incisor root
B pointPoint at the deepest concavity on the mandibular symphysis between infradentale and pogonion
Pogonion (Pog)The most anterior midpoint of the chin of the mandibular symphysis
Menton (Me) The most inferior point of the mandibular symphysis
Gonion (Go)Most posterior inferior point on angle of mandible
Posterior occlusal plane point (Pop) Posterior occlusal plane point: mesio-buccal cusp of 1st molar
Table 2. System prediction results of 15 landmark identification in combination data.
Table 2. System prediction results of 15 landmark identification in combination data.
Landmark.MRE (mm)SD (mm)SDR (%)
2.0 mm2.5 mm3.0 mm4.0 mm
N0.561.49795.9796.5597.1298.27
Or0.691.39797.1298.2798.2798.85
Po1.431.68978.7384.4889.6594.82
ANS0.760.72693.6797.1298.2799.42
A1.222.69789.0891.9596.5598.27
Mx1r1.591.30869.5478.7386.295.4
Is0.611.08193.194.2594.8298.27
Id0.791.08090.894.2595.496.55
Ii0.650.92793.6794.2595.9797.7
Md1r1.391.26681.0388.589.6594.25
B1.090.89183.3392.5295.9798.85
Pog0.580.58095.9797.798.85100
Me0.590.54797.1298.2798.85100
Go2.041.72762.6472.9877.5885.63
Pop1.311.89785.0587.9389.6592.52
Average 1.031.28887.1391.1993.5296.59
Table 3. Comparison of landmark identification between truth ground and AI prediction on the CBCT synthesized lateral cephalograms (CBCT-LC).
Table 3. Comparison of landmark identification between truth ground and AI prediction on the CBCT synthesized lateral cephalograms (CBCT-LC).
Truth GroundAI Prediction CBCT-LCp-Value *
LandmarkMeanSDMinMaxMeanSDMinMax
N_x1642.7158.461479.001866.001641.3355.811487.661849.160.868
N_y564.09124.36285.00865.00551.58122.73259.93828.510.488
Or_x1561.9255.801363.001755.001558.8354.221378.061732.800.703
Or_y819.48105.41628.001079.00805.06105.67604.011047.610.350
Po_x851.0281.24526.00991.00850.2778.83530.84989.410.950
Po_y827.38107.98633.001085.00819.01107.73612.381088.110.593
ANS_x1665.2455.631505.001875.001663.5152.621497.871838.490.829
ANS_y1052.1386.66883.001280.001042.4390.04834.111275.940.453
A_x1642.3654.261496.001851.001640.7352.121497.961831.310.833
A_y1105.7584.83946.001327.001095.8388.24914.741325.590.423
Mx1r_x1618.6453.641468.001813.001612.8050.991479.121779.620.457
Mx1r_y1134.8079.14969.001349.001128.0279.84951.911323.550.545
Is_x1697.2668.221534.001926.001697.3365.101543.721899.460.994
Is_y1314.3975.451150.001499.001306.3078.151123.211506.450.455
Id_x1655.1079.851475.001873.001650.8975.271482.441848.200.716
Id_y1392.6064.481262.001546.001384.9370.471224.561557.320.401
Ii_x1677.8575.421515.001902.001677.0270.941514.511875.320.939
Ii_y1305.0569.121164.001475.001294.4772.271129.551466.300.282
Md1r_x1596.4582.521414.001788.001590.1778.161415.731778.640.605
Md1r_y1453.8758.291336.001575.001445.5064.191283.631621.690.326
B_x1607.0986.881424.001813.001602.5982.051424.101785.400.723
B_y1506.3764.141337.001628.001497.5171.161294.641650.670.358
Pog_x1613.8593.191419.001837.001610.0788.901419.371809.180.785
Pog_y1601.3458.201471.001704.001600.1158.441442.151712.630.880
Me_x1555.6696.741352.001783.001553.7791.111355.661748.780.894
Me_y1663.9355.231526.001770.001658.0655.991510.331773.850.445
Go_x992.7587.06725.001184.00988.6279.60719.041152.650.748
Go_y1373.9765.901202.001537.001355.2764.391213.181528.060.053
Pop_x1434.1558.521264.001616.001425.4956.551261.201588.020.314
Pop_y1263.2965.401130.001426.001253.5868.571104.161427.760.292
* Paired t-test was done for comparison of manual identification and AI prediction on CBCT-LC. Unit of measurement: Pixel.
Table 4. Comparison of landmark identification between truth ground and AI prediction on CBCT synthesized MIP lateral cephalograms (MIP-LC).
Table 4. Comparison of landmark identification between truth ground and AI prediction on CBCT synthesized MIP lateral cephalograms (MIP-LC).
Truth GroundAI prediction MIP-LC p-Value *
LandmarkMeanSDMinMaxMeanSDMinMax
N_x1642.7158.461479.001866.001646.1759.461479.271862.750.754
N_y564.09124.36285.00865.00562.05118.36288.38830.260.793
Or_x1561.9255.801363.001755.001565.4857.181369.461756.590.649
Or_y819.48105.41628.001079.00819.07100.80630.061048.800.959
Po_x851.0281.24526.00991.00850.4881.05517.92986.900.920
Po_y827.38107.98633.001085.00822.26101.70635.181053.430.156
ANS_x1665.2455.631505.001875.001670.2056.721505.481875.800.406
ANS_y1052.1386.66883.001280.001052.2286.83875.051278.580.752
A_x1642.3654.261496.001851.001645.5355.471496.601849.640.778
A_y1105.7584.83946.001327.001107.5283.06950.801328.570.752
Mx1r_x1618.6453.641468.001813.001621.6254.201479.401825.120.771
Mx1r_y1134.8079.14969.001349.001138.5477.21980.841351.760.333
Is_x1697.2668.221534.001926.001701.6768.601542.861927.340.601
Is_y1314.3975.451150.001499.001314.7974.571149.571501.820.599
Id_x1655.1079.851475.001873.001655.8478.721482.701873.200.799
Id_y1392.6064.481262.001546.001397.2965.691262.711561.400.381
Ii_x1677.8575.421515.001902.001679.9074.651512.091899.680.986
Ii_y1305.0569.121164.001475.001307.1569.111164.501474.310.770
Md1r_x1596.4582.521414.001788.001596.1982.441408.861815.760.684
Md1r_y1453.8758.291336.001575.001456.7162.781310.591610.440.997
B_x1607.0986.881424.001813.001608.8385.341422.781812.240.979
B_y1506.3764.141337.001628.001507.2867.611326.611642.090.435
Pog_x1613.8593.191419.001837.001615.0291.871415.981837.820.915
Pog_y1601.3458.201471.001704.001607.1157.091467.661703.100.294
Me_x1555.6696.741352.001783.001556.1794.121353.871767.150.862
Me_y1663.9355.231526.001770.001666.4255.771522.651768.890.657
Go_x992.7587.06725.001184.00992.7181.71719.391169.110.941
Go_y1373.9765.901202.001537.001374.0463.351186.131531.910.493
Pop_x1434.1558.521264.001616.001429.5360.571253.621616.570.064
Pop_y1263.2965.401130.001426.001261.5665.561131.641427.020.182
* Paired t-test was done for comparison of manual identification and AI prediction on CBCT-MIP. Unit of measurement: Pixel.
Table 5. Comparison of automatic landmark identification system prediction results.
Table 5. Comparison of automatic landmark identification system prediction results.
Method.LandmarkTotal Data SDR (%)
2.0 mm2.5 mm3.0 mm4.0 mm
SDD221940075.681.384.788.1
YOLOv32480131180.487.492.096.2
Web-based4523207584.590.193.296.8
Proposed1586087.191.293.596.6
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kim, M.-J.; Liu, Y.; Oh, S.H.; Ahn, H.-W.; Kim, S.-H.; Nelson, G. Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images. Sensors 2021, 21, 505. https://doi.org/10.3390/s21020505

AMA Style

Kim M-J, Liu Y, Oh SH, Ahn H-W, Kim S-H, Nelson G. Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images. Sensors. 2021; 21(2):505. https://doi.org/10.3390/s21020505

Chicago/Turabian Style

Kim, Min-Jung, Yi Liu, Song Hee Oh, Hyo-Won Ahn, Seong-Hun Kim, and Gerald Nelson. 2021. "Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images" Sensors 21, no. 2: 505. https://doi.org/10.3390/s21020505

APA Style

Kim, M. -J., Liu, Y., Oh, S. H., Ahn, H. -W., Kim, S. -H., & Nelson, G. (2021). Automatic Cephalometric Landmark Identification System Based on the Multi-Stage Convolutional Neural Networks with CBCT Combination Images. Sensors, 21(2), 505. https://doi.org/10.3390/s21020505

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop