Next Article in Journal
Hazardous Effects of SiO2 Nanoparticles on Liver and Kidney Functions, Histopathology Characteristics, and Transcriptomic Responses in Nile Tilapia (Oreochromis niloticus) Juveniles
Previous Article in Journal
Increased Circulating H3 Histone in Response to Repeated Bouts of Exercise Does Not Associate with Parallel Alterations of Cell-Free DNA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Convolutional Neural Networks and Geometric Moments to Identify the Bilateral Symmetric Midplane in Facial Skeletons from CT Scans

by
Rodrigo Dalvit Carvalho da Silva
1,2,*,
Thomas Richard Jenkyn
1,2,3,4,5 and
Victor Alexander Carranza
1,6
1
Craniofacial Injury and Concussion Research Laboratory, Western University, London, ON N6A 3K7, Canada
2
School of Biomedical Engineering, Faculty of Engineering, Western University, London, ON N6A 3K7, Canada
3
Department of Mechanical and Materials Engineering, Western University, London, ON N6A 3K7, Canada
4
School of Kinesiology, Faculty of Health Sciences, Western University, London, ON N6A 3K7, Canada
5
Wolf Orthopaedic Biomechanics Laboratory, Fowler Kennedy Sport Medicine Clinic, London, ON N6A 3K7, Canada
6
School of Biomedical Engineering, Faculty of Engineering, Collaborative Specialization in Musculoskeletal Health Research, and Bone and Joint Institute, Western University, London, ON N6A 3K7, Canada
*
Author to whom correspondence should be addressed.
Biology 2021, 10(3), 182; https://doi.org/10.3390/biology10030182
Submission received: 3 January 2021 / Revised: 11 February 2021 / Accepted: 25 February 2021 / Published: 2 March 2021

Abstract

:

Simple Summary

The bilateral symmetry midplane of the facial skeleton plays a critical role in reconstructive craniofacial surgery. By accurately locating the midplane, surgeons can use the undeformed side of the face as a template for the malformed side. However, the location of the midline is still a subjective procedure, despite its importance. This study aimed to present a 3D technique for automatically calculating the craniofacial symmetry midline of the facial skeleton from CT scans using deep learning techniques. A total of 195 skull images were evaluated and were found to be reliable and provided good accuracy in symmetric images.

Abstract

In reconstructive craniofacial surgery, the bilateral symmetry of the midplane of the facial skeleton plays an important role in surgical planning. Surgeons can take advantage of the intact side of the face as a template for the malformed side by accurately locating the midplane to assist in the preparation of the surgical procedure. However, despite its importance, the location of the midline is still a subjective procedure. The aim of this study was to present a 3D technique using a convolutional neural network and geometric moments to automatically calculate the craniofacial midline symmetry of the facial skeleton from CT scans. To perform this task, a total of 195 skull images were assessed to validate the proposed technique. In the symmetry planes, the technique was found to be reliable and provided good accuracy. However, further investigations to improve the results of asymmetric images may be carried out.

1. Introduction

Craniomaxillofacial reconstructive surgery is a complex and difficult multidisciplinary technique due to the intricate anatomy of the skull. The aim of craniomaxillofacial reconstruction surgery is to preserve the patient’s appearance, strengthen facial functions, and regain the bilateral symmetry of the craniofacial skeleton. However, craniomaxillofacial reconstructive surgery may lead to complications such as bone disorders, congenital deformities, trauma, pathologies, genetic abnormalities, and cancers. A precise recognition of the bilateral symmetry facial midplane is an imperative step for pre-surgical planning and implant design techniques. For facial restoration, this midplane plays a major role when one side of the image is replicated and used as a guide to recreate the deformed or injured side. The authors of [1] have established a widely agreed approach for defining the midline of the craniofacial skeleton. To date, however, the most popular two-dimensional image application method, or the midsagittal plane (MSP) for a three-dimensional object, is the method introduced by [2].
There are a few approaches that aim to simplify the task of locating the midline plane of the facial skeleton. One technique proposed by [3] describes a semiautomatic system that, in conjunction with surface models reconstructed from computed tomography images (CT), uses principal component analysis (PCA) and the iterative closest point (ICP) alignment method. The first step is to determine the direction of the mirror plane correctly. This was achieved by using PCA to match the replicated mesh and the initial mesh roughly. Then, the ICP algorithm was described by a refined registration. The downside of this approach was the dependency on the central point of the image for the approximation of the symmetrical plane (obtained using the average of the vertices of the facial mesh). If the central point was in the wrong position due to any external factors (such as imperfect symmetry), this approach would lead to a symmetrical plane in the wrong direction and position. In addition, this algorithm is not able to adjust and learn from previous images to improve its performance, limiting its capabilities.
Alternatively, [4] determines the midline symmetry plane by using boney landmarks to create a midline representing facial symmetry. For a stack of horizontal lines crossing bilaterally through the facial skeleton containing boney landmarks, this approach essentially measures the midline symmetry plane as a perpendicular midpoint. This approach involves the manual collection of a variety of cephalometric boney landmarks in the dataset by either specifically locating the landmarks on the plane (which requires great attention by an expert user) or by using the midline as a reference and locating the landmarks at equal distances from the midline. However, manual skeletal landmark selection is ineffective, time-consuming, and reliant on an expert operator, resulting in errors in the measurement of the symmetry plane [5] outlines an ICP-based process for automated symmetry plane detection of 3D asymmetrically scanned human faces that uses particle swarm optimization. This approach starts with a discrete 3D model. The symmetry plane is tested by a tentative first attempt using a PCA algorithm. The model is then refined iteratively by a Levenberg–Marquardt algorithm before its final prediction is obtained. This revised version enhances the shortcomings of [6], but the current implementation also struggles to integrate self-learning to maximize the result of the model and misses the ability to learn from previous versions.
By minimizing the error-index of the symmetry plane, an automated method based on an iterative process was recently proposed by [7]. To automatically correct the initial symmetry plane, with a significant contribution to the use of the rotation matrix derived from the registration process, this method performs analytical data analysis in 3D point sets derived from CT images. First, the plane was divided into two groups by the initial symmetry plane estimated by the PCA and the collection of skull points. Then, to match two point-sets, the ICP registration method was used.
Most recently, ref. [8] introduced a novel automatic concept for determining the bilateral symmetry midline of the facial skeleton based on invariant moments. This technique creates a dataset from images aligned using cephalometric landmarks. The images are then rotated from 14° to 15° with a resolution of 0.5°. Then, after comparing different feature extractors, pseudo-zernike moments were selected for having the best accuracy using the k-nearest neighbors classifier. Finally, after detecting the rotation degree of the image, the midpoint is calculated using geometric moments. However, this model still has some limitations. For instance, this method uses 2D images with an image resolution of 128 × 128 which becomes difficult in real applications on different image modalities such as computed tomography and magnetic resonance imaging (MRI). Additionally, this technique was not tested on non-symmetrical skull images which may affect its results.
Thus, this study aims to present a 3D technique for automatically calculating the craniofacial symmetry midplane from CT scans using convolutional neural network (CNN) and geometric moments. Figure 1 shows the overview of the proposed method. First, using 3D U-net, the skull is removed from CT images to create a dataset. Then, based on the cephalometric landmarks, the CT image is aligned in the coronal and transverse planes. The image is then duplicated and two datasets of 441 images, per image, of 0.5° resolution is created from −5° to 5°. These sets of images are presented to a 3D rotate invariant CNN. After CNN determines the rotation degree of these images in the coronal and transverse planes, the skull midpoints are calculated using 3D geometric moments. Finally, by joining the midpoints and grades described by the CNN, the midplanes can be constructed.

2. Materials and Methods

2.1. Data Processing

The dataset used to validate the proposed method was acquired from the qure.ai CQ500 dataset [9]. From this dataset, 195 images with 512 × 512 and varied depths were selected for training, validation, and test purposes. To create the ground truth labels, CT images were imported into Mimics Medical Imaging Software (Materialise, Leuven, Belgium). First, individual thresholding with manual corrections was applied for each of the 3D volumetric CT images. Then, region growing was applied to create the 3D model mesh. This process allowed for the creation of the standard tessellation language (STL) file format which was converted into a matrix using voxelization method [10] so we can easily process the file in MATLAB R2019b software (Mathworks, USA) (Figure 2).

2.2. CNN Architecture and Implementation Details

2.2.1. CNN Framework for Biomedical Image Segmentation

The framework chosen in this paper for biomedical image segmentation was the U-Net [11]. U-Net has been used in a number of biomedical image segmentation applications such as kidney segmentation [12], prostate and prostate zones segmentation [13], brain tumor segmentation [14], brain segmentation [15], and so forth. Its name emerged from the idea of a U-shape architecture where in the first step, downsampling path, the spatial information is reduced while feature information is increased. In the next step, upsampling path, contracting path concatenate high-resolution features with spatial information and features. The result is a CNN that can work with few training samples and has the possibility to apply large images. We adopted a 3D U-net modified version of the code [16] initially implemented for brain tumor segmentation in MRI. The parameters adopted in this work are presented in Table 1. These parameters were chosen to avoid computational crash and error, while obtaining a good accuracy for the training set explored in this work.

2.2.2. CNN Framework for Rotation Invariant

By nature, CNNs are not rotation invariant, however, with a combination of convolutional, max pooling, average pooling, relu, and fully connected layers, the CNN framework can be transformed into rotation invariant. A number of papers have exploited the rotation invariant [17,18,19], however, the adopted framework presented in Table 2 worked very well in the dataset proposed using the Adam optimizer and a mini-batch size of 128.
Both CNN models were performed on an Intel i7-9700 (3.00 GHz) computer with 64 gigabyte (GB) of ram memory, and two 8GB Video RAM graphics processing units (GPUs) from NVIDIA (one RTX 2070 SUPER and one RTX 2080). The source code was implemented and tested in MATLAB R2019b.

2.3. Model Performance Evaluation and Statistical Analysis

For biomedical segmentation evaluation, the Dice Similarity Coefficient (DSC) [20] is the most popular metric to evaluate segmentation models. The DSC is a statistical method to gauge the similarity between two sample sets. In biomedical segmentation, the DSC measures the overlap between the ground truth and the predicted segmentation where 0 represents no overlap and 1 indicates complete overlap. Equation (1) defines the DSC, where the area of overlap is divided by the total pixels combined (TP—true positives, FP—false positives, and FN—false negatives).
DSC   =   Area   of   Overlap Total   Pixels   Combined =   2 TP 2 TP   +   FP   +   FN .
Symmetric Volume Difference (SVD) [21,22] is the corresponding error metric.
SVD = 1 − DSC.
Hausdorff Distance (HD) is a size-based method that describes the maximum distances between the boundaries of the segmented regions and the ground truth. This metric can be defined as:
HD   =   max h S ,   GT ,   h GT ,   S ,
where h S , GT   =   max a S min b GT | | a b | | [23].
To evaluate the performances of the CNN framework for rotation invariant, we implemented the following measures: average difference (AD), image quality index (IQI), Laplacian mean square error (LMSE), maximum difference (MD), mean-squared error (MSE), normalized absolute error (NAE), normalized cross-correlation (NK), structural content (SC), and structural similarity index (SSIM). To evaluate the classification, a sensitivity analysis was performed by using positive predictive value (PPV) and negative predictive value (NPV) defined as:
Sensitivity = TP TP + FN ,
Specificity = TN TN + FP ,
PPV = TP TP + FP ,
NPV = TN TN + FN .

2.4. 3D Geometric Moments

To avoid manual intervention, 3D geometric moments is applied for the automatic extraction of the central point. Three-dimensional p + q + r th order moments of a digitally sampled 3D image that has the gray function f x , y , z [24] is given as:
M pqr = x y z x p y q z r f x ,   y ,   z ,
where p , q , r = 0 ,   1 ,   2 ,   3 ,   . As described in [25], the mass and area of the zeroth order moment, M 000 , of a digital image is defined as:
M 000 = x y z f x ,   y ,   z ,
The center of mass of the image f x , y , z is represented by the two first moments:
M 100 = x y z xf x ,   y ,   z ,
M 010 = x y z yf x ,   y ,   z ,
M 001 = x y z zf x ,   y ,   z ,
Thus, the centroid of an image can be calculated by:
  X ¯ = M 100 M 000 ,   Y ¯ = M 010 M 000 ,   and   Z ¯ = M 001 M 000 .
As best practice, the center of mass was chosen to represent the position of an image in the field of view. The centroid of the image f x , y , z , given by Equation (13), can be used to describe the position of the image in space by using the point as a reference point

3. Results and Discussion

3.1. Skull Segmentation

The CT volumetric dataset and 3D mesh models were presented to the 3D U-Net with the parameters described by Table 1. From the 195 images, 190 were used for training and 5 for validation/testing. Table 3 shows the DSCs, SVDs, and HDs, in terms of mean ± standard deviation (SD) after being trained and tested 10 times, acquired from the testing set. When using two GPUs as specified, the CNN took 57 min in 15 epochs to converge.
These results are close to those DSCs reported by [26] (mean DSC of 0.92), and slightly lower than the results reported by [27] (mean DSC of 0.98). Regarding HD, its discrepant values may be directly related to segmentation errors due to bright artifacts found in the original image, which may be caused by dental filling and components of the CT scan machine, as shown in Figure 3. However, any necessary modification was performed by manual corrections after the segmentation. These predicted labels play an important role in the coronal and transverse alignment.

3.2. Transverse and Coronal Angles

To identify the transverse and coronal angles through the 3D CNN, 101 volumetric images, from the 195 segmented images, were selected to create a database. The first step was to identify cephalometric landmarks to help align the predicted labels Figure 3a. In the coronal plane, we selected the 1-crista galli, 2-frontozygomatic suture, and 3-orbitale while in the sagittal plane, we selected 4-lambda and 5-opistocranion. To identify these points and make the necessary alignment, two grids were generated as a reference in the transverse and coronal planes. A number of slices were verified, and the necessary adjustments were performed (Figure 4b,c).
After the alignment, for each of the 101 images, a set of 441 images with inclination angles from −5° to 5°, with 0.5° increments, along the coronal and transverse planes was created. In total, 44541 images were created and were divided into 21 labels. These labels represented the 0.5° of variation in the coronal and transverse planes from −5° to 5°, as shown in Figure 5.
To reduce the computational and processing time, these images were reduced to 128 × 128 using the nearest neighbor interpolation method. Additionally, the volumetric image was divided into four rectangular sub-cubes and only one-quarter of the whole image space was used to predict the angles as shown in Figure 6. As this step aims to identify the coronal and transverse angles in symmetrical skulls, these steps do not affect the output image. Figure 7 summarizes this process.
In this phase, we used the 3D rotation invariant CNN introduced in Section 2.2.2 and detailed in Table 2. To optimize the CNN convergence, computational time, and accuracy, two identical datasets were created using the 44,541 images and represented by 21 labels. Thus, 21 labels represent the rotation in the transverse plane and 21 labels in the coronal 201 plane. 90% of the dataset was used for training, and 10% for validation/testing. Training and testing were performed 10 times. Table 4 shows the analytical performance for these two. It took 30 min in five epochs for the transverse and 70 min in 12 epochs for the coronal CNN to converge using two GPUs.
As seen in Table 4, we can see that CNN performed well and can represent a rotation invariant image descriptor for these scenarios. In fact, the accuracy reached ≈99% with simple hyperparameter, which allows the construction of a simple 3D CNN. After the discovery of the transverse and coronal deviation angles, the original image was then rotated accordingly with these two found angles.

3.3. Geometric Moments Image Center

Finally, to calculate the center of the volumetric images, Equations (8) to (13) were used. As there are no patterns to validate the accuracy of the center-point, visual evaluation was used and compared with cephalometric landmarks. Figure 8 and Figure 9 show the cross-sectional plane created from the geometric moments. Figure 9a presents the perspective view and Figure 9b shows the front view of one aligned sample with measured dimensions of frontozygomatic suture and orbitale in the sagittal plane displayed in Materialise MiniMagic software. We used [28] to convert from voxel into STL file format.

3.4. Deformed Skull Test

To validate this method in deformed images, eight defected CT images were used from two different datasets [29,30,31,32] found in the cancer imaging archive data collections (TCIA) [33]. In the first step, STL files were generated by the 3D U-Net using the same parameters as the Table 1. DSC, SVD, and HD results are presented in Table 5 and a sample is shown in Figure 10.
The CQ500 database does not contain deformed images which may have caused the discrepancy in the DSC results. Unfortunately, there are no databases of deformed skulls for analysis. Furthermore, the 195 images used for training were not enough to improve the 3D U-Net prediction. Reference [27] reported a 6% mean improvement compared to [26] and associated the improvement with the size of the image dataset for training purposes. Moreover, in the sampled image (Figure 10), part of the vertebral column and small segmented parts that do not belong to the skull were segmented by mistake, which generated the worst DSC value and high HD values. This is likely associated with the C1 vertebrae labeled during the creation of the ground truth and small bright artifacts on the outside of the skull in the CT scans. Even though it seems like an error, this can be disregarded since Figure 10 shows that the ground truth and predicted label to be very close.
Finally, using the eight models predicted, transverse and coronal angles were calculated, and the center of the image was acquired using 3D geometric moments. Figure 11 shows the results for the deformed images.
The proposed method shows good results in obtaining the bilateral symmetric midplane of regular/symmetric images. However, for deformed images, it failed to identify the rotation for some images along with the image center. These errors are likely due to some factors:
  • the small database size, which is already reported in [27];
  • to the best of our knowledge, there are no deformed CT database available which restricts the possibility to train the system with deformed images;
  • during the ground truth segmentation process and voxelization, a few regions of interest (ROIs) may have not been incorporated in the 3D model. The first may be caused by the manual selection of the ROI, performed by an expert, which leads to the CNN generating the defects. Secondly, a quantity of information from the skull voxel may be lost due to the smoothing of the edges and noise residuals removal performed in the segmentation process;
  • regarding the center of the 3D images, as reported by [8], when the image suffers from a lack of symmetry, non-uniform brightness, deformation, interference, or incompleteness, the calculation of the image center using geometric moments becomes complex and finds some restrictions as this technique is a quantitative measure of an image’s function or structure.
It is possible to state that the proposed method obtained good results from symmetric CT images datasets. However, for deformed images, an improvement is necessary to achieve better results. For this purpose, an increase in the CT database size may be performed as well as the inclusion of non-uniform and deformed CT images which may also lead to an improvement in the detection of the transverse and coronal angles.
Finally, a modification in the method to identify the center of the image may be carried out by creating a 3D U-net to segment the nasal bone instead of the presented geometric moments technique. This method will allow for the definition of the center of the image using the center of the nasal bone.

4. Conclusions

This study aimed to introduce a 3D automatic technique for determining the craniofacial symmetry midplane from CT scans using the convolutional neural network and gemetric moments. A total of 195 symmetric CT images were used to evaluate this method using the CQ500 database while eight asymmetric CT images from TCIA database were used to evaluate the performance in asymmetric images. For symmetric images, this method obtained results close to 99%. However, for asymmetric images, the method needs further development to improve its results. CNNs offer an effective alternative to the pseudo-zernike moments method and conventional landmark-based symmetry scores that depend on the general positions of cephalometric landmarks. CNNs are also an alternative to PCA-ICP techniques, which depend on the manual selection of the central point which cannot be improved. With the proposed technique, the central point could be found as the centroid of an image, and then the symmetrical midplane can be constructed for symmetric images. In this study, we have shown the proposed technique to be reliable and to provide the midplane symmetry plane with great accuracy in symmetric images. This method can be used as a tool to aid surgeons in reconstructive craniofacial surgeries.

Author Contributions

Conceptualization, R.D.C.d.S.; Project administration, T.R.J.; Writing—review and editing, V.A.C. All authors have read and agreed to the published version of the manuscript.

Funding

No external funding was used for this study.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets presented in this study are openly available at: CQ500- http://headctstudy.qure.ai/dataset (accessed on 1 June 2020), and TCIA- https://www.cancerimagingarchive.net/ (accessed on 1 June 2020).

Acknowledgments

Part of the data used in this publication were created by the National Cancer Institute Clinical Proteomic Tumor Analysis Consortium (CPTAC). Part of the results shown in this work are in part based upon data created by the TCGA Research Network.

Conflicts of Interest

The authors declare no conflict of interest.
Disclosure Statement: This study does not contain any experiment with human participants performed by any of the authors.

Abbreviations

The following abbreviations are used in this manuscript:
ADaverage difference
CNNconvolutional neural network
CTcomputerized tomography
DSCdice similarity coefficient
FNfalse negative
FPfalse positive
GBgigabyte
GPUsgraphics processing units
HDhausdorff distance
ICPiterative closest point
IQIimage quality index
LMSElaplacian mean square error
MDmaximum difference
MRImagnetic resonance imaging
MSEmean-square error
MSPmidsagittal plane
NAEnormalized absolute error
NKnormalized cross-correlation
NPVnegative predictive value
PCAprincipal component analysis
PPVpositive predictive value
ROIsregions of interest
SCstructural content
SDstandard deviation
SSIMstructural similarity index
STLstandard tessellation language
SVDsymmetric volume difference
TPtrue positive

References

  1. Damstra, J.; Fourie, Z.; De Wit, M.; Ren, Y. A three-dimensional comparison of a morphometric and conventional cephalometric midsagittal planes for craniofacial asymmetry. Clin. Oral Investig. 2011, 16, 285–294. [Google Scholar] [CrossRef] [Green Version]
  2. Kim, T.-Y.; Baik, J.-S.; Park, J.-Y.; Chae, H.-S.; Huh, K.-H.; Choi, S.-C. Determination of midsagittal plane for evaluation of facial asymmetry using three-dimensional computed tomography. Imaging Sci. Dent. 2011, 41, 79–84. [Google Scholar] [CrossRef] [Green Version]
  3. Willing, R.; Roumeliotis, G.; Jenkyn, T.; Yazdani, A. Development and evaluation of a semi-automatic technique for determining the bilateral symmetry plane of the facial skeleton. Med. Eng. Phys. 2013, 35, 1843–1849. [Google Scholar] [CrossRef]
  4. Roumeliotis, G.; Willing, R.; Neuert, M.; Ahluwali, R.; Jenkyn, T.; Yazdani, A. Application of a novel semi-automatic technique for determining the bilateral symmetry plane of the facial skeleton of normal adult males. J. Craniofacial Surg. 2015, 26, 1997–2001. [Google Scholar] [CrossRef]
  5. Di Angelo, L.; Stefano, P.D.; Governi, L.; Marzola, A.; Volpe, Y. A robust and automatic method for the best symmetry plane detection of craniofacial skeletons. Symmetry 2019, 11, 245. [Google Scholar] [CrossRef]
  6. Di Angelo, L.; Di Stefano, P. A computational method for bilateral symmetry recognition in asymmetrically scanned human faces. Comput. Aided Des. Appl. 2014, 11, 275–283. [Google Scholar] [CrossRef]
  7. Noori, S.M.R.; Farnia, P.; Bayat, M.; Bahrami, N.; Shakourirad, A.; Ahmadian, A. Automatic detection of symmetry plane for computer-aided surgical simulation in craniomaxillofacial surgery. Phys. Eng. Sci. Med. 2020, 43, 1087–1099. [Google Scholar] [CrossRef]
  8. Dalvit Carvalho da Silva, R.; Jenkyn, T.R.; Carranza, V.A. Application of a novel automatic method for determining the bilateral symmetry midline of the facial skeleton based on invariant moments. Symmetry 2020, 12, 1448. [Google Scholar] [CrossRef]
  9. Chilamkurthy, S.; Ghosh, R.; Tanamala, S.; Biviji, M.; Campeau, N.G.; Venugopal, V.K.; Mahajan, V.; Rao, P.; Warier, P. Deep learning algorithms for detection of critical findings in head CT scans: A retrospective study. Lancet 2018, 392, 2388–2396. [Google Scholar] [CrossRef]
  10. Patil, S.; Ravi, B. Voxel-based representation, display and thickness analysis of intricate shapes. In Proceedings of the Ninth International Conference on Computer Aided Design and Computer Graphics (CAD-CG’05), Hong Kong, China, 7–10 December 2005; p. 6. [Google Scholar] [CrossRef]
  11. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  12. Kim, T.; Lee, K.; Ham, S.; Park, B.; Lee, S.; Hong, D.; Kim, G.B.; Kyung, Y.S.; Kim, C.-S.; Kim, N. Active learning for accuracy enhancement of semantic segmentation with CNN-corrected label curations: Evaluation on kidney segmentation in abdominal CT. Sci. Rep. 2020, 10, 1–7. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Aldoj, N.; Biavati, F.; Michallek, F.; Stober, S.; Dewey, M. Automatic prostate and prostate zones segmentation of magnetic resonance images using DenseNet-like U-net. Sci. Rep. 2020, 10, 1–17. [Google Scholar] [CrossRef]
  14. Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic brain tumor detection and segmentation using U-Net based fully convolutional networks. In Medical Image Understanding and Analysis; Valdés Hernández, M., González-Castro, V., Eds.; Communications in Computer and Information Science; Springer: Cham, Switzerland, 2017; Volume 723. [Google Scholar] [CrossRef] [Green Version]
  15. Lee, B.; Yamanakkanavar, N.; Choi, J.Y. Automatic segmentation of brain MRI using a novel patch-wise U-net deep architecture. PLoS ONE 2020, 15, e0236493. [Google Scholar] [CrossRef]
  16. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. arXiv 2016, arXiv:1606.06650. [Google Scholar]
  17. Chen, Y.; Lyu, Z.X.; Kang, X.; Wang, Z.J. A rotation-invariant convolutional neural network for image enhancement forensics. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 2111–2115. [Google Scholar]
  18. Chidester, B.; Zhou, T.; Do, M.N.; Ma, J. Rotation equivariant and invariant neural networks for microscopy image analysis. Bioinformatics 2019, 35, i530–i537. [Google Scholar] [CrossRef] [PubMed]
  19. Kim, J.; Jung, W.; Kim, H.; Lee, J. CyCNN: A rotation invariant CNN using polar mapping and cylindrical convolutional layers. arXiv 2020, arXiv:2007.10588. [Google Scholar]
  20. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  21. Schenk, A.; Prause, G.; Peitgen, H.-O. Efficient semiautomatic segmentation of 3D objects in medical images. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2000; Delp, S.L., DiGoia, A.M., Jaramaz, B., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2000; Volume 1935, pp. 186–195. [Google Scholar]
  22. Yeghiazaryan, V.; Voiculescu, I. Family of boundary overlap metrics for the evaluation of medical image segmentation. J. Med. Imaging 2018, 5, 1. [Google Scholar] [CrossRef]
  23. Vania, M.; Mureja, D.; Lee, D. Automatic spine segmentation from CT images using Convolutional Neural Network via redundant generation of class labels. J. Comput. Des. Eng. 2019, 6, 224–232. [Google Scholar] [CrossRef]
  24. Mercimek, M.; Gulez, K.; Mumcu, T.V. Real object recognition using moment invariants. Sadhana 2005, 30, 765–775. [Google Scholar] [CrossRef] [Green Version]
  25. Xu, D.; Li, H. Geometric moment invariants. Pattern Recognit. 2008, 41, 240–249. [Google Scholar] [CrossRef]
  26. Minnema, J.; Van Eijnatten, M.; Kouw, W.M.; Diblen, F.; Mendrik, A.; Wolff, J. CT image segmentation of bone for medical additive manufacturing using a convolutional neural network. Comput. Biol. Med. 2018, 103, 130–139. [Google Scholar] [CrossRef] [Green Version]
  27. Kodym, O.; Španěl, M.; Herout, A. Segmentation of defective skulls from CT data for tissue modelling. arXiv 2019, arXiv:1911.08805. [Google Scholar]
  28. Adam, A. Converting a 3D logical array into an STL surface mesh. Available online: https://www.mathworks.com/matlabcentral/fileexchange/27733-converting-a-3d-logical-array-into-an-stl-surface-mesh (accessed on 1 November 2020).
  29. Kinahan, P.; Muzi, M.; Bialecki, B.; Coombs, L. Data from ACRIN-FMISO brain. Cancer Imaging Arch. 2018. [Google Scholar] [CrossRef]
  30. Gerstner, E.R.; Zhang, Z.; Fink, J.R.; Muzi, M.; Hanna, L.; Greco, E.; Prah, M.; Schmainda, K.M.; Mintz, A.; Kostakoglu, L.; et al. ACRIN 6684: Assessment of tumor hypoxia in newly diagnosed glioblastoma using 18F-FMISO PET and MRI. Clin. Cancer Res. 2016, 22, 5079–5086. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Ratai, E.-M.; Zhang, Z.; Fink, J.; Muzi, M.; Hanna, L.; Greco, E.; Richards, T.; Kim, D.; Andronesi, O.C.; Mintz, A.; et al. ACRIN 6684: Multicenter, phase II assessment of tumor hypoxia in newly diagnosed glioblastoma using magnetic resonance spectroscopy. PLoS ONE 2018, 13, e0198548. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Zuley, M.L.; Jarosz, R.; Kirk, S.; Lee, Y.; Colen, R.; Garcia, K.; Aredes, N.D. Radiology data from the cancer genome atlas head-neck squamous cell carcinoma [tcga-hnsc] collection. Cancer Imaging Arch. 2016. [Google Scholar] [CrossRef]
  33. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The Cancer Imaging Archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Overview of the proposed method. CNN: convolutional neural network.
Figure 1. Overview of the proposed method. CNN: convolutional neural network.
Biology 10 00182 g001
Figure 2. (a) Computed tomography (CT) scan, (b) thresholding applied, (c) region growing, and (d) 3D mesh model.
Figure 2. (a) Computed tomography (CT) scan, (b) thresholding applied, (c) region growing, and (d) 3D mesh model.
Biology 10 00182 g002
Figure 3. Errors in the predicted images caused by machine components and dental fillings.
Figure 3. Errors in the predicted images caused by machine components and dental fillings.
Biology 10 00182 g003
Figure 4. (a) Identification of cephalometric landmarks in coronal (top) and sagittal plane (bottom), (b) CT slices misaligned, and (c) CT slices after the alignment procedure (coronal plane-top and transverse plane-bottom).
Figure 4. (a) Identification of cephalometric landmarks in coronal (top) and sagittal plane (bottom), (b) CT slices misaligned, and (c) CT slices after the alignment procedure (coronal plane-top and transverse plane-bottom).
Biology 10 00182 g004
Figure 5. Examples of images rotated in the coronal and transverse planes where parentheses represents (coronal, sagittal), respectively.
Figure 5. Examples of images rotated in the coronal and transverse planes where parentheses represents (coronal, sagittal), respectively.
Biology 10 00182 g005
Figure 6. Volumetric image divided into rectangular subcubes and the selected one-quarter subcube.
Figure 6. Volumetric image divided into rectangular subcubes and the selected one-quarter subcube.
Biology 10 00182 g006
Figure 7. Flowchart representing the process of transversal and coronal alignment and database creation.
Figure 7. Flowchart representing the process of transversal and coronal alignment and database creation.
Biology 10 00182 g007
Figure 8. 12 samples with their respective cross-sectional plane created using geometric moments.
Figure 8. 12 samples with their respective cross-sectional plane created using geometric moments.
Biology 10 00182 g008
Figure 9. (a) Perspective and (b) front view.
Figure 9. (a) Perspective and (b) front view.
Biology 10 00182 g009
Figure 10. (a) Ground truth and (b) predicted.
Figure 10. (a) Ground truth and (b) predicted.
Biology 10 00182 g010
Figure 11. Results of the implemented method in deformed/asymmetric images.
Figure 11. Results of the implemented method in deformed/asymmetric images.
Biology 10 00182 g011
Table 1. Skull CT segmentation implementation details.
Table 1. Skull CT segmentation implementation details.
ParameterValue
OptimizerAdam
Encoder Depth4
Filter Size5
Number of First Encoder Filter6
Patch Per Image4
Min Batch Size128
Initial Learning Rate10−2
Table 2. Rotation-invariant CNN framework adopted.
Table 2. Rotation-invariant CNN framework adopted.
LayersSizeNumber FilterStride
3D Conv1 × 131
1BN + Relu---
3D Max pooling5 × 5-2
3D Conv5 × 581
2BN + Relu---
3D Max pooling5 × 5-2
3D Conv7 × 7161
3BN + Relu---
3D Max pooling3 × 3-2
3D Conv5 × 5321
4BN + Relu---
3D Max pooling2 × 2-2
3D Conv5 × 5641
5BN + Relu---
3D Average pooling2 × 2-2
63D Conv1 × 11281
BN + Relu---
7FC (25 neurons)---
Relu---
8FC (50 neurons)---
Relu---
9FC (labels neurons)---
Softmax---
Table 3. Skull and Background DSCs, SVDs, and HDs Values of 5 samples and their mean. DSC: Dice Similarity Coefficient; SVD: Symmetric Volume Difference; HD: Hausdorff Distance.
Table 3. Skull and Background DSCs, SVDs, and HDs Values of 5 samples and their mean. DSC: Dice Similarity Coefficient; SVD: Symmetric Volume Difference; HD: Hausdorff Distance.
DSC-SkullDSC-BackgroundSVD-SkullHD-Skull
0.8993 ± 0.0040.9927 ± 0.00030.1007 ± 0.00467.7 ± 09.20
0.9093 ± 0.0080.9948 ± 0.00050.0907 ± 0.00827.81 ± 31.19
0.9150 ± 0.0080.9941 ± 0.00060.0850 ± 0.00838.78 ± 20.26
0.9349 ± 0.0080.9958 ± 0.00050.0651 ± 0.00849.92 ± 37.69
0.9362 ± 0.0060.9953 ± 0.00040.8844 ± 0.00639.99 ± 44.40
0.9189 ± 0.0160.9945 ± 0.00120.0811 ± 0.01644.73 ± 14.79
Table 4. Statistical analysis of the coronal and transverse CNNs. PPV: positive predictive value; NPV: negative predictive value; AD: average difference; IQI: image quality index; LMSE: Laplacian mean square error; MD: maximum difference; MSE: mean-squared error; NAE: normalized absolute error; NK: normalized cross-correlation; SC: structural content; SSIM: structural similarity index.
Table 4. Statistical analysis of the coronal and transverse CNNs. PPV: positive predictive value; NPV: negative predictive value; AD: average difference; IQI: image quality index; LMSE: Laplacian mean square error; MD: maximum difference; MSE: mean-squared error; NAE: normalized absolute error; NK: normalized cross-correlation; SC: structural content; SSIM: structural similarity index.
IndexIdeal ValueCoronal CNN ValueTransverse CNN Value
Accuracy10.9909 ± 0.00380.9947 ± 0.0034
Sensitivity10.9811 ± 0.01700.9969 ± 0.0054
Specificity10.9982 ± 0.00120.9994 ± 0.0005
PPV10.9646 ± 0.02360.9892 ± 0.0106
NPV10.991 ± 0.00090.9998 ± 0.0003
AD00.0124 ± 0.00850.0004 ± 0.0095
IQI10.9979 ± 0.00140.9994 ± 0.0004
LMSE00.9926 ± 0.67170.9571 ± 0.0814
MD011.667 ± 1.15479.3333 ± 5.6862
MSE00.2710 ± 0.18830.0723 ± 0.0561
NAE00.0030 ± 0.00180.0012 ± 0.0007
NK10.9985 ± 0.00100.9998 ± 0.0004
SC10.9982 ± 0.00081.0000 ± 0.0008
SSIM10.9285 ± 0.03130.9523 ± 0.0290
Table 5. Skull and background DSCs, SVDs, and HDs values of the 8 defected skulls and their mean.
Table 5. Skull and background DSCs, SVDs, and HDs values of the 8 defected skulls and their mean.
DSC-SkullDSC-BackgroundSVD-SkullHD-Skull
0.8206 ± 0.0800.9902 ± 0.00030.1794 ± 0.08048.56 ± 03.69
0.8114 ± 0.0050.9831 ± 0.00110.1886 ± 0.00549.71 ± 06.21
0.8294 ± 0.0120.9875 ± 0.00070.1706 ± 0.01247.57 ± 08.61
0.8626 ± 0.0160.9890 ± 0.00080.1374 ± 0.01646.60 ± 07.76
0.8974 ± 0.0110.9946 ± 0.00040.1026 ± 0.01149.20 ± 10.86
0.8302 ± 0.0110.9817 ± 0.00100.1698 ± 0.01156.17 ± 07.55
0.7888 ± 0.0100.9898 ± 0.00040.2112 ± 0.01050.81 ± 09.57
0.8361 ± 0.0140.9891 ± 0.00120.1639 ± 0.01434.27 ± 14.06
0.8346 ± 0.0330.9881 ± 0.00410.1654 ± 0.03347.86 ± 06.21
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dalvit Carvalho da Silva, R.; Jenkyn, T.R.; Carranza, V.A. Convolutional Neural Networks and Geometric Moments to Identify the Bilateral Symmetric Midplane in Facial Skeletons from CT Scans. Biology 2021, 10, 182. https://doi.org/10.3390/biology10030182

AMA Style

Dalvit Carvalho da Silva R, Jenkyn TR, Carranza VA. Convolutional Neural Networks and Geometric Moments to Identify the Bilateral Symmetric Midplane in Facial Skeletons from CT Scans. Biology. 2021; 10(3):182. https://doi.org/10.3390/biology10030182

Chicago/Turabian Style

Dalvit Carvalho da Silva, Rodrigo, Thomas Richard Jenkyn, and Victor Alexander Carranza. 2021. "Convolutional Neural Networks and Geometric Moments to Identify the Bilateral Symmetric Midplane in Facial Skeletons from CT Scans" Biology 10, no. 3: 182. https://doi.org/10.3390/biology10030182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop