Next Article in Journal
Construction of Chitosan-Modified Naphthalimide Fluorescence Probe for Selective Detection of Cu2+
Previous Article in Journal
LidPose: Real-Time 3D Human Pose Estimation in Sparse Lidar Point Clouds with Non-Repetitive Circular Scanning Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cervical Spondylosis Diagnosis Based on Convolutional Neural Network with X-ray Images

1
Department of Medical Imaging, China Rehabilitation Research Center and Capital Medical University School of Rehabilitation Medicine, Beijing 100068, China
2
Department of Electronics Design, Mid Sweden University, 85170 Sundsvall, Sweden
3
Department of Spinal and Neural Function Reconstruction, China Rehabilitation Research Center and Capital Medical University School of Rehabilitation Medicine, Beijing 100068, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3428; https://doi.org/10.3390/s24113428
Submission received: 24 April 2024 / Revised: 18 May 2024 / Accepted: 23 May 2024 / Published: 26 May 2024
(This article belongs to the Special Issue Anomaly Detection and Fault Diagnosis in Sensor Networks)

Abstract

:
The increase in Cervical Spondylosis cases and the expansion of the affected demographic to younger patients have escalated the demand for X-ray screening. Challenges include variability in imaging technology, differences in equipment specifications, and the diverse experience levels of clinicians, which collectively hinder diagnostic accuracy. In response, a deep learning approach utilizing a ResNet-34 convolutional neural network has been developed. This model, trained on a comprehensive dataset of 1235 cervical spine X-ray images representing a wide range of projection angles, aims to mitigate these issues by providing a robust tool for diagnosis. Validation of the model was performed on an independent set of 136 X-ray images, also varied in projection angles, to ensure its efficacy across diverse clinical scenarios. The model achieved a classification accuracy of 89.7%, significantly outperforming the traditional manual diagnostic approach, which has an accuracy of 68.3%. This advancement demonstrates the viability of deep learning models to not only complement but enhance the diagnostic capabilities of clinicians in identifying Cervical Spondylosis, offering a promising avenue for improving diagnostic accuracy and efficiency in clinical settings.

1. Introduction

Cervical Spondylosis (CS) is a chronic degenerative condition of the cervical spine, manifesting through various clinical symptoms such as neck pain, stiffness, radiating pain, headaches, and the potential onset of neurological dysfunction [1]. Significantly, neck pain was identified as the fourth leading cause of disability-adjusted life years (DALYs) in 2015, highlighting the global impact of CS [2]. Epidemiologically, the prevalence of CS varies by age, with rates of 12.4% in individuals aged 18–29, escalating to 100% in those over 70 years, indicating a widespread and age-related increase in incidence. Early stages of CS might not disrupt daily life significantly, but as the condition progresses, symptoms worsen, leading to increased treatment complexity and healthcare costs, thereby underscoring the importance of early detection [3].
The diagnostic process for CS primarily employs imaging modalities such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI), and X-ray [4], with X-ray screening being the preferred method due to its accessibility, low radiation exposure, and cost-effectiveness [5]. Despite its advantages, X-ray diagnosis of CS is hindered by its lower accuracy and the reliance on the subjective judgment and experience of the interpreting physician [6]. This situation calls for improved diagnostic methodologies to overcome these limitations [7].
Recent advancements in artificial intelligence (AI), particularly in machine learning (ML) and deep learning (DL), have made significant strides in medical diagnostics, including the field of computer-aided diagnosis (CAD) [8,9,10]. These technologies have demonstrated exceptional utility in enhancing X-ray-based CS classification, through their ability to autonomously identify and analyze CS-relevant features within X-ray images. The strengths of ML and DL lie in their capacity to manage and analyze large datasets, thereby improving the performance and generalizability of diagnostic algorithms. By minimizing the influence of human subjectivity and fatigue, these technologies ensure higher diagnostic precision and consistency. DL models, in particular, offer rapid analysis capabilities, essential in urgent care settings, and evolve over time through continuous training, further refining their accuracy. Consequently, ML and DL represent invaluable tools for healthcare professionals, providing critical support in achieving more accurate diagnoses and treatment decisions.
This research employs a ResNet-34 convolutional neural network (CNN) model to evaluate the presence of CS in patients, utilizing cervical spine X-ray imaging features. Notably, this study distinguishes itself by being the first to leverage multi-angle, multi-label data, enhancing the model’s accuracy significantly. The results confirm that the proposed deep learning neural network model achieves a diagnostic accuracy rate that is 21.4% higher than conventional manual diagnostic approaches [11], thereby offering a substantial improvement in the accuracy of CS diagnosis and potentially reducing the clinical workload.
In this work, our main contributions can be summarized as the following:
  • Improved Patient Outcomes: The enhanced accuracy of the ResNet-34 CNN model, which is 21.4% higher than manual methods, means that it can more accurately diagnose conditions from cervical spine X-ray images. This leads to more accurate treatments being prescribed, improving patient outcomes.
  • Efficiency in Diagnosis: The increased diagnostic accuracy reduces the reliance on manual methods, which can be time-consuming and prone to human error. This not only speeds up the diagnostic process but also reduces the workload of clinical staff, allowing them to focus on other critical tasks.
  • Leveraging Multi-angle, Multi-label Data: This is the first model to leverage multi-angle, multi-label data, which provides a more comprehensive view of the patient’s condition. This can lead to the discovery of conditions that might be missed using traditional methods, further improving the accuracy of the diagnosis.

2. Proposed Method

2.1. Materials

This retrospective study meticulously compiled data from 595 patients treated at a rehabilitation research center from January 2021 to November 2022, resulting in a collection of 1371 anonymized images in line with established research protocols. The cohort consisted of 566 males and 805 females, with an average age of 46.4 ± 16.4 years, ensuring all selected medical images met the required diagnostic standards.

2.1.1. Inclusion Criteria

For a case to be included in the study, patients had to meet the radiological diagnostic criteria for Cervical Spondylosis (CS). These criteria encompass a range of degenerative changes indicative of CS, including [12,13] degeneration observed in the intervertebral facet joints and vertebral costal joints, intervertebral disc degeneration, ligament degeneration, uncinate joint degeneration, osteoporosis, and scoliosis. Such comprehensive criteria ensure a robust identification of positive CS cases based on radiological evidence.

2.1.2. Exclusion Criteria

The study excluded any patient showing degenerative changes in the cervical spine that could be attributed to congenital malformations, spinal cord injuries, tuberculosis, tumors, ankylosing spondylitis, or previous surgical interventions. Furthermore, X-ray images that did not meet the quality criteria due to artifacts, overexposure, underexposure, or any factor that could affect the clarity and accuracy of cervical spine imaging diagnosis were also excluded [14].

2.2. Experimental Design

This study strategically divided the 1371 collected image data into two sets: a test set (136 images) and a training-validation set (1235 images), following a 9:1 ratio. Each image was manually labeled by expert radiologists, noting the cervical spine X-ray projection position and CS classification outcomes.
Within the training-validation dataset, 80% (988 images) was randomly selected for the training of the ResNet-34 convolutional neural network (CNN) model [15], while the remaining 20% (247 images) constituted the validation set, which was strictly used for performance evaluation without influencing the training phase. The selection of the model for final experimentation was determined by the highest F1-score achieved on this validation set (Figure 1).
The performance of the ResNet-34 CNN model in accurately classifying cervical spine X-ray projection positions and diagnosing CS was evaluated using the test set. Detailed methodologies and the results of this evaluation are elaborated upon in Section 3, accompanied by illustrative Figure 2 that maps out the process.

2.2.1. Data Preprocessing

To prepare the X-ray images for CS research, each image underwent anonymization using the freely available RadiAnt DICOM Viewer 2020.2.3 software, effectively eliminating any patient identifiers. Data augmentation prepossessing includes randomly cropping the image to a size of 224 × 224 pixels, applying random flips, random rotations, and normalization.

2.2.2. Data Labeling

For the purpose of image diagnosis, two expert radiologists, with eight and twenty-five years of specialized experience, respectively, independently assessed each patient’s set of images. These sets included anteroposterior, lateral, left oblique, and right oblique views, presented as separate, anonymized collections. The RadiAnt DICOM Viewer 2020.2.3 software facilitated the image evaluations. In instances of divergent assessments, the two experts engaged in discussions to reach a consensus. Cases without unanimous agreement were subsequently excluded from the analysis.
The process began with the initial labeling of the X-ray projection positions for each diagnostic image. Following this, CS was identified based on the distinct radiological diagnostic features observable in different positions. The labeled imaging characteristics for the various positions are outlined as follows:
Anteroposterior: (a) Degenerative changes in the intervertebral facet joints and vertebral costal joints, presenting as marginal osteophyte development, articular cartilage sclerosis, cystic degeneration, and narrowing of the joint space; (b) osteoporosis; (c) scoliosis.
Lateral: (a) Intervertebral disc degeneration, presenting as intervertebral space narrowing, the formation of Schmorl’s nodes, and the development of osteophytes along the vertebral body edges; (b) ligament degeneration, presenting as ligament thickening and calcification; (c) osteoporosis; (d) scoliosis, presenting as vertebral slippage and abnormal cervical curvature.
Left Oblique and Right Oblique: (a) Uncinate joint degeneration, presenting with osteophyte formation and intervertebral foramen narrowing; (b) osteoporosis.

2.2.3. Deep Learning Model

The CNN stands as a prominent subcategory within the realm of DL algorithms and serves as the primary neural network architecture harnessed in the field of medical imaging [16]. Among various CNN architectures, ResNet stands out as a cutting-edge model for image classification, renowned for its efficacy and precision [17,18]. The adoption of ResNet-34 in this study is underpinned by its prior training on the ImageNet dataset, which is notably extensive and diverse, providing a solid foundation for image classification tasks.
A key innovation of ResNet-34 is its incorporation of residual learning, namely, a novel approach that emphasizes learning the residual or difference between the input and output. This strategy effectively addresses the perennial issues of vanishing and exploding gradients that often hamper the performance of traditional neural networks [19]. By fostering direct information flow across layers through cross-layer connections, ResNet-34 ensures the preservation of input data integrity throughout the network. This attribute not only streamlines the model’s mapping process but also diminishes training complexity.
The architectural hallmark of ResNet-34 is its assembly of residual blocks, which collectively forge a deep CNN. Such depth empowers the network to capture complex high-level features and patterns within the data, facilitating nuanced image analysis. Additionally, the modular design of ResNet-34’s residual blocks enhances the model’s versatility, making it apt for a wide array of computer vision tasks beyond image classification, including object detection and semantic segmentation. This versatility not only augments model reusability but also supports the application of transfer learning techniques, thereby broadening the scope of ResNet-34’s utility in diverse settings [20].

2.3. Statistical Analysis

Data analysis was conducted using SPSS 26. Measurement data were represented as x ¯ ± s or M ( P 25 , P 75 ) . Counting data were presented as the number of cases and percentages (%).
Various evaluation metrics were employed to assess model performance. Accuracy was the primary metric used for evaluation, with additional criteria including sensitivity, precision, and F1-score (Table 1). The reliability and significance of these metrics were further scrutinized through χ 2 testing, with a P > 0.01 indicating the absence of statistically significant differences within the dataset. For the execution of statistical analyses, the Scikit-learn library was employed, ensuring robust and reliable analysis [21].
In essence, classifiers that exhibit superior performance across the metrics of accuracy, sensitivity, precision, and F1-score are deemed more efficacious in the context of performance evaluation, providing a comprehensive measure of the model’s diagnostic capability.

3. Results

The dataset for this study comprises 1371 X-ray images, including 566 male and 805 female participants, with a mean age of (46.4 ± 16.4) years. Statistical analyses to assess the comparability of the training-validation and test sets with the overall population sample were performed using the χ 2 test for categorical variables and the T test for continuous variables. These analyses covered parameters such as the number of images, gender distribution, age, and rates of positive diagnosis. The resulting P > 0.01, indicating the absence of statistically significant differences between the training-validation set, the test set, and the population sample. This demonstrates the external validity of the study findings; refer to Table 2.

3.1. Cervical Vertebra X-ray Projection Position Classification Results

Upon applying the neural network model to the test dataset, exemplary performance metrics were observed. The weighted average (WA), based on the proportions of images in anteroposterior (A), lateral (B), left oblique (L), and right oblique (R) positions, exhibited average accuracy, sensitivity, precision, and F1-score of 97.8%, 97.8%, 98.1%, and 97.8% respectively.
Detailed analysis of the four specific imaging positions revealed optimal scores: positions A, B, and L demonstrated perfect accuracy and sensitivity (100.0%), underscoring the model’s exceptional efficacy. Positions A, B, and R recorded the highest precision (100.0%), and positions A and B achieved the top F1-score (100.0%). These indicators were validated through χ 2 testing, confirming their statistical reliability as depicted in Figure 3 and Table 3).

3.2. CS Classification Results

When analyzing the test dataset with the neural network model, the performance metrics revealed are as follows: the weighted average (WA) results yield an average accuracy of 89.7%, sensitivity of 92.4%, precision of 91.1%, and F1-score of 91.2%.
Further dissection of the Cervical Spondylosis (CS) classification outcomes across four distinct imaging positions indicates exceptional performance in the right oblique (R) position, with perfect scores of 100.0% in accuracy, sensitivity, precision, and F1-score. The lateral (B) position also recorded a precision rate of 100.0%. These performance metrics were subjected to χ 2 testing, affirming that the CS classification accuracy across different positions does not significantly deviate from the overall observed accuracy. This analysis is visually supported in Figure 4 and Figure 5, and Table 4.

4. Discussion

Overcoming the subjective limitations inherent in manual X-ray diagnosis of Cervical Spondylosis (CS) and enhancing diagnostic accuracy have been consistent objectives within the research community. It has been indicated that the accuracy of manual X-ray diagnosis for CS is only 68.3% [11]. Conversely, the application of artificial intelligence (AI) offers a promising avenue for the prediction and diagnosis of CS [22]. Yu [23] developed a CS classification model using fuzzy computing theory, achieving an accuracy of 80.33%, thus demonstrating the potential of machine learning in classifying and processing various imaging features effectively.
This study extends these findings by integrating data from multiple cervical spine projection positions and conducting a comprehensive analysis of imaging features prevalent in CS diagnoses. The development of an innovative CS classification model based on ResNet-34 is reported, achieving a classification accuracy of 89.7%. The research is characterized by several advantages: (a) the integration of data from various X-ray projection positions, and (b) the exhaustive analysis of numerous imaging features.

4.1. Multi- X-ray Projection Position Imaging Data

Cervical spine X-ray imaging data from various positions, including anterior-posterior, lateral, left oblique, and right oblique, were consolidated. Identifying image features from different positions is crucial for model training, as each image feature has a primary observation position [24]. Initially, CS X-ray projection positions are identified to ensure the deep learning model learns the most important CS image features from these specific positions, aiding in precise feature localization and thus enhancing CS diagnosis accuracy.
In the field of AI-based CS prediction and diagnosis, the lateral view of the cervical spine remains a primary perspective due to the rich diagnostic information it provides and the minimal overlap and relatively simple anatomical structures of lateral cervical spine images, which simplify data processing [25]. Lee’s [25] study reported an accuracy of 87.1% for CS classification in lateral positions. A classification accuracy of 90.2% for CS in lateral positions was achieved by the model (Table 4).
Conversely, the anterior–posterior position of the cervical spine requires the classification of numerous CS imaging features, with added complexity due to the overlap of anatomical structures such as the calcification of thyroid cartilage, the hyoid bone, and the mandible. Limited research focuses on anterior–posterior positions of the cervical spine, mostly on segmentation-related studies. The model achieved an accuracy of 85.5% for CS classification in anterior-posterior positions (Table 4).
Oblique positions of the cervical spine present challenges due to overlapping anatomical structures and strict irradiation positioning requirements, leading to unstable feature extraction. However, their advantage lies in the fewer image features involved in the model, related primarily to hook joint degeneration and osteoporosis, simplifying data processing. Park [26] proposed a cervical spine oblique position classification model based on ResNet50, with an identification accuracy of 77% for foraminal stenosis. The model demonstrated an accuracy of 95.0% for CS classification in oblique positions, showing superior performance (Table 4).
To validate the performance consistency of the model in CS classification across different X-ray projection positions, indicators were compared between positions. Statistical results showed no significant difference in various indicators of CS classification between different positions, confirming the stable and consistent performance of the model (Table 5).

4.2. Multi-Label CNN Model

Current research predominantly examines single imaging features of Cervical Spondylosis (CS); however, the radiological diagnosis of CS necessitates a comprehensive analysis of multiple imaging features. This study aims to empower artificial intelligence to autonomously identify CS image features and perform image diagnosis. Unlike models that concentrate on a singular feature, this approach involves training the model with images that display various CS imaging features in different numbers and positions. A multi-label neural network model was adopted, designed to output the probability of each category, thus facilitating the flexible classification of multiple labels. Additionally, the efficiency of the algorithm and the generalization capability of the model were enhanced through the design of loss functions tailored for optimizing multi-task learning.
While analyses based on single image features have produced satisfactory outcomes as evidenced by studies such as Jebri’s [27] use of a machine learning-based random forest classifier model for detecting intervertebral space narrowing and osteophyte formation, Tamai’s [28] segmentation of cervical ligament calcification using the EfficientNet-B2 model, Fujimori’s [29] application of the EfficientNet-B4 model to identify cervical lordosis, and Chen’s [30] diagnosis of cervical spine scoliosis using the ResNet model. These contributions notwithstanding, CS cannot be adequately represented by single image features alone but rather as a constellation of multiple imaging features. Therefore, a comprehensive analysis of varied imaging features is more reflective of the clinical diagnostic requirements for CS. Given the complexities associated with processing data characterized by multiple imaging features, classifying multi-feature patterns remains a challenging endeavor. The high accuracy achieved in this study is primarily attributed to the neural network model’s extensive coverage of essential X-ray imaging features of CS, positioning this research at the forefront of clinical relevance.
Despite the model’s exceptional performance in accuracy, sensitivity, precision, and F1-score, opportunities for enhancement remain. Notably, this retrospective study did not utilize standardized cervical spine imaging projection patterns. Variabilities such as hardware differences across devices, individual disparities resulting from X-ray imaging parameters (e.g., X-ray intensity, irradiation distance, and X-ray incidence angle), subjective imaging diagnostic criteria, and variations in patient positions complicate data processing (Figure 6). To mitigate overfitting associated with complex data, a novel model-processing method was employed. Parameters were saved not after every 100 training cycles but from the training sessions that exhibited the best performance within each set of 100 cycles. This approach, by minimizing the overfitting’s impact on model parameters, improved the model’s accuracy by 1.5% compared to traditional methods.

4.3. Comparative Performance Analysis

Table 6 showcases a comparative analysis of implementation accuracy among various deep convolutional neural network models. Notably, our study stands out with the ResNet34 model, achieving remarkable accuracy across all three views: anteroposterior (85.50%), lateral (90.20%), and oblique (95.00%). This comprehensive performance is unparalleled when compared to the other studies listed, which focus solely on the lateral view.
For instance, Yüksel’s [31] work with the VGG-16 model in 2022 achieved an accuracy of 93.90% in the lateral view, which is commendable but does not address the anteroposterior and oblique views. Similarly, other studies like those by Miura [32] and Tamai [28] using EfficientNetB4 and EfficientNetB2, respectively, show lower accuracies in the lateral view (86.00% and 88.00%) and do not provide solutions for the other views.
Moreover, Ogawa [33] and Lee’s [25] research with CNN models in 2022 also focus on the lateral view, with accuracies of 90.00% and 87.10% respectively, which are surpassed by our ResNet34’s performance. Park’s [26] study using ResNet50 is the only other work addressing a different view, the oblique but with a significantly lower accuracy of 77.00%.
Our method’s superiority is evident not only in the highest accuracy achieved in the lateral view but also in its versatility and robustness, demonstrated by its performance in the anteroposterior and oblique views. This indicates a significant advancement in the field, suggesting that our ResNet34 model can be a more reliable and comprehensive tool for image analysis in various orientations. The breadth and depth of accuracy in our study underscore its potential as the leading methodology in deep convolutional neural network research for image classification.

5. Conclusions

This study underscores the escalating prevalence of Cervical Spondylosis (CS) amidst evolving societal dynamics and lifestyle modifications, illuminating the consequential surge in the necessity for efficient and accurate CS diagnostic methodologies. The paramount importance of early intervention in the management and prognosis of CS is highlighted against the backdrop of the current limitations encountered in the diagnostic accuracy of CS via X-ray imaging. These limitations are primarily attributed to the non-distinct nature of X-ray imaging features and the substantial reliance on the diagnostic acumen of medical practitioners.
The deployment of the proposed neural network model represents a significant stride toward mitigating these challenges. Model selection is very important, and model optimization algorithms [34] are also the direction to improve accuracy in the future. By harnessing advanced machine learning techniques, the model demonstrates a robust capacity to accurately identify CS from X-ray images, thus offering a critical tool to support clinicians, particularly those at the nascent stages of their careers or those with limited diagnostic experience. Consequently, this model stands to substantially reduce the incidence of diagnostic errors, such as misdiagnoses or missed diagnoses, and to alleviate the workload burden on healthcare professionals.
Furthermore, the findings from this study lay a foundation for future research endeavors aimed at refining AI-driven diagnostic models, with the potential to enhance the precision and reliability of CS detection and classification. It opens avenues for the integration of such models into clinical workflows, thereby augmenting the diagnostic process with a level of accuracy and efficiency that aligns closely with the nuanced demands of contemporary medical practice.
In conclusion, the development and application of the neural network model for CS diagnosis not only advance the field of medical diagnostics but also promise significant improvements in patient care outcomes. By addressing the challenges associated with the current diagnostic modalities for CS, this study contributes to the broader narrative of leveraging artificial intelligence to enrich healthcare delivery, underscored by a commitment to accuracy, efficiency, and accessibility.

Author Contributions

Y.X. performed the data collection, experiments and statistical analysis. Y.N. performed the Model selection, code writing and parameter adjustment. Y.X. and Z.C. performed Imaging diagnosis. Y.X. and Y.N. drafted the article, J.L., M.Y. and Y.Z. revised the article. Z.C. designed and guided the experiments. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Capital’s Funds for Health Improvement and Research [grant number 2020-2-6012].

Institutional Review Board Statement

The research protocol was approved by the Ethics Review Committee of Beijing Boai Hospital, China Rehabilitation Research Center (approval number 2020-083-1).

Informed Consent Statement

In accordance with the retrospective review policy, informed consent was not required.

Data Availability Statement

The datasets generated and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this paper:
CScervical spondylosis
AIartificial intelligence
MLmachine learning
DLdeep learning
CNNconvolutional neural network
Athe anteroposterior position of the cervical spine
Bthe lateral position of the cervical spine
Lthe left oblique position of the cervical spine
Rthe right oblique position of the cervical spine
Othe set of right and left oblique positions of the cervical spine
WAthe weighted average based on the proportions of images in A, B, L, and R positions

References

  1. Theodore, N. Degenerative cervical spondylosis. N. Engl. J. Med. 2020, 383, 159–168. [Google Scholar] [CrossRef]
  2. Hurwitz, E.L.; Randhawa, K.; Yu, H.; Côté, P.; Haldeman, S. The Global Spine Care Initiative: A summary of the global burden of low back and neck pain studies. Eur. Spine J. 2018, 27, 796–801. [Google Scholar] [CrossRef]
  3. Bakhsheshian, J.; Mehta, V.A.; Liu, J.C. Current diagnosis and management of cervical spondylotic myelopathy. Glob. Spine J. 2017, 7, 572–586. [Google Scholar] [CrossRef] [PubMed]
  4. Kim, G.U.; Chang, M.C.; Kim, T.U.; Lee, G.W. Diagnostic modality in spine disease: A review. Asian Spine J. 2020, 14, 910. [Google Scholar] [CrossRef]
  5. Hesni, S.; Baxter, D.; Saifuddin, A. The imaging of cervical spondylotic myeloradiculopathy. Skelet. Radiol. 2023, 52, 2341–2365. [Google Scholar] [CrossRef] [PubMed]
  6. Urrutia, J.; Zamora, T.; Yurac, R.; Campos, M.; Palma, J.; Mobarec, S.; Prada, C. An independent inter-and intraobserver agreement evaluation of the AOSpine subaxial cervical spine injury classification system. Spine 2017, 42, 298–303. [Google Scholar] [CrossRef] [PubMed]
  7. Christe, A.; Läubli, R.; Guzman, R.; Berlemann, U.; Moore, R.; Schroth, G.; Vock, P.; Lövblad, K. Degeneration of the cervical disc: Histology compared with radiography and magnetic resonance imaging. Neuroradiology 2005, 47, 721–729. [Google Scholar] [CrossRef]
  8. Chea, P.; Mandell, J.C. Current applications and future directions of deep learning in musculoskeletal radiology. Skelet. Radiol. 2020, 49, 183–197. [Google Scholar] [CrossRef]
  9. Huber, F.A.; Guggenberger, R. AI MSK clinical applications: Spine imaging. Skelet. Radiol. 2022, 51, 279–291. [Google Scholar] [CrossRef]
  10. Akay, G.; Akcayol, M.A.; Özdem, K.; Güngör, K. Deep convolutional neural network—The evaluation of cervical vertebrae maturation. Oral Radiol. 2023, 39, 629–638. [Google Scholar] [CrossRef]
  11. Yu, X.; Liu, M.; Meng, L.; Xiang, L. Classifying cervical spondylosis based on X-ray quantitative diagnosis. Neurocomputing 2015, 165, 222–227. [Google Scholar] [CrossRef]
  12. Wang, Z.; Liu, Z.; Wang, Z.W.; Ding, W.Y.; Yang, D.L. Does cervical facet degeneration impact surgical outcomes and sagittal balance in patients with radiculopathy? BMC Surg. 2021, 21, 225. [Google Scholar] [CrossRef] [PubMed]
  13. Murata, K.; Endo, K.; Suzuki, H.; Matsuoka, Y.; Takamatsu, T.; Nishimura, H.; Yamamoto, K. Spinal sagittal alignment and trapezoidal deformity in patients with degenerative cervical spondylolisthesis. Sci. Rep. 2019, 9, 4992. [Google Scholar] [CrossRef] [PubMed]
  14. Martin, C.; Sharp, P.; Sutton, D. Measurement of image quality in diagnostic radiology. Appl. Radiat. Isot. 1999, 50, 21–38. [Google Scholar] [CrossRef]
  15. Al-Waisy, A.S.; Al-Fahdawi, S.; Mohammed, M.A.; Abdulkareem, K.H.; Mostafa, S.A.; Maashi, M.S.; Arif, M.; Garcia-Zapirain, B. COVID-CheXNet: Hybrid deep learning framework for identifying COVID-19 virus in chest X-rays images. Soft Comput. 2023, 27, 2657–2672. [Google Scholar] [CrossRef] [PubMed]
  16. Zhuang, Q.; Gan, S.; Zhang, L. Human-computer interaction based health diagnostics using ResNet34 for tongue image classification. Comput. Methods Programs Biomed. 2022, 226, 107096. [Google Scholar] [CrossRef] [PubMed]
  17. Miao, K.; Zhao, N.; Lv, Q.; He, X.; Xu, M.; Dong, X.; Li, D.; Shao, X. Prediction of benign and malignant ovarian tumors using Resnet34 on ultrasound images. J. Obstet. Gynaecol. Res. 2023, 49, 2910–2917. [Google Scholar] [CrossRef] [PubMed]
  18. Imagawa, K.; Shiomoto, K. Performance change with the number of training data: A case study on the binary classification of COVID-19 chest X-ray by using convolutional neural networks. Comput. Biol. Med. 2022, 142, 105251. [Google Scholar] [CrossRef]
  19. Shakya, K.S.; Priti, K.; Jaiswal, M.; Laddi, A. Segmentation of Sella Turcica in X-ray Image based on U-Net Architecture. Procedia Comput. Sci. 2023, 218, 828–835. [Google Scholar] [CrossRef]
  20. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  21. Hao, J.; Ho, T.K. Machine learning made easy: A review of scikit-learn package in python programming language. J. Educ. Behav. Stat. 2019, 44, 348–361. [Google Scholar] [CrossRef]
  22. Sreeraj, M.; Joy, J.; Jose, M.; Varghese, M.; Rejoice, T. Comparative analysis of Machine Learning approaches for early stage Cervical Spondylosis detection. J. King Saud-Univ.-Comput. Inf. Sci. 2022, 34, 3301–3309. [Google Scholar] [CrossRef]
  23. Yu, X.; Xiang, L. Classifying cervical spondylosis based on fuzzy calculation. In Abstract and Applied Analysis; Hindawi: London, UK, 2014; Volume 2014. [Google Scholar]
  24. Meacock, J.; Schramm, M.; Selvanathan, S.; Currie, S.; Stocken, D.; Jayne, D.; Thomson, S. Systematic review of radiological cervical foraminal grading systems. Neuroradiology 2021, 63, 305–316. [Google Scholar] [CrossRef] [PubMed]
  25. Lee, G.W.; Shin, H.; Chang, M.C. Deep learning algorithm to evaluate cervical spondylotic myelopathy using lateral cervical spine radiograph. BMC Neurol. 2022, 22, 147. [Google Scholar] [CrossRef] [PubMed]
  26. Park, J.; Yang, J.; Park, S.; Kim, J. Deep learning-based approaches for classifying foraminal stenosis using cervical spine radiographs. Electronics 2022, 12, 195. [Google Scholar] [CrossRef]
  27. Jebri, B.; Phillips, M.; Knapp, K.; Appelboam, A.; Reuben, A.; Slabaugh, G. Detection of degenerative change in lateral projection cervical spine X-ray images. In Medical Imaging 2015: Computer-Aided Diagnosis; SPIE: Bellingham, WA, USA, 2015; Volume 9414, pp. 18–25. [Google Scholar]
  28. Tamai, K.; Terai, H.; Hoshino, M.; Yabu, A.; Tabuchi, H.; Sasaki, R.; Nakamura, H. A deep learning algorithm to identify cervical ossification of posterior longitudinal ligaments on radiography. Sci. Rep. 2022, 12, 2113. [Google Scholar] [CrossRef] [PubMed]
  29. Fujimori, T.; Suzuki, Y.; Takenaka, S.; Kita, K.; Kanie, Y.; Kaito, T.; Ukon, Y.; Watabe, T.; Nakajima, N.; Kido, S.; et al. Development of artificial intelligence for automated measurement of cervical lordosis on lateral radiographs. Sci. Rep. 2022, 12, 15732. [Google Scholar] [CrossRef]
  30. Chen, P.; Zhou, Z.; Yu, H.; Chen, K.; Yang, Y. Computerized-assisted scoliosis diagnosis based on faster R-CNN and resnet for the classification of spine X-ray images. Comput. Math. Methods Med. 2022, 2022, 3796202. [Google Scholar] [CrossRef] [PubMed]
  31. Maraş, Y.; Tokdemir, G.; Üreten, K.; Atalar, E.; Duran, S.; Maraş, H. Diagnosis of osteoarthritic changes, loss of cervical lordosis, and disc space narrowing on cervical radiographs with deep learning methods. Jt. Dis. Relat. Surg. 2022, 33, 93. [Google Scholar] [CrossRef]
  32. Miura, M.; Maki, S.; Miura, K.; Takahashi, H.; Miyagi, M.; Inoue, G.; Murata, K.; Konishi, T.; Furuya, T.; Koda, M.; et al. Automated detection of cervical ossification of the posterior longitudinal ligament in plain lateral radiographs of the cervical spine using a convolutional neural network. Sci. Rep. 2021, 11, 12702. [Google Scholar] [CrossRef]
  33. Ogawa, T.; Yoshii, T.; Oyama, J.; Sugimura, N.; Akada, T.; Sugino, T.; Hashimoto, M.; Morishita, S.; Takahashi, T.; Motoyoshi, T.; et al. Detecting ossification of the posterior longitudinal ligament on plain radiographs using a deep convolutional neural network: A pilot study. Spine J. 2022, 22, 934–940. [Google Scholar] [CrossRef] [PubMed]
  34. Liu, J.; Wei, J.; Heidari, A.A.; Kuang, F.; Zhang, S.; Gui, W.; Chen, H.; Pan, Z. Chaotic simulated annealing multi-verse optimization enhanced kernel extreme learning machine for medical diagnosis. Comput. Biol. Med. 2022, 144, 105356. [Google Scholar] [CrossRef] [PubMed]
Figure 1. F1 score of ResNet-34 model from validation results. The abscissa is the F1-score of each validation from training the model, and the ordinate is the number of times the model is trained. We saved the final CNN from the validation with the best performance within 100 epochs.
Figure 1. F1 score of ResNet-34 model from validation results. The abscissa is the F1-score of each validation from training the model, and the ordinate is the number of times the model is trained. We saved the final CNN from the validation with the best performance within 100 epochs.
Sensors 24 03428 g001
Figure 2. CNN model flow chart. We categorize the input data into a test set and a training-validation set. The data in the training-validation set undergo manual labeling, and the results are fed into the subsequent step. Subsequently, preprocessing is applied to the training-validation set data. Following preprocessing, the training set data are employed to train the ResNet-34 model, while the validation set data are utilized for selecting the final model. The ultimately saved model is then evaluated for performance using the test dataset.
Figure 2. CNN model flow chart. We categorize the input data into a test set and a training-validation set. The data in the training-validation set undergo manual labeling, and the results are fed into the subsequent step. Subsequently, preprocessing is applied to the training-validation set data. Following preprocessing, the training set data are employed to train the ResNet-34 model, while the validation set data are utilized for selecting the final model. The ultimately saved model is then evaluated for performance using the test dataset.
Sensors 24 03428 g002
Figure 3. Confusion matrix for classification of cervical vertebrae X-ray projection position on test dataset.
Figure 3. Confusion matrix for classification of cervical vertebrae X-ray projection position on test dataset.
Sensors 24 03428 g003
Figure 4. Confusion matrix for classification of CS on test dataset (P: positive, N: negative).
Figure 4. Confusion matrix for classification of CS on test dataset (P: positive, N: negative).
Sensors 24 03428 g004
Figure 5. Confusion matrix for classification of CS from different positions. (a) is from anteroposterior positions; (b) is from lateral positions; (c) is from left oblique positions; (d) is from right oblique positions.
Figure 5. Confusion matrix for classification of CS from different positions. (a) is from anteroposterior positions; (b) is from lateral positions; (c) is from left oblique positions; (d) is from right oblique positions.
Sensors 24 03428 g005
Figure 6. Cervical vertebra X-ray projection position classification and CS classification results. Figure (ac) represent that the cervical spine position is correctly classified but CS is incorrectly classified. Figure (d) represents that the cervical spine position is classified incorrectly but CS is correctly classified. Figure (eh) represent images with correct cervical vertebra position classification and CS classification. (“N” indicates CS is negative; “P” indicates CS is positive; the red line indicates the physiological curvature of the cervical spine; the red dotted arrow indicates the hyoid bone blocking the cervical vertebrae. the yellow arrow indicates the intervertebral space; the yellow dotted arrow indicates osteophyte formation; the blue dotted arrow indicates ligament calcification; and the blue arrow indicates the intervertebral foramen).
Figure 6. Cervical vertebra X-ray projection position classification and CS classification results. Figure (ac) represent that the cervical spine position is correctly classified but CS is incorrectly classified. Figure (d) represents that the cervical spine position is classified incorrectly but CS is correctly classified. Figure (eh) represent images with correct cervical vertebra position classification and CS classification. (“N” indicates CS is negative; “P” indicates CS is positive; the red line indicates the physiological curvature of the cervical spine; the red dotted arrow indicates the hyoid bone blocking the cervical vertebrae. the yellow arrow indicates the intervertebral space; the yellow dotted arrow indicates osteophyte formation; the blue dotted arrow indicates ligament calcification; and the blue arrow indicates the intervertebral foramen).
Sensors 24 03428 g006
Table 1. Calculation formula for evaluation indicators.
Table 1. Calculation formula for evaluation indicators.
IndicatorsDefinition and Calculation Formula
Accuracy(True Positive + True Negative) / (Population Sample)
SensitivityTrue Positive / (True Positive + False Negative)
PrecisionTrue Positive / (True Positive + False Positive)
F1 Score2 × (Precision × Sensitivity) / (Precision + Sensitivity)
Table 2. Demographic information.
Table 2. Demographic information.
PopulationTraining-Verification SetStatistics P Test SetStatistics P
No. of images137112350.002>0.9991360.0360.998
A 1565509//55//
B 2410370//41//
L 3196176//20//
R 4200180//20//
GenderFemale 805(58.7%)Female 724(58.6%)0.0020.962Female 81(59.6%)0.0360.849
Age46.4 ± 16.446.2 ± 16.0−0.2560.79847.8 ± 20.30.9760.329
Positive57.8%57.7%<0.0010.98558.1%0.0050.943
The test method for number of images, gender and positivity rate indicators is the χ 2 test. The test method for age indicators is the T test. 1 represents the anteroposterior position of the cervical spine. 2 represents the lateral position of the cervical spine. 3 represents the left oblique position of the cervical spine. 4 represents the right oblique position of the cervical spine.
Table 3. Statistical table for classification of cervical vertebrae X-ray projection position.
Table 3. Statistical table for classification of cervical vertebrae X-ray projection position.
NumberAccuracyStatistics P SensitivityStatistics P PrecisionStatistics P F1 ScoreStatistics P
A55100.0%1.2330.267100.0%1.2330.267100.0%1.2330.267100.0%1.2330.267
B41100.0%0.9200.337100.0%0.9200.337100.0%0.9200.337100.0%0.9200.337
L20100.0%0.4500.502100.0%0.4500.50287.0%7.7170.00593.0%0.5450.460
R2085.0%7.7170.00585.0%7.7170.005100.0%0.4500.50291.9%3.4140.065
O 14092.5%2.6310.10592.5%2.6310.10593.5%2.6310.10592.5%2.6310.105
WA 213697.8%//97.8%//98.1%//97.8%//
A, B, L, R and O are independently contrasted with the WA. The test method used χ 2 test; if P > 0.01, it indicated no statistically significant differences in the data. 1 represents the set of right and left oblique positions of the cervical spine. 2 represents the weighted average according to the proportion of A, B, L, R.
Table 4. Statistical table for classification of CS.
Table 4. Statistical table for classification of CS.
NumberAccuracyStatistics P SensitivityStatistics P PrecisionStatistics P F1 ScoreStatistics P
A5585.5%0.6950.40596.8%0.9190.33881.1%3.3650.06788.2%0.2000.655
B4190.2%0.0100.92082.6%3.4280.064100.0%3.8810.04990.5%0.0330.855
L2090.0%0.0020.96892.3%0.1720.67892.3%0.0300.86492.3%0.0300.864
R20100.0%2.2620.133100.0%1.5710.210100.0%1.9120.167100.0%1.9120.167
O4095.0%1.0480.30696.2%0.2690.60496.2%0.6170.43296.2%0.6170.432
WA13689.7%//92.4%//91.1%//91.2%//
Table 5. Performance comparison of deep learning models for CS classification at different projection positions.
Table 5. Performance comparison of deep learning models for CS classification at different projection positions.
CompareAccuracySensitivityPrecisionF1 Score
Statistics P Statistics P Statistics P Statistics P
A–B0.4930.4834.9920.0256.4870.0110.0330.855
A–O2.2400.1340.1070.7443.6460.0560.0190.306
B–O0.6680.4142.9880.0842.1020.1470.6680.414
L–R2.1050.1472.1050.1472.1050.1472.1050.147
A–L0.2620.6091.1760.2780.7310.3930.0130.910
A–R3.2560.0710.7470.3874.1960.0412.3720.124
B–L0.0010.9760.5350.4654.2390.0400.0010.976
B–R2.0880.1483.8750.050<0.001>0.9992.0880.148
A, B, L, R, and O were compared pairwise. The test method used χ 2 test; if P > 0.01, it indicated no statistically significant differences in the data. Dash (–) represents the comparison between two projected positions.
Table 6. Implementation accuracy comparison with previous research works using deep convolutional neural network.
Table 6. Implementation accuracy comparison with previous research works using deep convolutional neural network.
AuthorYearModelAccuracy
Anteroposterior Lateral Oblique
Yüksel et al. [31]2022VGG-16-93.90%-
Miura et al. [32]2021EfficientNetB4-86.00%-
Tamai et al. [28]2022EfficientNetB2-88.00%-
Ogawa et al. [33]2022CNN-90.00%-
Lee et al. [25]2022CNN-87.10%-
Park et al. [26]2022ResNet50--77.00%
Our study ResNet3485.50%90.20%95.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, Y.; Nie, Y.; Lundgren, J.; Yang, M.; Zhang, Y.; Chen, Z. Cervical Spondylosis Diagnosis Based on Convolutional Neural Network with X-ray Images. Sensors 2024, 24, 3428. https://doi.org/10.3390/s24113428

AMA Style

Xie Y, Nie Y, Lundgren J, Yang M, Zhang Y, Chen Z. Cervical Spondylosis Diagnosis Based on Convolutional Neural Network with X-ray Images. Sensors. 2024; 24(11):3428. https://doi.org/10.3390/s24113428

Chicago/Turabian Style

Xie, Yang, Yali Nie, Jan Lundgren, Mingliang Yang, Yuxuan Zhang, and Zhenbo Chen. 2024. "Cervical Spondylosis Diagnosis Based on Convolutional Neural Network with X-ray Images" Sensors 24, no. 11: 3428. https://doi.org/10.3390/s24113428

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop