Next Article in Journal
Development of a Deep-Learning-Based Computerized Scoring Algorithm
Previous Article in Journal
High-Performance Self-Powered Photodetector Enabled by Te-Doped GeH Nanostructures Engineering
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Classification of Grades of Subchondral Sclerosis from Knee Radiographic Images Using Artificial Intelligence

by
Soo-Been Kim
1,
Young Jae Kim
2,
Joon-Yong Jung
3 and
Kwang Gi Kim
4,*
1
Medical Devices R&D Center, Gil Medical Center, Gachon University, Incheon 21565, Republic of Korea
2
Gachon Biomedical & Convergence Institute, Gil Medical Center, Gachon University, Incheon 21565, Republic of Korea
3
Department of Radiology, College of Medicine, The Catholic University of Korea, Seoul St. Mary’s Hospital, Seoul 06591, Republic of Korea
4
Department of Biomedical Engineering, College of Medicine, Gil Medical Center, Gachon University, Incheon 21565, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(8), 2535; https://doi.org/10.3390/s25082535
Submission received: 11 March 2025 / Revised: 15 April 2025 / Accepted: 15 April 2025 / Published: 17 April 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Osteoarthritis (OA) is the most common joint disease, affecting over 300 million people worldwide. Subchondral sclerosis is a key indicator of OA. Currently, the diagnosis of subchondral sclerosis is primarily based on radiographic images; however, reliability issues exist owing to subjective evaluations and inter-observer variability. This study proposes a novel diagnostic method that utilizes artificial intelligence (AI) to automatically classify the severity of subchondral sclerosis. A total of 4019 radiographic images of the knee were used to train the 3-Layer CNN, DenseNet121, MobileNetV2, and EfficientNetB0 models. The best-performing model was determined based on sensitivity, specificity, accuracy, and area under the curve (AUC). The proposed model exhibited outstanding performance, achieving 84.27 ± 1.03% sensitivity, 92.46 ± 0.49% specificity, 84.70 ± 0.98% accuracy, and 95.17 ± 0.41% AUC. The analysis of variance confirmed significant performance differences across models, age groups, and sexes (p < 0.05). These findings demonstrate the utility of AI in diagnosing and treating knee subchondral sclerosis and suggest that this approach could provide a new diagnostic method in clinical medicine. By precisely classifying the grades of subchondral sclerosis, this method contributes to improved overall diagnostic accuracy and offers valuable insights for clinical decision-making.

1. Introduction

Osteoarthritis (OA) is the most prevalent joint disease, affecting over 300 million people worldwide or approximately 4% of the global population [1,2]. The global prevalence of OA is estimated to be 16% in individuals aged 15 years and older, and around 30% in those aged 60 years and above [3]. Notably, women are more susceptible to OA, and the incidence of OA is expected to increase with the increase in the aging population [4,5,6,7].
The image classification capabilities of deep learning have improved substantially with the development of convolutional neural networks (CNNs) [8]. The visual assessment of radiographic images for subchondral sclerosis is often subject to reliability issues due to subjective evaluations and interobserver variability [9,10]. To mitigate these uncertainties in diagnosis, AI-based research has been increasingly explored in the medical field [11] with promising outcomes observed in studies on the importance of knee OA feature analysis [12] and knee OA grading [13,14].
Tiulpin et al. proposed a deep learning-based approach for the automatic diagnosis of knee OA using plain radiographs, achieving a classification accuracy of 66.71% on the osteoarthritis initiative (OAI) and multicenter osteoarthritis study (MOST) datasets using the ResNet34 model [15]. Patron et al. utilized a ResNet50-32×4d model to study the tibial region, an early sign of arthritis, and achieved an overall accuracy of 86.9%, sensitivity of 90.9%, and specificity of 75% using the OAI and MOST datasets [16]. Von Schacky et al. employed the RetinaNet model for hip OA feature grading using DenseNet to learn specific features, such as femoral osteophytes, acetabular osteophytes, joint space narrowing, subchondral sclerosis, and subchondral cysts, achieving subchondral sclerosis accuracy of 95.8% on an internal test set and 88.5% on an external test set [17]. Abdullah and Rajasekaran used Faster R-CNN to detect knee joint space width and AlexNet for classification, achieving an accuracy of 98.9% [18]. Yoon et al. leveraged OAI data for automatic knee position detection and tibial width measurement and developed an AI model (MediAI-OA) for Kellgren–Lawrence (KL) grade classification and osteophyte detection using the NASNet model, with an overall accuracy of 92% [19]. Gornale et al. developed an algorithm to calculate the cartilage area and thickness from knee radiographic images, achieving an accuracy of 99.81% with a K-NN classifier and 95.09% with a decision tree classifier [20]. Subha et al. proposed a dual convolutional neural network based on the Gaussian Aquila Optimizer, which achieved an accuracy of 98.77%, sensitivity of 98.25%, and specificity of 98.93% on 2283 knee images [21]. Mohammed et al. performed binary and three-class classifications of OA presence and severity using pre-trained models, with ResNet101 achieving maximum classification accuracies of 69%, 83%, and 89% for each task [22]. Mahum et al. conducted automatic OA detection and classification using the DenseNet pre-trained model on the OAI dataset, recording an accuracy of 98.22% [23]. Song et al. proposed KOA-CAD, integrating a Laplacian-based strategy and AMD-CNN for the detection and grading of knee OA using vibroarthrography and physiological signals, achieving an automatic detection accuracy of 93.6%, early detection accuracy of 92.1%, and grading detection accuracy of 84.2% [24]. Boniatis et al. introduced a classification ensemble for assessing the severity of hip OA by manually segmenting 64 regions of interests (ROIs) corresponding to the hip space and generating new texture features, achieving accuracies of 100% and 95.7% for normal/OA (stage 1) and mild/moderate-to-severe (stage 2) classifications, respectively [25]. Michael Fei et al. proposed a regression-based deep learning model using the EfficientNet architecture to predict continuous KL scores from radiographs, achieving an AUC of 0.83 [26].
Currently, the most widely used method for radiographic evaluation of osteoarthritis (OA) is the Kellgren–Lawrence (KL) grading system, which classifies OA severity on a scale from 0 (normal) to 4 (severe) based on radiographic images. However, the KL grade primarily focuses on osteophytes, which has been pointed out as a limitation [27]. For a more accurate assessment of OA, detailed investigations of specific features such as osteophytes, joint space narrowing, subchondral sclerosis, bone cysts, and joint deformity are needed. Although studies have attempted to quantify OA severity, research specifically analyzing subchondral sclerosis remains limited, and most existing studies only evaluate its presence or absence. Cooper et al. emphasized the importance of the reproducibility of individual component features in the development of radiographic systems for grading OA and reported low reliability in the assessment of subchondral sclerosis, leading to poor reproducibility [28]. Yoon et al. highlighted the need for the development of related models to evaluate OA features because the MediAI-OA model cannot assess subchondral sclerosis or bone abnormalities [19].
To address these issues, this study proposes a deep-learning-based radiograph data-driven classification model for knee subchondral sclerosis. The clinical utility of this model was validated by comparing it with the existing KL grading system using metrics such as sensitivity, specificity, accuracy, and AUC. Thus, we aimed to develop a more objective and precise evaluation method for subchondral sclerosis, thereby enhancing the reliability of the diagnosis and improving clinical decision support systems.

2. Methods

2.1. Data Collection and Labeling

In this study, 4019 knee joint radiographic images were collected from 4019 patients at the Catholic University Hospital. The dataset consisted of 643 men and 3376 women participants, with a mean age of 66.81 years (±11.83 years), ranging from 6 to 96 years. All the procedures conducted in this study adhered to the ethical principles outlined in the Declaration of Helsinki and were approved by the Institutional Review Board (IRB No. KC23RIDI0485).
The ROI was delineated by a specialist on the knee radiographic images following the criteria illustrated in Figure 1. The labeling process was independently conducted and verified by two specialists. In cases in which different grades were assigned, the final decision was reached through consensus. Figure 1 presents examples of subchondral labeling and subchondral sclerosis grades, with the subchondral sclerosis areas highlighted. Grade 0 indicated no abnormalities in the subchondral region, Grade 1 indicated mild sclerosis or partial destruction of the subchondral bone surface, and Grade 2 indicated significant destruction that could impair joint function.
The dataset was distributed across three grades to avoid any imbalance: 1470 images in Grade 0, 1282 images in Grade 1, and 1267 images in Grade 2. Thereafter, the dataset was divided into training (80%), validation (10%), and testing (10%) datasets. This resulted in 3215 images for training and 402 images each for validation and testing. Table 1 provides a detailed breakdown of the data split according to subchondral sclerosis grades for the training, testing, and validation subsets. The distribution of subchondral sclerosis grades according to sex and age within the test dataset is presented in Table 2.

2.2. Training Environment and Data Preprocessing

The system used for training consisted of an NVIDIA GeForce GTX 1660 (NVIDIA, Santa Clara, CA, USA) graphics processing unit, an Intel Core i7 10700 CPU (Intel Corporation, Santa Clara, CA, USA), and 32 GB of RAM. Deep learning training was conducted on a Windows 10 Pro operating system using Python (Ver. 3.12.1) and the Keras framework (Ver. 2.10.0, with TensorFlow backend).
The DICOM images were converted to an 8-bit scale by adjusting the pixel value range to 0–255. This conversion was performed to reduce the memory usage and facilitate easier data processing and analysis. To study the subchondral regions, the ROI was designated as a mask to extract the relevant area from the images. Contrast limited adaptive histogram equalization (CLAHE), a contrast enhancement technique used in image processing, was applied to equalize the histogram of the image, thereby improving the local contrast of the images [29,30,31,32]. CLAHE was specifically employed to enhance the visibility of the features of subchondral sclerosis in the X-ray images. Finally, zero padding was applied to preserve the spatial dimensions of the images, prevent distortion, and improve training efficiency. Figure 2 illustrates the data preprocessing steps.

2.3. Subchondral Sclerosis Classification Models and Statistical Analysis

Four deep learning models were employed for the classification of subchondral sclerosis grades: 3-Layer CNN, DenseNet, MobileNet, and EfficientNet. These models were selected based on their proven effectiveness in medical image analysis [33,34,35]. The 3-Layer CNN model utilizes a series of convolutional and pooling layers to sequentially extract features and perform classification. DenseNet features a densely connected network architecture in which each layer receives the output of all preceding layers as input, enhancing feature propagation and efficiency. MobileNet combines depth-wise separable convolutions and pointwise convolutions to significantly reduce the model size and computation and achieve model lightness and speed improvements. EfficientNet is designed to improve network performance and efficiency by balancing the model depth, width, and resolution. To prevent overfitting during training, dropout was applied at a fixed rate of 0.4. A seed size of 42 was set to ensure reproducibility of the training process. Each model was pretrained on the ImageNet dataset and then fine-tuned for the subchondral sclerosis grading classification. Common training parameters included a batch size of 32, 50 epochs, and an Adam optimizer with a learning rate of 0.00001. Figure 3 illustrates the overall workflow for training of subchondral sclerosis grading classification models used in this study.
To assess differences in subchondral sclerosis grades by age group and sex, analysis of variance (ANOVA) was performed, followed by Tukey’s HSD test to identify significant differences. Additionally, the Mann–Whitney U test was used to evaluate differences in subchondral sclerosis detection accuracy between specific groups. All statistical tests were conducted with a significance level set at p-value < 0.05.

3. Results

In this study, four models were trained to classify subchondral sclerosis grades using a radiographic dataset, and their performances were evaluated using a test set of 402 images. The best-performing model was selected based on a 10-fold cross-validation. The sensitivity, specificity, accuracy, and AUC values were calculated and are presented in Table 3. The AUC for the 3-Layer CNN, DenseNet121, MobileNetV2, and EfficientNetB0 models were 89.52 ± 0.46%, 94.68 ± 0.84%, 94.45 ± 0.71%, and 95.17 ± 0.41%, respectively. ANOVA was conducted to assess the differences in accuracy among the models, revealing statistically significant differences in performance. The results of the ANOVA are presented in Table 3. To further investigate the significant differences identified by ANOVA, Tukey’s Honestly significant difference (HSD) test was performed. The results are presented in Table 4. The analysis revealed that the performance differences between the 3-Layer CNN model and all the other models (DenseNet121, MobileNetV2, and EfficientNetB0) were statistically significant (p < 0.05). However, no significant differences were observed in the performance between DenseNet121 and MobileNetV2 (p = 0.87), DenseNet121 and EfficientNetB0 (p = 0.15), or MobileNetV2 and EfficientNetB0 (p = 0.49).
Significant differences are highlighted, indicating that the models showed statistically significant differences in performance.
To analyze the results, the cases predicted by the models were sampled. Various examples of correctly classified subchondral sclerosis grades are shown in Figure 4. Figure 5 shows the image regions that the model focused on for classification using Gradient-weighted class activation mapping (Grad-CAM). Grad-CAM activation was observed in areas prone to subchondral sclerosis, allowing visual confirmation of the joint structural features associated with subchondral sclerosis.
Confusion matrices and receiver operating characteristic (ROC) curves were derived using the final test dataset to visually analyze the performance of the models in classifying subchondral sclerosis grades. The results are presented in Figure 6 and Figure 7.
The detection accuracy of the EfficientNetB0 model was compared based on age and sex. The age groups were divided into Group A (individuals in their 20s to 50s) and Group B (individuals in their 60s to 90s), and sex was categorized into Group C (men) and Group D (women). To ensure a fair assessment despite the data imbalance, the macro average accuracy was computed for each group. The results were as follows: the accuracy for age Groups A and B was 0.77 ± 0.09 and 0.84 ± 0.04, respectively, and that for men (Group C) and women (Group D) was 0.78 ± 0.10 and 0.84 ± 0.04, respectively. The results are summarized in Table 5. To determine whether the differences in accuracy between age groups A and B and sex groups C and D were statistically significant, Mann–Whitney U tests were conducted. The 12 results indicated statistically significant differences between the two groups (p < 0.05).

4. Discussion

This study proposes a novel method for evaluating the subchondral sclerosis grades based on knee radiographs using AI. A majority of studies have focused on knee joint grading or cartilage detection, with research on subchondral sclerosis typically addressing only its presence or absence. In contrast, this study utilized radiographic images collected from hospitals, incorporating diverse data from both men and women aged 6–96 years. It refines the grading of subchondral sclerosis into Grades 0, 1, and 2 for a more precise classification. Various AI models, including 3-Layer CNN, DenseNet121, MobileNetV2, and EfficientNetB0, were developed and compared. The EfficientNetB0 model demonstrated the best performance with an AUC of 95.17 ± 0.41% and an accuracy of 84.70 ± 0.98%. The ANOVA and Tukey’s HSD tests confirmed that the EfficientNetB0 model significantly outperformed the other models. Analysis of model performance by age group and sex using the Mann–Whitney U test revealed statistically significant differences (p < 0.05). In particular, the accuracy for women was 0.84 ± 0.04, and that for individuals in their 60s to 90s was also 0.84 ± 0.04. The difference in accuracy according to age was likely due to degenerative changes. In the 60–90 years age group (Group B), a higher incidence of arthritis and more advanced cartilage damage and sclerosis resulted in more distinct subchondral sclerosis lesions on radiographs. Conversely, in the 20–50 years age group (Group A) the relatively less cartilage damage suggests that subchondral sclerosis is less advanced, potentially leading to lower model detection accuracy. This supports the notion that the progression of subchondral sclerosis with age is reflected in the radiographic images [36]. Differences in accuracy by sex may be attributed to physiological changes, such as decreased bone density following menopause, which can accelerate cartilage damage in women and make subchondral sclerosis features more pronounced on radiography. This finding aligns with previous research on the relationship between knee OA and hormonal changes in postmenopausal women [37]. Subchondral sclerosis lesion characteristics vary according to age and sex, and the model can reflect these population characteristics. Von Schacky et al. [17] graded hip OA using deep learning models but only assessed the presence of subchondral sclerosis. In contrast, this study used actual patient data to subdivide subchondral sclerosis into grades 0, 1, and 2 for a more detailed evaluation. Additionally, while Patron et al. [16] used the OAI and MOST datasets, this study employed hospital data to develop a model suited to clinical environments. Mohammed et al. [22] classified OA severity into three grades; however, this study utilized the distinct radiographic features to classify subchondral sclerosis more precisely. Fei, M et al. [26] developed a regression-based model using EfficientNet to predict continuous KL scores, achieving an AUC of 0.83. However, their model revealed inconsistencies between KL labels and radiographic findings, such as visible sclerosis in KL 0 cases. In contrast, our model directly targets subchondral sclerosis as an independent feature, enabling a more precise evaluation of structural changes. As a result, it achieved an AUC of 95.17 ± 0.41%. Tiulpin et al. [15] achieved a multi-class accuracy of 66.71% and an AUC of 0.93 using probabilistic KL grade prediction with joint localization. In contrast, our model focuses on subchondral sclerosis as an independent feature, achieving a higher accuracy of 84.70 ± 0.98% and improved diagnostic specificity.
This study has several limitations. First, the amount of data used is relatively limited. The number of cases used (approximately 4000) may be insufficient for model training and generalization. Future research should collect a larger and more diverse set of patient data to improve model performance and data representativeness. Second, in this study, the CLAHE technique was applied to enhance image contrast, and all experiments were conducted using CLAHE-processed images to ensure consistent experimental conditions. As a result, the model was not evaluated using original images without CLAHE processing. This may be considered a limitation, and future research should include performance comparisons using original images as well. Third, to improve model generalization, it is necessary to collect data that are balanced across sex and age groups and to include radiographic images acquired from a variety of imaging equipment. This will be an important factor in assessing model performance and enhancing its reliability in clinical settings. Finally, the development of an automated system for detecting and classifying subchondral regions in knee radiographic images is required. Addressing these limitations requires further research and additional development.

5. Conclusions

This study demonstrates the potential of a radiograph-based classification model for subchondral sclerosis and proposes a novel approach for detailed grading of subchondral sclerosis using artificial intelligence. The EfficientNetB0 model achieved high accuracy (84.70 ± 0.98%) and AUC results (95.17 ± 0.41%) through 10-fold cross-validation, showing superior performance compared to that of existing studies. This indicates that the model has the potential to deliver reliable performance in clinical settings. Future research involving larger datasets and diverse models could further optimize the performance of AI-based classification models for subchondral sclerosis. This will enable a more objective and reliable classification of subchondral sclerosis and advance the development of systems that support clinical decision-making and diagnostic processes.

Author Contributions

S.-B.K. contributed to the methodology, software development, and drafting of the original manuscript. Y.J.K. and J.-Y.J. were involved in validation and participated in the review and editing of this manuscript. K.G.K. supervised this study and contributed to the review and editing process. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Korea Medical Device Development Fund grant funded by the Korean government (the Ministry of Science and ICT, the Ministry of Trade, Industry and Energy, the Ministry of Health & Welfare, and the Ministry of Food and Drug Safety) (Project Number: 1711196789, RS-2023-00252804) and by the Export-Oriented “2023 Small and Medium Business Technology Development (R&D) Support Project” (Grant No: RS-2023-00280710), funded by the Ministry of SMEs and Startups (MSS, Republic of Korea).

Institutional Review Board Statement

All procedures conducted in this study adhered to the ethical principles outlined in the Declaration of Helsinki and were approved by the Institutional Review Board (IRB No. KC23RIDI0485).

Informed Consent Statement

The requirement for informed consent was waived due to the retrospective design of this study, which posed minimal risk to participants.

Data Availability Statement

The data presented in this study are available from the corresponding author upon reasonable request. The data are not publicly available due to privacy and ethical restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Vos, T.; Flaxman, A.D.; Naghavi, M.; Lozano, R.; Michaud, C.; Ezzati, M.; Shibuya, K.; Salomon, J.A.; Abdalla, S.; Aboyans, V.; et al. Years lived with disability (YLDs) for 1160 sequelae of 289 diseases and injuries 1990–2010: A systematic analysis for the Global Burden of Disease Study 2010. Lancet 2012, 380, 2163–2196. [Google Scholar] [CrossRef]
  2. Safiri, S.; Kolahi, A.-A.; Smith, E.; Hill, C.; Bettampadi, D.; Mansournia, M.A.; Hoy, D.; Ashrafi-Asgarabad, A.; Sepidarkish, M.; Almasi-Hashiani, A.; et al. Global, regional and national burden of osteoarthritis 1990–2017: A systematic analysis of the Global Burden of Disease Study 2017. Ann. Rheum. Dis. 2020, 79, 819–828. [Google Scholar] [CrossRef]
  3. Cui, A.; Li, H.; Wang, D.; Zhong, J.; Chen, Y.; Lu, H. Global, regional prevalence, incidence and risk factors of knee osteoarthritis in population-based studies. eClinicalMedicine 2020, 29–30, 100587. [Google Scholar] [CrossRef] [PubMed]
  4. Felson, D.T.; Naimark, A.; Anderson, J.; Kazis, L.; Castelli, W.; Meenan, R.F. The prevalence of knee osteoarthritis in the elderly. The Framingham Osteoarthritis Study. Arthritis Rheum. Off. J. Am. Coll. Rheumatol. 1987, 30, 914–918. [Google Scholar] [CrossRef]
  5. Bagge, E.; Bjelle, A.; Valkenburg, H.A.; Svanborg, A. Prevalence of radiographic osteoarthritis in two elderly European populations. Rheumatol. Int. 1992, 12, 33–38. [Google Scholar] [CrossRef] [PubMed]
  6. Maiese, K. Picking a bone with WISP1 (CCN4): New strategies against degenerative joint disease. J. Transl. Sci. 2016, 2, 83–85. [Google Scholar] [CrossRef]
  7. Pal, C.P.; Singh, P.; Chaturvedi, S.; Pruthi, K.K.; Vij, A. Epidemiology of knee osteoarthritis in India and related factors. Indian J. Orthop. 2016, 50, 518–522. [Google Scholar] [CrossRef] [PubMed]
  8. Olsson, S.; Akbarian, E.; Lind, A.; Razavian, A.S.; Gordon, M. Automating classification of osteoarthritis according to Kellgren-Lawrence in the knee using deep learning in an unfiltered adult population. BMC Musculoskelet. Disord. 2021, 22, 844. [Google Scholar] [CrossRef]
  9. Podsiadlo, P.; Wolski, M.; Stachowiak, G.W. Automated selection of trabecular bone regions in knee radiographs. Med. Phys. 2008, 35, 1870–1883. [Google Scholar] [CrossRef]
  10. Chen, P.; Gao, L.; Shi, X.; Allen, K.; Yang, L. Fully automatic knee osteoarthritis severity grading using deep neural networks with a novel ordinal loss. Comput. Med. Imaging Graph. 2019, 75, 84–92. [Google Scholar] [CrossRef]
  11. Xue, Y.; Zhang, R.; Deng, Y.; Chen, K.; Jiang, T. A preliminary examination of the diagnostic value of deep learning in hip osteoarthritis. PLoS ONE 2017, 12, e0178992. [Google Scholar] [CrossRef]
  12. Shin, Y.C.; Kim, S.W.; Chae, D.S.; Yoo, S.K. Analysis of Feature Importance for Knee Osteoarthritis Severity Classification Using Machine Learning. J. Inst. Electron. Inf. Eng. 2020, 57, 99–106. [Google Scholar] [CrossRef]
  13. Swiecicki, A.; Li, N.; O’Donnell, J.; Said, N.; Yang, J.; Mather, R.C.; Jiranek, W.A.; Mazurowski, M.A. Deep learning-based algorithm for assessment of knee osteoarthritis severity in radiographs matches performance of radiologists. Comput. Biol. Med. 2021, 133, 104334. [Google Scholar] [CrossRef]
  14. Mikhaylichenko, A.; Demyanenko, Y. Automatic grading of knee osteoarthritis from plain radiographs using densely connected convolutional networks. In Recent Trends in Analysis of Images, Social Networks and Texts, Proceedings of the 9th International Conference, AIST 2020, Skolkovo, Moscow, Russia, 15–16 October 2020; Revised Supplementary Proceedings 9; Springer International Publishing: Berlin/Heidelberg, Germany, 2021; pp. 149–161. [Google Scholar] [CrossRef]
  15. Tiulpin, A.; Thevenot, J.; Rahtu, E.; Lehenkari, P.; Saarakkala, S. Automatic Knee Osteoarthritis Diagnosis from Plain Radiographs: A Deep Learning-Based Approach. Sci. Rep. 2018, 8, 1727. [Google Scholar] [CrossRef]
  16. Patron, A.; Annala, L.; Lainiala, O.; Paloneva, J.; Äyrämö, S. Method for Automatic Assessment of Spiking of Tibial Tubercles Associated with Knee Osteoarthritis. SSRN 2025, ssrn:4155105. [Google Scholar] [CrossRef]
  17. von Schacky, C.E.; Sohn, J.H.; Liu, F.; Ozhinsky, E.; Jungmann, P.M.; Nardo, L.; Posadzy, M.; Foreman, S.C.; Nevitt, M.C.; Link, T.M.; et al. Development and Validation of a Multitask Deep Learning Model for Severity Grading of Hip Osteoarthritis Features on Radiographs. Radiology 2020, 295, 136–145. [Google Scholar] [CrossRef]
  18. Abdullah, S.S.; Rajasekaran, M.P. Automatic detection and classification of knee osteoarthritis using deep learning approach. Radiol. Med. 2022, 127, 398–406. [Google Scholar] [CrossRef]
  19. Yoon, J.S.; Yon, C.-J.; Lee, D.; Lee, J.J.; Kang, C.H.; Kang, S.-B.; Lee, N.-K.; Chang, C.B. Assessment of a novel deep learning-based software developed for automatic feature extraction and grading of radiographic knee osteoarthritis. BMC Musculoskelet. Disord. 2023, 24, 869. [Google Scholar] [CrossRef]
  20. Gornale, S.S.; Patravali, P.U.; Hiremath, P.S. Early Detection of Osteoarthritis based on Cartilage Thickness in Knee X-ray Images. Int. J. Image Graph. Signal Process. 2019, 11, 56–63. [Google Scholar] [CrossRef]
  21. Subha, B.; Jeyakumar, V.; Deepa, S.N. Gaussian Aquila optimizer based dual convolutional neural networks for identification and grading of osteoarthritis using knee joint images. Sci. Rep. 2024, 14, 7225. [Google Scholar] [CrossRef]
  22. Mohammed, A.S.; Hasanaath, A.A.; Latif, G.; Bashar, A. Knee Osteoarthritis Detection and Severity Classification Using Residual Neural Networks on Preprocessed X-ray Images. Diagnostics 2023, 13, 1380. [Google Scholar] [CrossRef]
  23. Mahum, R.; Irtaza, A.; El-Meligy, M.A.; Sharaf, M.; Tlili, I.; Butt, S.; Mahmood, A.; Awais, M.; El-Sherbeeny, A.M. A Robust Framework for Severity Detection of Knee Osteoarthritis Using an Efficient Deep Learning Model. Int. J. Pattern Recognit. Artif. Intell. 2023, 37, 2352010. [Google Scholar] [CrossRef]
  24. Song, J.; Zhang, R. A novel computer-assisted diagnosis method of knee osteoarthritis based on multivariate information and deep learning model. Digit. Signal Process. 2022, 133, 103863. [Google Scholar] [CrossRef]
  25. Boniatis, I.; Costaridou, L.; Cavouras, D.; Kalatzis, I.; Panagiotopoulos, E.; Panayiotakis, G. Osteoarthritis severity of the hip by computer-aided grading of radiographic images. Med. Biol. Eng. Comput. 2006, 44, 793–803. [Google Scholar] [CrossRef] [PubMed]
  26. Fei, M.; Lu, S.; Chung, J.H.; Hassan, S.; Elsissy, J.; Schneiderman, B.A. Diagnosing the Severity of Knee Osteoarthritis Using Regression Scores from Artificial Intelligence Convolution Neural Networks. Orthopedics 2024, 47, E247–E254. [Google Scholar] [CrossRef]
  27. Hunter, D.J.; Bierma-Zeinstra, S. Osteoarthritis. Lancet 2019, 393, 1745–1759. [Google Scholar] [CrossRef] [PubMed]
  28. Cooper, C.; Cushnaghan, J.; Kirwan, J.; Dieppe, P.; Rogers, J.; McAlindon, T.; McCrae, F. Radiographic assessment of the knee joint in osteoarthritis. Ann. Rheum. Dis. 1992, 51, 80–82. [Google Scholar] [CrossRef]
  29. Wen, H.; Qi, W.; Shuang, L. Medical X-ray image enhancement based on wavelet domain homomorphic filtering and CLAHE. In Proceedings of the 2016 International Conference on Robots & Intelligent System (ICRIS), Zhangjiajie, China, 27–28 August 2016; IEEE: New York, NY, USA, 2016; pp. 249–254. [Google Scholar] [CrossRef]
  30. Sahu, S.; Singh, A.K.; Ghrera, S.; Elhoseny, M. An approach for de-noising and contrast enhancement of retinal fundus image using CLAHE. Opt. Laser Technol. 2019, 110, 87–98. [Google Scholar] [CrossRef]
  31. Muniyappan, S.; Allirani, A.; Saraswathi, S. A novel approach for image enhancement by using contrast limited adaptive histogram equalization method. In Proceedings of the 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT), Tiruchengode, India, 4–6 July 2013; pp. 1–6. [Google Scholar] [CrossRef]
  32. Yadav, G.; Maheshwari, S.; Agarwal, A. Contrast limited adaptive histogram equalization based enhancement for real time video system. In Proceedings of the 2014 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Delhi, India, 24–27 September 2014; IEEE: New York, NY, USA, 2014; pp. 2392–2397. [Google Scholar] [CrossRef]
  33. Dong, K.; Zhou, C.; Ruan, Y.; Li, Y. MobileNetV2 model for image classification. In Proceedings of the 2020 2nd International Conference on Information Technology and Computer Application (ITCA), Guangzhou, China, 18–20 December 2020; IEEE: New York, NY, USA, 2020; pp. 476–480. [Google Scholar] [CrossRef]
  34. Revathi, S.A.; Babu, B.S. Enhanced image classification with integrating DenseNet121 with Mixup augmentation and attention mechanisms for knee OA. In Proceedings of the 2024 Second International Conference on Advanced Computing & Communication Technologies (ICACCTech), Sonipat, India, 16–17 November 2024; IEEE: New York, NY, USA, 2024; pp. 889–894. [Google Scholar] [CrossRef]
  35. Jahan, N.; Anower, M.S.; Hassan, R. Automated diagnosis of pneumonia from classification of chest x-ray images using EfficientNet. In Proceedings of the 2021 International Conference on Information and Communication Technology for Sustainable Development (ICICT4SD), Dhaka, Bangladesh, 27–28 February 2021; IEEE: New York, NY, USA, 2021; pp. 235–239. [Google Scholar] [CrossRef]
  36. Li, Y.P.; Wei, X.C.; Zhou, J.M.; Wei, L. The Age-Related Changes in Cartilage and Osteoarthritis. BioMed Res. Int. 2013, 2013, 916530. [Google Scholar] [CrossRef]
  37. Jung, J.H.; Bang, C.H.; Song, G.G.; Kim, C.; Kim, J.-H.; Choi, S.J. Knee osteoarthritis and menopausal hormone therapy in postmenopausal women: A nationwide cross-sectional study. Menopause 2019, 26, 598–602. [Google Scholar] [CrossRef]
Figure 1. Subchondral labeling and images for each sclerosis grade. The red squares in the image highlight the ROI where subchondral sclerosis is most prominent. Grade 0: no abnormalities in the subchondral bone. Grade 1: mild sclerosis or partial destruction may be observed on the surface of the subchondral bone. Grade 2: significant sclerosis is present, with substantial destruction in multiple areas, potentially impacting joint function. The arrows in the grade 2 image indicate the areas of subchondral sclerosis and OA. These points represent the key lesions characteristic of grade 2 in this study, typically observed in more advanced stages of OA. Notably, subchondral sclerosis appears as white areas in X-ray images.
Figure 1. Subchondral labeling and images for each sclerosis grade. The red squares in the image highlight the ROI where subchondral sclerosis is most prominent. Grade 0: no abnormalities in the subchondral bone. Grade 1: mild sclerosis or partial destruction may be observed on the surface of the subchondral bone. Grade 2: significant sclerosis is present, with substantial destruction in multiple areas, potentially impacting joint function. The arrows in the grade 2 image indicate the areas of subchondral sclerosis and OA. These points represent the key lesions characteristic of grade 2 in this study, typically observed in more advanced stages of OA. Notably, subchondral sclerosis appears as white areas in X-ray images.
Sensors 25 02535 g001
Figure 2. Data preprocessing steps. (a) Convert to 8-bit scale: this step involves adjusting the pixel values of the data to an 8-bit scale, reducing memory usage and facilitating easier data processing and analysis. (b) CROP: the ROI for subchondral sclerosis data is masked and cropped to the size of the relevant area for further use. (c) CLAHE (contrast limited adaptive histogram equalization): enhances image contrast to make the features of joint tissues more prominent. (d) Zero padding: preserves the spatial dimensions of the images to prevent distortion and improve training efficiency.
Figure 2. Data preprocessing steps. (a) Convert to 8-bit scale: this step involves adjusting the pixel values of the data to an 8-bit scale, reducing memory usage and facilitating easier data processing and analysis. (b) CROP: the ROI for subchondral sclerosis data is masked and cropped to the size of the relevant area for further use. (c) CLAHE (contrast limited adaptive histogram equalization): enhances image contrast to make the features of joint tissues more prominent. (d) Zero padding: preserves the spatial dimensions of the images to prevent distortion and improve training efficiency.
Sensors 25 02535 g002
Figure 3. Workflow for subchondral sclerosis grading classification model training. Data collection, data preprocessing, model training and validation, and model evaluation.
Figure 3. Workflow for subchondral sclerosis grading classification model training. Data collection, data preprocessing, model training and validation, and model evaluation.
Sensors 25 02535 g003
Figure 4. Images where the actual grades match the predicted grades classified by the EfficientNetB0 model.
Figure 4. Images where the actual grades match the predicted grades classified by the EfficientNetB0 model.
Sensors 25 02535 g004
Figure 5. Grad-CAM results for randomly selected test dataset images. This visualization highlights the features of joint structures related to subchondral sclerosis. In our visualization, cooler colors such as blue indicate higher importance, while warmer colors such as red and yellow denote lower relevance, according to the color scale applied.
Figure 5. Grad-CAM results for randomly selected test dataset images. This visualization highlights the features of joint structures related to subchondral sclerosis. In our visualization, cooler colors such as blue indicate higher importance, while warmer colors such as red and yellow denote lower relevance, according to the color scale applied.
Sensors 25 02535 g005
Figure 6. ROC curves for multiclass classification models derived from test data. (a) 3-Layer CNN, (b) DenseNet121, (c) MobileNetV2, and (d) EfficientNetB0.
Figure 6. ROC curves for multiclass classification models derived from test data. (a) 3-Layer CNN, (b) DenseNet121, (c) MobileNetV2, and (d) EfficientNetB0.
Sensors 25 02535 g006
Figure 7. Confusion matrices for multiclass classification models derived from test data. (a) 3-Layer CNN, (b) DenseNet121, (c) MobileNetV2, and (d) EfficientNetB0.
Figure 7. Confusion matrices for multiclass classification models derived from test data. (a) 3-Layer CNN, (b) DenseNet121, (c) MobileNetV2, and (d) EfficientNetB0.
Sensors 25 02535 g007
Table 1. Data partitioning of training, testing, and validation subsets for subchondral sclerosis grading.
Table 1. Data partitioning of training, testing, and validation subsets for subchondral sclerosis grading.
Subchondral Sclerosis
Grade
TrainValidationTest
Sample, nRatio, %Sample, nRatio, %Sample, nRatio, %
Grade 0117836.6414235.3215037.31
Grade 1101131.4513734.0813433.33
Grade 2102631.9112330.6011829.36
Total3215100402100402100
Note: This table outlines the sample sizes and ratios for different grades of subchondral sclerosis across the training, validation, and test datasets.
Table 2. Distribution of subchondral sclerosis grades in the test data set, categorized by age group and sex.
Table 2. Distribution of subchondral sclerosis grades in the test data set, categorized by age group and sex.
Grade 0Grade 1Grade 2Total (Men, Women)Proportion (%)
20s4004 (0, 4)1.00
30s8008 (4, 4)2.01
40s151016 (2, 14)4.02
50s4211760 (17, 43)15.10
60s434244129 (21, 108)32.39
70s316854153 (19, 134)38.53
80s6111128 (3, 25)7.04
90s1124 (1, 3)1.00
Total150134118402100
Note: This table presents the distribution of different subchondral sclerosis grades in the test dataset, including breakdown by age and sex.
Table 3. Performance evaluation results of models.
Table 3. Performance evaluation results of models.
ModelSensitivity (%)Specificity (%)Accuracy (%)AUC *p-Value **
3-Layer CNN73.38 ± 1.3887.12 ± 0.7073.81 ± 1.5489.52 ± 0.46<0.05
DenseNet12182.98 ± 1.5591.75 ± 0.7883.31 ± 1.6294.68 ± 0.84
MobileNetV283.41 ± 1.2091.98 ± 0.6183.78 ± 1.2294.45 ± 0.71
EfficientNetB084.27 ± 1.0392.46 ± 0.4984.70 ± 0.9895.17 ± 0.41
* AUC: area under the curve. ** The p-value was calculated using analysis of variance.
Table 4. Tukey’s HSD test results for comparison of model performance.
Table 4. Tukey’s HSD test results for comparison of model performance.
3-Layer CNNDenseNet121MobileNetV2EfficientNetB0
3-Layer CNN1.00<0.001<0.001<0.001
DenseNet121<0.0011.000.87410.1523
MobileNetV2<0.0010.87411.000.4892
EfficientNetB0<0.0010.15230.48921.00
p-value was calculated using Tukey’s HSD test.
Table 5. Comparison of detection accuracy for subchondral bone sclerosis by age group and sex using the EfficientNetB0 model.
Table 5. Comparison of detection accuracy for subchondral bone sclerosis by age group and sex using the EfficientNetB0 model.
CategoryGrade 0Grade 1Grade 2Accuracy (%)p-Value *
DetectedTotalDetectedTotalDetectedTotal
Age GroupGroup A
(20s–50s)
63691012470.77 ± 0.09<0.05
Group B
(60s–90s)
778190122921110.84 ± 0.04
SexGroup C
(Men)
373881212170.78 ± 0.10<0.05
Group D (Women)10311292122841010.84 ± 0.04
* p-value was calculated using the Mann–Whitney U test. The accuracy values are presented as standard deviations, and the p-values indicate statistical significance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, S.-B.; Kim, Y.J.; Jung, J.-Y.; Kim, K.G. Classification of Grades of Subchondral Sclerosis from Knee Radiographic Images Using Artificial Intelligence. Sensors 2025, 25, 2535. https://doi.org/10.3390/s25082535

AMA Style

Kim S-B, Kim YJ, Jung J-Y, Kim KG. Classification of Grades of Subchondral Sclerosis from Knee Radiographic Images Using Artificial Intelligence. Sensors. 2025; 25(8):2535. https://doi.org/10.3390/s25082535

Chicago/Turabian Style

Kim, Soo-Been, Young Jae Kim, Joon-Yong Jung, and Kwang Gi Kim. 2025. "Classification of Grades of Subchondral Sclerosis from Knee Radiographic Images Using Artificial Intelligence" Sensors 25, no. 8: 2535. https://doi.org/10.3390/s25082535

APA Style

Kim, S.-B., Kim, Y. J., Jung, J.-Y., & Kim, K. G. (2025). Classification of Grades of Subchondral Sclerosis from Knee Radiographic Images Using Artificial Intelligence. Sensors, 25(8), 2535. https://doi.org/10.3390/s25082535

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop