Next Article in Journal
Patellar Tendon Elasticity and Temperature Following after a 448 Kilohertz Radiofrequency Intervention on Active Healthy Subjects: An Open Controlled Clinical Trial
Next Article in Special Issue
Regional Variations in the Intra- and Intervertebral Trabecular Microarchitecture of the Osteoporotic Axial Skeleton with Reference to the Direction of Puncture
Previous Article in Journal
Temporal Stability of Ciliary Beating Post Nasal Brushing, Modulated by Storage Temperature
Previous Article in Special Issue
Delta Cord as a Radiological Localization Sign of Postoperative Adhesive Arachnoiditis: A Case Report and Literature Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Nerve Root Compression Analysis to Find Lumbar Spine Stenosis on MRI Using CNN

1
Department of Information Technology, University of Gujrat, Gujrat 50700, Pakistan
2
Department of Computer Science, University of Gujrat, Gujrat 50700, Pakistan
3
Universidad Europea del Atlántico, Isabel Torres 21, 39011 Santander, Spain
4
Universidad Internacional Iberoamericana, Campeche 24560, Mexico
5
Universidad Internacional do Cuanza, Cuito EN250, Bié, Angola
6
Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Republic of Korea
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(18), 2975; https://doi.org/10.3390/diagnostics13182975
Submission received: 21 August 2023 / Revised: 14 September 2023 / Accepted: 14 September 2023 / Published: 18 September 2023
(This article belongs to the Special Issue Advances in Spine Imaging)

Abstract

:
Lumbar spine stenosis (LSS) is caused by low back pain that exerts pressure on the nerves in the spine. Detecting LSS is a significantly important yet difficult task. It is detected by analyzing the area of the anteroposterior diameter of the patient’s lumbar spine. Currently, the versatility and accuracy of LSS segmentation algorithms are limited. The objective of this research is to use magnetic resonance imaging (MRI) to automatically categorize LSS. This study presents a convolutional neural network (CNN)-based method to detect LSS using MRI images. Radiological grading is performed on a publicly available dataset. Four regions of interest (ROIs) are determined to diagnose LSS with normal, mild, moderate, and severe gradings. The experiments are performed on 1545 axial-view MRI images. Furthermore, two datasets—multi-ROI and single-ROI—are created. For training and testing, an 80:20 ratio of randomly selected labeled datasets is used, with fivefold cross-validation. The results of the proposed model reveal a 97.01% accuracy for multi-ROI and 97.71% accuracy for single-ROI. The proposed computer-aided diagnosis approach can significantly improve diagnostic accuracy in everyday clinical workflows to assist medical experts in decision making. The proposed CNN-based MRI image segmentation approach shows its efficacy on a variety of datasets. Results are compared to existing state-of-the-art studies, indicating the superior performance of the proposed approach.

1. Introduction

Lumbar spine stenosis (LSS) is a severe back pain disease caused by a degenerative process that compresses the spinal cord and exiting nerve roots, also known as central and foraminal or lateral stenosis [1]. The lumbar contains five vertebrae labeled L1 to L5, the different areas of which may be affected by chronic low back pain (CLBP). CLBP can be caused by a number of factors, including fractures, lumbar disc degeneration, lumbar disc herniation, or infection of the nerve roots. Any of the factors listed above might lead to LSS; however, CLBP is a general term used to define the LSS cause [2]. CLBP negatively affects the health of millions of people around the globe, severely disturbing and employment, personal, and social lives [3]. According to statistics reported in [4], 50% to 80% of adults experience LBP at some point in their lives, representing the most prevalent illness in the world.
CLBP is a frequent consequence of a less serious but acute type of LBP. If the underlying cause is significant and ignored, LBP can develop from acute to chronic status [5]. Acute back pain lasts for a few days to weeks and can be resolved by self-care, as it can last for 12 weeks or longer [6]. According to a National Health Service (NHS) study of the economic impact of CLBP in the United Kingdom (UK) [7], the probability of a patient’s rehabilitation success is primarily based on a timely diagnosis of the LBP cause. Verbiest [8] proposed the term neurogenic claudication to characterize the symptoms that individuals with CLBP experience. Numbness, coldness, burning, and cramping are a few symptoms of the ailment. LBP might begin in the buttocks and extend to the thigh and leg. LSS is a narrowing of the spinal column or vertebral foramina that puts pressure on the thecal sac and posterior nerve roots, either directly or indirectly [9].
An LSS diagnosis is usually carried out utilizing imaging modalities including magnetic resonance imaging (MRI), radiographic myelography (RM) scans, and intravascular ultrasound (IVUS). X-rays can only show the bony part of the spine; soft tissue structures such as the intervertebral discs, muscles around the spine, spinal cord, and spinal nerves cannot be directly evaluated by X-rays alone. MRI produces multiple images of different views that can help provide clearer pictures for the doctor to analyze. Another approach for computer-aided medical diagnosis is the use of IVUS images, which has been well recognized as a powerful imaging technique to evaluate stenosis inside the coronary arteries. These approaches can not only relieve the burden on radiologists but also increase the certainty of a precise diagnosis. MRI is much more common [10] in hospitals than other imaging modalities such as RM and IVUS. MRI is more ubiquitous, since it is the only technique for back pain diagnosis [11], while RM detects the most challenging cases of LSS. A lumbar spine MRI of a patient can be viewed in two ways: from a sagittal (side) or axial (top-down) view [12].
A major source of concern is the fact that in the previous decade, there has been a significant shortage of neuroradiologists [13]. As a result, early disease detection may not always be achievable, since diagnosis may take several weeks due to the need to obtain a referral and the time it takes to see a medical specialist and perform medical scans and analysis. A report from the Royal College of Radiologists [14] showed that three-quarters of UK medical imaging departments lack radiology specialists to provide healthcare services, resulting in increased spending on outsourcing, overtime, and expert doctors to substitute radiologist duties each year. Since 1995, the demand for radiographic imaging, including MRI and computed tomography (CT) scans, has increased at an average annual rate of 12.3%. As a result, improved methods for computer-aided diagnosis (CAD) systems are required to extract diagnostic results from MRI and other modalities. Due to a lack of time to study a case, a radiologist may overlook a disease that a CAD system may detect. This insight drives us to improve the efficiency and effectiveness of automated lumbar abnormality detection systems.
Artificial-intelligence-based approaches have achieved great success in different domains, like image processing, image segmentation, text analysis, etc. [15,16]. In particular, medical image analysis, disease prognosis and detection, and medical data analysis have witnessed a great deal of success using such methods [17,18,19]. Machine learning and deep learning models have been widely adopted for disease detection and prediction. Several CAD approaches have been presented in the past few years for LSS diagnosis [20,21,22]. These studies show that convolutional neural networks (CNNs), which are particularly good at processing image data, show promising results for LSS detection. CNN models are widely applied in combination with medical imaging modalities such as MRI and CT, with a high success rate. However, current LSS detection models suffer from low accuracy and require further investigation and improvement. In this regard, this study makes the following contributions.
  • This study aims to automatically diagnose the types of foraminal LSS by analyzing axial MRI images. In this regard, a custom CNN architecture is proposed.
  • For automated spine labeling and foraminal stenosis grading, this method is trained using large-scale data.
  • Multi-input data are generated for single-region and multiregion LSS classification. Multiregion data include normal, mild, moderate, and severe classes.
  • Assessment of the proposed model is carried out using quantitative and qualitative methods on a publicly available MRI dataset. Afterward, the results are compared to those of existing approaches.
In this study, Section 2 presents the proposed approach, a description of the dataset used in this study, and the details of the proposed CNN model. Experimental results and discussions are presented in Section 3. Conclusions are presented in Section 4.

2. Materials and Methods

This section describes the proposed approach for lumbar spine stenosis detection, the proposed CNN model used for experiments, the dataset used in the study, and performance evaluation metrics.

2.1. Overview of Proposed System

The proposed system uses a deep learning approach for feature extraction that classifies the foraminal or lateral stenosis into normal, mild, moderate, and severe spine stenosis using a CNN, as presented in Figure 1. Several image preprocessing techniques are also used to improve the model’s performance.

2.1.1. Description of Lumbar Spine MR Image Dataset

The dataset [23] used in this study is associated with a clinical study of 515 patients suffering from back pain symptoms. The data were collected and analyzed by experienced radiologists and contain notes regarding the analyzed characteristics and conditions of the lower back, as well as the presence of disease. The data of each patient data are associated with one or more MRI studies. Each study contains slices (i.e., individual images taken from a either sagittal or axial view) of the lowest three vertebrae and the three lowest IVDs. The axial view slices are mainly taken from these last three IVDs. In most cases, the total number of slices in the axial view ranges from 12 to 15 [10]. The best images were extracted from the three lowest vertebrae in the axial view of IVDs, like L3–D3, L4–D4, and L5–D5. The image resolution is 320 × 320 pixels. Image pixels have a precision of 12 bits per pixel, but grayscale pixels have a precision of 8 bits per pixel, which is greater. Further grading was performed by an expert radiologist based on the clinical notes provided within the dataset; MRI lumbar spine stenosis scans were evaluated as normal, mild, moderate, or severe. This evaluation was based on the observer’s characteristics, as well as the condition of the lumbar spine and the presence of disease, such as bone marrow disease, endplate degeneration, IVD bulges, TS compression, central vs. FS, annular tears, scoliosis, endplate defects, facet joint, and ligamentum flavum hypertrophy and spondylosis.

2.1.2. Region of Interest Extraction

To extract the desired region, a unique identification (ID) was allocated to each region, such as 1 (IVD), 2 (PE), 3 (TS), and 4 (AAP), as described in [9]. These regions were extracted by their associated IDs. Two datasets were created from a single dataset, each with a distinct region based on its ID. In the first dataset, IVD, PE, AAP, and TS were chosen, whereas, in the second dataset, only single-ROI (AAP) was extracted. LSS causes AAP compression, which puts pressure on the central spinal canal or nerve roots because stenosis can occur anywhere along the AAP. A general practitioner primarily evaluates three distances in the AAP to diagnose LSS: the AAP diameter and the left and right foramen widths.

2.1.3. MR Image Cropping

Cropping a significant ROI of each image is a method used as a processing step for image data with both height and width dimensions. Additionally, random cropping is used to minimize the size of the input. The computation time is heavily influenced by the image resolution. The resolution of the available dataset is 320 × 320 pixels. The images were cropped with a resolution of 240 × 240 pixels. Despite the lack of considerable difference in image size, the computing time was significantly impacted by the reduced size. The primary images were cropped to a uniform image while maintaining their aspect ratio.

2.1.4. Dataset Augmentation Techniques

Deep learning models require large datasets to reduce the possibilities of overfitting and provide better results [24,25]. Medical image analysis domains, on the other hand, do not have access to such big datasets. Consequently, depending on the need to expand the amount of data, different augmentation techniques have been used in the existing literature [26,27,28]. In this study, the size of the training dataset was increased using these techniques.
The left, right, top, or bottom translation of images is chosen to prevent positional bias in the data while maintaining the image dimensions. Translation was used on the training dataset, whereas the testing dataset was solely made up of the original images. Zoom augmentation casually enlarges the image and adds new pixel values to the surrounding area. The zoom range is [1 − value, 1 + value], which means that if the range is 0.9 to 0.7, the image is 90 to 70% zoomed-in, and if the range is 1.1 to 1.3, the image is 110 to 130% zoomed-out. For random rotation augmentation, the images are rotated right or left on an axis of 1 to 359 . The rotation degree parameter, which is set between 1 and 20 or −1 to −20, has a remarkable effect on the safety of rotation augmentations.

2.1.5. Architecture of the Proposed CNN Model

CNN is the most well-known and widely used type of algorithm in the field of deep learning, particularly for image processing [29,30,31,32]. A CNN automatically finds significant features without the need for human intervention. A CNN is a multilayered model with several convolution layers leading to subsampling (pooling) layers and fully connected layers at the end. Figure 2 shows the architecture of the proposed CNN for image categorization.
The first layer of the proposed architecture is a convolution layer that detects and extracts patterns and features from an input image. It keeps the pixels together by learning image patterns via small squares of input data. In a CNN model, the input (x) of each layer is structured in three aspects: height, width, and depth (or m × m × r ), where the height (m) equals the width of 240 × 240 pixels. The channel number is another term for depth. The depth (r) of a grayscale image is set as 1. Each convolutional layer comprises a number of feature maps (filters) that are designated by k, with three dimensions ( n × n × q ), related to the input image.
Filters learn patterns such as edge detection, texture, corners, blur, and sharpening. In the next step, the stride shifts the number of pixels per step on the input matrix. Here, the stride (1,1) passes the filters as 1 pixel gradually. The model padding parameter is set as ‘same’. If the filter occasionally does not fit well with the input image, then there is the choice of padding so that it fits according to the requirement; otherwise valid padding drops, are used for the part where it does not fit perfectly.
The convolution and pooling layer is accountable for reducing the spatial size of the features convolved by applying a filter that also decreases the computational requirement to process the data by reducing dimensionality. Furthermore, it is useful for extracting features, since augmentation techniques are applied to data that have been rotated and shifted around without losing information. Furthermore, it maintains the process of efficient model training. The rectified linear unit (ReLU) activation function is utilized to map the input to the output. ReLU is the most-often-used function in the CNN context, since it reduces the model’s computational complexity.
In a fully connected layer that is located at the end of the CNN architecture, the matrix is turned into a vector, similar to a neural network. We linked these features together to form a model, since each neuron in the input layer is connected to the neurons in the output layer. Ultimately, we used a softmax activation function to categorize the outputs as normal, mild, moderate, or severe. Complete architectural details of the proposed CNN model are presented in Table 1.

2.2. Performance Evaluation

A range of evaluation criteria was employed to examine how well the model performed on augmented and non-augmented datasets with varied data values. These measures include class-specific metrics, as well as performance metrics, like accuracy, precision, sensitivity, and F1 score, which provide us with specific values to compare the algorithm efficiently. The following equations were used for these metrics:
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1 s c o r e = 2 P r e c i s i o n R e c a l l P r e c i s i o n + R e c a l l
where TP, TN, FP, and FN refer to true positive, true negative, false positive, and false negative, respectively.

3. Results, Analysis, and Discussion

This section provides details of the experimental setup, results of the proposed approach, and discussions of the results.

3.1. Experimental Setup

The proposed model was evaluated on a system with a Windows 10 operating system, 8 GB of RAM, and a Core i5 (3.6 GHz) CPU. Google Colab Pro was acquired with 38 GB of RAM and TPU processing. Python programming was utilized for experiments. The training took an average of 810 s (3 s/steps) for each image.

3.2. Data Preparation

The initial step in using deep learning models is to prepare the training data for the classifiers. The deep learning approach is tremendously data-hungry because it also incorporates representation learning [33]. Multi-ROI and single-ROI training datasets were employed in our experiment. The first mask shown in Figure 3, contained IVD, PE, and TS, but the second mask only included the AAP region, so stenosis diagnosis could be performed after measuring this region.
To improve model performance, training was performed using various augmented sets of data including 5 k, 10 k, and 12.5 k, and testing was performed on 330 labeled images. The training was carried out using an 80:20 ratio; the classifier was trained on 80% and tested on 20% of the data. The training and testing ratio was chosen randomly, which was found to be most effective in prior research [10].
The model takes modest steps to reduce the negative gradient of the loss function, which is specified as the categorical cross-entropy probability distribution of each class. The learning rate parameter, which is 0.001, alters the step size, and the Adam optimizer is utilized. In the 12.5 k dataset, the batch size utilized for training was 256 per image, for a total of 34 epochs. A batch size of 128 with 34 epochs was applied in the 10 k dataset, with a batch size of 64 with 200 epochs in the case of the 5 k dataset.
The proposed model outperformed the compared models because its pruning neural networks reduced computational complexity and training inputs to some extent. The model was trained using 34 epochs instead of 100 epochs with a GPU, as the model took less than 12 h to complete the training process and produce the results, whereas prior models [34] took two or more days to complete the training process. After these epochs, the model showed no improvement in performance.

3.3. Results Using 12.5 K Dataset

Plots of the model’s training and testing accuracy plot are shown in Figure 4; the model was trained for 34 epochs for both multi-ROI and single-ROI datasets. The model learns rapidly as training data are fed into it, and the training curve steadily increases until all epochs are completed. As epochs increase, the validation accuracy is increased in lockstep with the training accuracy curve. To compute model loss from the plot, the model travels through the same epochs. The model’s validation loss is significant at the start of the epochs, but as the number of epochs increases, the loss decreases.
Table 2 demonstrates multi-ROI and single-ROI classification reports of classes using the CNN, with precision of 0.94, recall of 0.92, and F1 score of 0.93 for the mild class. Precision and recall for the moderate class are 0.96 and 0.98, respectively. Precision for the normal class is recorded as 1.00. The average multiclass accuracy of the model for the multi-ROI dataset is 0.97. The macro average and weighed average were also calculated for precision, recall, and F1 score as 0.97 each for multi-ROI data.
The other half of Table 2 shows precision and recall values of 0.96 for the moderate class and precision of 0.97 for the normal class. The F1 score and recall for the normal class and F1 score for severe class are both 0.98. For the mild class, the precision, recall, and F1 score are each 0.99. The overall accuracy score of the model for all single-ROI is 0.97.

3.4. Results Using 10 K Dataset

The CNN accuracy plot in Figure 5 shows the training process of the model for 34 epochs for both the multi-ROI and single-ROI datasets. The model learns gradually as training progresses, although its accuracy graphs in both datasets exhibit variance towards the conclusion. Multi-ROI data training and validation accuracy fluctuate, whereas the accuracy of single-ROI training and validation increases steadily as the number of epochs increases. The model traverses the same epochs to compute model loss from the plot. The validation loss of the model was initially substantial, but as the number of epochs increased, the quantity of loss dropped. The results of the proposed model are varied in comparison to previously reported results.
Table 3 displays multi-ROI and single-ROI classification reports of classes using the CNN, which shows a precision of 0.92, recall of 0.96, and F1 score of 0.94, for the moderate class using multi-ROI data. The precision for the severe class is 0.94, while the F1 score for the mild and moderate classes are also 0.94. The precision, recall, and F1 score for the normal class are 0.98, 1.00 and 0.99, respectively. The average accuracy for the multi-ROI data using 10 k data is 0.96, which is slightly lower than that achieved using 12.5 k data.
The weighted average for precision, recall, and F1 scores is 0.98 each for single-ROI data using 10 k data. The precision for the normal class is 0.97, while the recall and F1 scores for the severe class and F1 score for normal class is 0.98. For the moderate class, precision, recall, and F1 scores are 0.96 each. The CNN model achieves a 0.98 accuracy score for a single-ROI dataset with 10 k data.

3.5. Results Using 5 K Dataset

Figure 6 shows the training and validation accuracy and loss for the models using 200 epochs for both multi-ROI and single-ROI datasets. In this scenario, the model yields the lowest results compared to the other datasets. The model learns slowly as training proceeds, with the number of epochs increasing. With the increase in epochs, the training accuracy outperforms validation accuracy. Both datasets have a validation accuracy of 86.30% for multi-ROI and 93.15% for single-ROI. To depict the model loss, it went through the same number of epochs. The validation loss of the model was initially significant, but as the number of epochs increased, the amount of loss decreased slightly.
Table 4 illustrates the multi-ROI and single-ROI classification report produced by the CNN model, yielding a precision of 0.74 and an F1 score of 0.79 for the moderate class. The precision, recall, and F1 scores for the mild class are all 0.81, whereas those for normal and moderate classes are 0.84 and 0.85, respectively. The precision and recall for severe and normal classes are 0.86, 0.89, and 0.96 and 0.84, respectively.
Results for the single-ROI dataset indicate that for the normal and moderate classes, the recall and precision score is 0.86. The precision for the severe class is 0.89, while the recall and F1 scores are 0.99 and 0.94, respectively. In terms of accuracy, the CNN model achieves 0.86 and 0.92 accuracy scores for all multi-ROI and single-ROI classes, respectively. This performance is substantially lower compared to results achieved using a 12.5 k dataset.

3.6. Performance Comparison Using Different Dataset Sizes

Precision, recall, F1 score, and accuracy were calculated to compare the results of different dataset values, as presented in Table 5. For LSS detection, superior performance is obtained when a 12.5 k dataset is used with the proposed CNN model, achieving an accuracy score of 0.97, which is superior to that achieved when using both 10 k and 5 k datasets.
Figure 7 provides a visual illustration of the performance comparison of the proposed CNN model using different dataset sizes. It can be observed that using the 5 k dataset, the model shows significantly low performance compared to the 10 k and 12.5 k datasets. The achieved performance slightly differs between the 10 k and 12.5 k datasets, and the best performance is obtained using the 12.5 k dataset.

3.7. Performance Comparison for Augmented vs. Non-Augmented Data

Deep learning models are data-intensive and require significantly larger datasets to obtain complex features to achieve improved performance. The original dataset size was smaller, and the CNN model could not be trained well to achieve satisfactory accuracy for LSS detection. Consequently, we performed data augmentation to resolve this issue.
Table 6 and Figure 8 show comparisons of the results obtained using the original dataset with those achieved using the augmented dataset. When original data was used, the evaluation accuracy was relatively low, whereas, when augmented data were utilized, the results were much better. Comparably, the single-ROI results are better than those obtained with multi-ROI dataset.

3.8. Performance with Existing Approaches

In order to further advocating for the performance of the proposed model, a performance comparison with existing state-of-the-art approaches was also carried out. For this purpose, several approaches reported in existing literature were selected. For example, the methods proposed in [22,35] both use a CNN model for LSS detection, obtaining accuracies of 87.75% and 84.5%, respectively. Pretrained models have also been deployed, such as ResNet in [36] and VGG16 in [37]. Higher performance was reported in [38], with 94% accuracy, while the authors of [39] recently reported 95% accuracy. The results comparison presented in Table 7 indicates that the proposed model outperforms existing approaches.

3.9. Discussion

Over the last decade, several CAD approaches [41,42,43,44,45] have been investigated for their potential to address the challenges of spinal MRI interpretation and full automation the LSS diagnostic procedure, which could help to improve detection accuracy. In this regard, the CNN model is often employed with medical imaging modalities such as MRI and computed tomography (CT), with a high success rate. Several previously proposed approaches for neural foraminal stenosis disease detection using binary and multigraded (normal, mild, moderate, and severe) classification are discussed herein.
Among the most current diagnostic frameworks for LSS is that proposed by Natalia et al. [20], who used the SegNet model to automatically assess the area between the anterior and posterior (AAP) diameter and foraminal widths in MRI-, T1-, and T2-weighted composite images. Six ROIs were extracted after semantic segmentation, including intervertebral disc (IVD), posterior element (PE), and thecal sac (TS), as well as auxiliary ROIs, such as AAP and others. The contour evaluation technique was used to increase the accuracy of the segmentation result in specified ROIs. The results demonstrate a 96.7% diameter agreement with the expert. Similarly, Sartoretti’s classification [21] is based on a six-point grading system for detecting lumbar foraminal stenosis (FS) on MRI images of high resolution. Grade A has no FS. The superior, posterior, inferior, and anterior boundaries of the lumbar foramen are graded B, C, D, and E, respectively, indicating nerve root contact with surrounding anatomical structures. The existence of FS in the nerve root with morphological changes was graded F in this research, in which we employed sagittal high-resolution T1-weighted and T2-weighted MRI data from 101 subjects.
A study in regard to grading of CAD systems by Salehi et al. [22] showed that a CNN can be utilized to diagnose disc herniation using MRI images. A performance evaluation was carried out for normal, bulge, protrusion, and extrusion images. The experiment was performed on 2329 axial-view lumbar MRI datasets collected from a local medical center. Experimental results reported an 87.75% accuracy with data augmentation. Lu et al. [38] used the U-Net architecture of the CNN model to grade central and FS as normal, mild, moderate, or severe based on both sagittal and axial MRI images. A large-scale dataset of 22,796 was used, which included data from 4075 patients. An accuracy of 94% was reported for this study.
A different technique proposed by Han et al. [40] localizes six vertebrae and disc T12 to S1 using a deep multiscale multitask learning network (DMML-Net) that integrated into a full convolution network that grades the lumbar neural FS into normal and abnormal cases. The experimental setup included a dataset comprising 200 T1- and T2-weighted MRI images from 200 patients, achieving an accuracy of 84.5% using the proposed approach. An approach recently proposed by Hallinan et al. [35] is to classify neural foraminal stenosis into normal, mild, moderate, or severe classes using a deep learning CNN model that achieved 84.5% accuracy using a dataset of T2-weighted axial MRI images and T1-weighted sagittal MRI images from 446 patients.
Using a deep learning ResNet-50 model, multitask classification was performed in [36], which demonstrated the automated grading of lumbar disc herniation (LDH), lumbar central canal stenosis (LCCS), and lumbar nerve roots compression (LNRC) in lumbar axial MRIs. An internal test dataset and an external test dataset were used for classification systems with four graded levels (grade 0, grade 1, grade 2, and grade 3). A total of 1115 patients (1015 patients from the internal dataset and 100 patients from the external test dataset) were evaluated, and the best MRI slices were obtained. The efficiency of the model on the given datasets was evaluated using precision, accuracy, sensitivity, specificity, F1 scores, confusion matrices, receiver operating characteristics, and inter-rater agreement (Gwet k). On the internal test dataset, the overall grading accuracy for LDH, LCCS, and LNRC were 84.17%, 86.99%, and 81.21%, respectively. For the external test data, 74.16%, 79.65%, and 81.21% accuracy are reported for LDH, LCCS, and LNRC, respectively.
Bharadwaj et al. [46] utilized a V-Net model to segment the dural sac and IVD and localize the facet and foramen. Big transfer (BiT) models were trained for classification tasks. Multievaluation metrics including Cohen’s Kappa score were used for the dural sac and IVD. The authors used axial T2-weighted MRI images of the lumbar spine obtained between 2008 and 2019. The area under the receiver operator characteristic curve (AUROC) values used for the binary classification of facet and neural foraminal stenosis were 0.92 and 0.93, respectively. Sinan et al. [37] proposed an LSS-VGG16 and U-Net model that detects LSS in MR and CT images and achieved 87.70% classification accuracy on VGG16. A total of 1560 MR images were used with U-Net, with a 0.93 DICE score.
The authors of [47] a 3D LSS segmentation framework that enables the complete determination of the regions of the body that cannot be fully opened during LSS surgeries, particularly in the nerve roots. The spinal disc, canal, thecal sac, posterior element, and other regions and backgrounds in the image that are crucial for LSS were all segmented and divided into a total of six classes in MRI images. The intersection over union (IoU) metric was deployed for each class to assess the success of segmentation, since the canal had an IoU value of 0.61. The study employed T2 sequence lumbar MRI images of 300 LSS patients in the digital imaging and communications in medicine (DICOM) format.
Abhinav et al. [48] also recently presented a U-Net-dependent CNN model to segment the IVD, PE, TS, and AAP regions of LSS on an axial MRI dataset [10] and performed binary classification. The performance of the model was evaluated by IoU metrics. Since IVD is the simplest region to label and PE has a particular shape that resembles the letter Y, the values of regions like IVD, PE, and IoU vary between 0.80 and 1.0. And because AAP was the most challenging to identify, its IoU metric value is 0.6568, which is lower than that of the other regions.
Another innovative study [39] compared conventional and ultrafast methods and analyzed sagittal T1-weighted, T2-weighted, short-TI inversion recovery, and axial T2-weighted MRI images of 58 patients. Cohen’s kappa metrics were used to assess foraminal stenosis in axial images, and the results were provided in a multigraded classification. The accuracy obtained using this method was 95%.
In this study, we investigated LSS detection using a customized CNN model. We evaluated the algorithm’s performance using a variety of metrics. Experiments were conducted using two datasets and with and without augmentation techniques using different data values. Multi-ROI and single-ROI datasets with 5 k achieved the lowest results in terms of accuracy scores: 0.85 and 0.92, respectively. The cure of model accuracy shows that the model could be trained more to prevent underfitting and inflection because the model was not overlearned for the training set. Due to inadequate training, the model loss exhibits a divergence from the training curve, which indicates why the overall loss is large in the results of both datasets.
The two 10 k datasets achieved accuracy scores of 0.96 and 0.98, respectively. Figure 6 demonstrates that the trained model fit well; however, the validation curve is slightly unsatisfactory, owing to underfitting for the multi-ROI dataset, requiring more training data samples to improve accuracy. The model loss curve shows that training significantly decreased the loss, although it remains high during the initial epochs.
Table 5 illustrates that the accuracy is similar between the two 12.5 k datasets: 0.97 and 0.98, respectively. The model accuracy curve show in Figure 4 indicates that while model training performs well on a single-ROI dataset, results on a multi-ROI dataset might be further enhanced by adding more training data and by further reducing model loss.
Table 8 shows an analytical summary of the discussed research works. It can be observed that for LSS detection and segmentation, the models suffer from low accuracy. The CNN model and its variants were tested, yet no CNN technique was able to more accurately categorize LSS disease, necessitating the development of automatic methods that better classify the disease.

4. Conclusions

In this study, we proposed a technique to assist doctors in detecting and grading neural foraminal LSS using MRI images. Four regions of IVD, PE, TS, and AAP were selected as areas of focus in this study. For LSS detection, we proposed a customized CNN model and performed experiments using a publicly available lumbar spine dataset. The dataset consists of a back pain characteristics report annotated by expert radiologists and 515 patient MRI scan images of L3–L5 in axial view. Observed characteristics in annotated studies were further classified as normal, mild, moderate, or severe to investigate the reliability of the proposed deep-learning-based stenosis grading system. We constructed two datasets—a multi-ROI and single-ROI dataset—then trained the model on a variety of dataset values. Experiments were conducted using 5 k, 10 k, and 12.5 k datasets, which were produced using data augmentation. Experimental results indicate that better performance can be obtained using an augmented dataset. The best performance was achieved with a 12.5 k dataset for both single-ROI and multi-ROI datasets, showing 97.71% and 97.01% accuracy, respectively. Performance comparison with existing state-of-the-art approaches validated the superior performance of the proposed approach.

Author Contributions

Conceptualization, T.S. and M.U.A.; Data curation, M.U.A. and F.M.; Formal analysis, T.S. and F.M.; Funding acquisition, R.M.D.; Investigation, M.U.S. and R.M.D.; Methodology, F.M.; Project administration, M.U.S.; Software, M.U.S. and M.A.S.; Supervision, I.A.; Validation, M.A.S. and I.A.; Visualization, R.M.D. and M.A.S.; Writing—original draft, T.S. and M.U.A.; Writing—review and editing, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the European the University of Atlantic.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Raja, A.; Hoang, S.; Patel, P.; Mesfin, F.B. Spinal Stenosis; StatPearls Publishing: Treasure Island, FL, USA, 2017. [Google Scholar]
  2. Al Kafri, A.S.; Sudirman, S.; Hussain, A.J.; Fergus, P.; Al-Jumeily, D.; Al-Jumaily, M.; Al-Askar, H. A framework on a computer assisted and systematic methodology for detection of chronic lower back pain using artificial intelligence and computer graphics technologies. In Proceedings of the Intelligent Computing Theories and Application: 12th International Conference, ICIC 2016, Lanzhou, China, 2–5 August 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 843–854. [Google Scholar]
  3. Hoy, D.; March, L.; Woolf, A.; Blyth, F.; Brooks, P.; Smith, E.; Vos, T.; Barendregt, J.; Blore, J.; Murray, C.; et al. The global burden of neck pain: Estimates from the Global Burden of Disease 2010 study. Ann. Rheum. Dis. 2014, 73, 1309–1315. [Google Scholar] [CrossRef]
  4. Fatoye, F.; Gebrye, T.; Odeyemi, I. Real-world incidence and prevalence of low back pain using routinely collected data. Rheumatol. Int. 2019, 39, 619–626. [Google Scholar] [CrossRef] [PubMed]
  5. Wu, A.; March, L.; Zheng, X.; Huang, J.; Wang, X.; Zhao, J.; Blyth, F.M.; Smith, E.; Buchbinder, R.; Hoy, D. Global low back pain prevalence and years lived with disability from 1990 to 2017: Estimates from the Global Burden of Disease Study 2017. Ann. Transl. Med. 2020, 8. [Google Scholar] [CrossRef] [PubMed]
  6. Hoy, D.; Bain, C.; Williams, G.; March, L.; Brooks, P.; Blyth, F.; Woolf, A.; Vos, T.; Buchbinder, R. A systematic review of the global prevalence of low back pain. Arthritis Rheum. 2012, 64, 2028–2037. [Google Scholar] [CrossRef] [PubMed]
  7. Versus Arthritis. The State of Musculoskeletal Health. 2019. Available online: https://www.versusarthritis.org/about-arthritis/data-and-statistics/the-state-of-musculoskeletal-health/ (accessed on 22 August 2023).
  8. Verbiest, H. Neurogenic Intermittent Claudication: With Special Reference to Stenosis of the Lumbar Vertebral Canal; Elsevier: Amsterdam, The Netherlands, 1976. [Google Scholar]
  9. Natalia, F.; Meidia, H.; Afriliana, N.; Al-Kafri, A.; Sudirman, S. Methodology to Determine Important-Points Location for Automated Lumbar Spine Stenosis Diagnosis Procedure. In Proceedings of the 2019 International Conference on Intelligent Medicine and Health, Ningbo, China, 1–3 July 2019; pp. 53–57. [Google Scholar] [CrossRef]
  10. Al-Kafri, A.S.; Sudirman, S.; Hussain, A.; Al-Jumeily, D.; Natalia, F.; Meidia, H.; Afriliana, N.; Al-Rashdan, W.; Bashtawi, M.; Al-Jumaily, M. Boundary delineation of MRI images for lumbar spinal stenosis detection through semantic segmentation using deep neural networks. IEEE Access 2019, 7, 43487–43501. [Google Scholar] [CrossRef]
  11. Savigny, P.; Kuntze, S.; Watson, P.; Underwood, M.; Ritchie, G.; Cotterell, M.; Hill, D.; Browne, N.; Buchanan, E.; Coffey, P.; et al. Low Back Pain: Early Management of Persistent Non-Specific Low Back Pain; National Collaborating Centre for Primary Care and Royal College of General Practitioners: London, UK, 2009; Volume 14, pp. 9–13. [Google Scholar]
  12. Nguyen, H.S.; Doan, N.; Shabani, S.; Baisden, J.; Wolfla, C.; Paskoff, G.; Shender, B.; Stemper, B. Upright magnetic resonance imaging of the lumbar spine: Back pain and radiculopathy. J. Craniovertebral Junction Spine 2016, 7, 31. [Google Scholar]
  13. Bhargavan, M.; Sunshine, J.H.; Schepps, B. Too few radiologists? Am. J. Roentgenol. 2002, 178, 1075–1082. [Google Scholar] [CrossRef]
  14. Royal College of Radiologists. Clinical radiology UK Workforce Census 2015 Report. 2016. Available online: https://www.rcr.ac.uk/publication/clinical-radiology-uk-workforce-census-2015-report (accessed on 22 August 2023).
  15. Reshi, A.A.; Ashraf, I.; Rustam, F.; Shahzad, H.F.; Mehmood, A.; Choi, G.S. Diagnosis of vertebral column pathologies using concatenated resampling with machine learning algorithms. PeerJ Comput. Sci. 2021, 7, e547. [Google Scholar] [CrossRef]
  16. Jalal, N.; Mehmood, A.; Choi, G.S.; Ashraf, I. A novel improved random forest for text classification using feature ranking and optimal number of trees. J. King Saud-Univ.-Comput. Inf. Sci. 2022, 34, 2733–2742. [Google Scholar] [CrossRef]
  17. Rupapara, V.; Rustam, F.; Aljedaani, W.; Shahzad, H.F.; Lee, E.; Ashraf, I. Blood cancer prediction using leukemia microarray gene data and hybrid logistic vector trees model. Sci. Rep. 2022, 12, 1000. [Google Scholar] [CrossRef]
  18. Umer, M.; Ashraf, I.; Ullah, S.; Mehmood, A.; Choi, G.S. COVINet: A convolutional neural network approach for predicting COVID-19 from chest X-ray images. J. Ambient. Intell. Humaniz. Comput. 2022, 13, 535–547. [Google Scholar] [CrossRef]
  19. Shafi, I.; Din, S.; Khan, A.; Díez, I.D.L.T.; Casanova, R.d.J.P.; Pifarre, K.T.; Ashraf, I. An effective method for lung cancer diagnosis from ct scan using deep learning-based support vector network. Cancers 2022, 14, 5457. [Google Scholar] [CrossRef] [PubMed]
  20. Natalia, F.; Meidia, H.; Afriliana, N.; Young, J.C.; Yunus, R.E.; Al-Jumaily, M.; Al-Kafri, A.; Sudirman, S. Automated measurement of anteroposterior diameter and foraminal widths in MRI images for lumbar spinal stenosis diagnosis. PLoS ONE 2020, 15, e0241309. [Google Scholar] [CrossRef] [PubMed]
  21. Sartoretti, E.; Wyss, M.; Alfieri, A.; Binkert, C.A.; Erne, C.; Sartoretti-Schefer, S.; Sartoretti, T. Introduction and reproducibility of an updated practical grading system for lumbar foraminal stenosis based on high-resolution MR imaging. Sci. Rep. 2021, 11, 12000. [Google Scholar] [CrossRef]
  22. Salehi, E.; Khanbare, S.; Yousefi, H.; Sharpasand, H.; Sojoodi Sheyjani, O. Deep Convolutional Neural Networks for Automated Diagnosis of Disc Herniation on Axial MRI. In Proceedings of the 2019 Scientific Meeting on Electrical-Electronics Biomedical Engineering and Computer Science (EBBT), Istanbul, Turkey, 24–26 April 2019; pp. 1–6. [Google Scholar] [CrossRef]
  23. Sudirman, S.; Kafri, A.A.; Natalia, F.; Meidia, H.; Afriliana, N.; Al-Rashdan, W.; Bashtawi, M.; Al-Jumaily, M. Lumbar Spine MRI Dataset. Mendeley Data 2019, 2. [Google Scholar]
  24. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  25. Xu, Q.; Zhang, M.; Gu, Z.; Pan, G. Overfitting remedy by sparsifying regularization on fully-connected layers of CNNs. Neurocomputing 2019, 328, 69–74. [Google Scholar] [CrossRef]
  26. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  27. Shafique, R.; Rustam, F.; Choi, G.S.; Díez, I.d.l.T.; Mahmood, A.; Lipari, V.; Velasco, C.L.R.; Ashraf, I. Breast cancer prediction using fine needle aspiration features and upsampling with supervised machine learning. Cancers 2023, 15, 681. [Google Scholar] [CrossRef]
  28. Hirahara, D.; Takaya, E.; Takahara, T.; Ueda, T. Effects of data count and image scaling on deep learning training. PeerJ Comput. Sci. 2020, 6, e312. [Google Scholar] [CrossRef]
  29. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates, Inc.: New York, NY, USA, 2012; Volume 25. [Google Scholar]
  30. Zhou, D.X. Theory of deep convolutional neural networks: Downsampling. Neural Netw. 2020, 124, 319–327. [Google Scholar] [CrossRef] [PubMed]
  31. Li, G.; Zhang, M.; Li, J.; Lv, F.; Tong, G. Efficient densely connected convolutional neural networks. Pattern Recognit. 2021, 109, 107610. [Google Scholar] [CrossRef]
  32. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
  33. Karimi, H.; Derr, T.; Tang, J. Characterizing the decision boundary of deep neural networks. arXiv 2019, arXiv:1912.11460. [Google Scholar]
  34. Al-Kafri, A. A Machine Learning and Computer Assisted Methodology for Diagnosing Chronic Lower Back Pain on Lumbar Spine Magnetic Resonance Images. Ph.D. Thesis, Liverpool John Moores University, Liverpool, UK, 2019. [Google Scholar]
  35. Hallinan, J.T.P.D.; Zhu, L.; Yang, K.; Makmur, A.; Algazwi, D.A.R.; Thian, Y.L.; Lau, S.; Choo, Y.S.; Eide, S.E.; Yap, Q.V.; et al. Deep learning model for automated detection and classification of central canal, lateral recess, and neural foraminal stenosis at lumbar spine MRI. Radiology 2021, 300, 204289. [Google Scholar] [CrossRef] [PubMed]
  36. Su, Z.H.; Liu, J.; Yang, M.S.; Chen, Z.Y.; You, K.; Shen, J.; Huang, C.J.; Zhao, Q.H.; Liu, E.Q.; Zhao, L.; et al. Automatic grading of disc herniation, central canal stenosis and nerve roots compression in lumbar magnetic resonance image diagnosis. Front. Endocrinol. 2022, 13, 890371. [Google Scholar] [CrossRef]
  37. Altun, S.; Alkan, A.; Altun, İ. LSS-VGG16: Diagnosis of Lumbar Spinal Stenosis With Deep Learning. Clin. Spine Surg. 2023, 36, E180–E190. [Google Scholar] [CrossRef]
  38. Lu, J.T.; Pedemonte, S.; Bizzo, B.; Doyle, S.; Andriole, K.P.; Michalski, M.H.; Gonzalez, R.G.; Pomerantz, S.R. Deep Spine: Automated Lumbar Vertebral Segmentation, Disc-Level Designation, and Spinal Stenosis Grading using Deep Learning. In Proceedings of the 3rd Machine Learning for Healthcare Conference, Palo Alto, CA, USA, 17–18 August 2018; Doshi-Velez, F., Fackler, J., Jung, K., Kale, D., Ranganath, R., Wallace, B., Wiens, J., Eds.; PMLR: Cambridge, MA, USA, 2018; Volume 85, pp. 403–419. [Google Scholar]
  39. Fujiwara, M.; Kashiwagi, N.; Matsuo, C.; Watanabe, H.; Kassai, Y.; Nakamoto, A.; Tomiyama, N. Ultrafast lumbar spine MRI protocol using deep learning–based reconstruction: Diagnostic equivalence to a conventional protocol. Skelet. Radiol. 2023, 52, 233–241. [Google Scholar] [CrossRef]
  40. Han, Z.; Wei, B.; Leung, S.; Nachum, I.B.; Laidley, D.; Li, S. Automated pathogenesis-based diagnosis of lumbar neural foraminal stenosis via deep multiscale multitask learning. Neuroinformatics 2018, 16, 325–337. [Google Scholar] [CrossRef]
  41. Ruiz-España, S.; Arana, E.; Moratal, D. Semiautomatic computer-aided classification of degenerative lumbar spine disease in magnetic resonance imaging. Comput. Biol. Med. 2015, 62, 196–205. [Google Scholar] [CrossRef]
  42. Alomari, R.S.; Chaudhary, V.; Dhillon, G. Computer Aided Diagnosis System for Lumbar Spine. In Proceedings of the 4th International Symposium on Applied Sciences in Biomedical and Communication Technologies, New York, NY, USA, 26–29 October 2011. [Google Scholar] [CrossRef]
  43. Alomari, R.; Corso, J.J.; Chaudhary, V.; Dhillon, G. Toward a clinical lumbar CAD: Herniation diagnosis. Int. J. Comput. Assist. Radiol. Surg. 2011, 6, 119–126. [Google Scholar] [CrossRef] [PubMed]
  44. Nikravan, M.; Ebrahimzadeh, E.; Izadi, M.R.; Mikaeili, M. Toward a computer aided diagnosis system for lumbar disc herniation disease based on MR images analysis. Biomed. Eng. Appl. Basis Commun. 2016, 28, 1650042. [Google Scholar] [CrossRef]
  45. Ghosh, S.; Chaudhary, V. Supervised methods for detection and segmentation of tissues in clinical lumbar MRI. Comput. Med. Imaging Graph. 2014, 38, 639–649. [Google Scholar] [CrossRef] [PubMed]
  46. Bharadwaj, U.U.; Christine, M.; Li, S.; Chou, D.; Pedoia, V.; Link, T.M.; Chin, C.T.; Majumdar, S. Deep learning for automated, interpretable classification of lumbar spinal stenosis and facet arthropathy from axial MRI. Eur. Radiol. 2023, 33, 3435–3443. [Google Scholar] [CrossRef] [PubMed]
  47. Altun, S.; Alkan, A. LSS-net: 3-dimensional segmentation of the spinal canal for the diagnosis of lumbar spinal stenosis. Int. J. Imaging Syst. Technol. 2023, 33, 378–388. [Google Scholar] [CrossRef]
  48. Shukla, A.; Bhardwaj, S.; Singh, M. Segmentation for Lumbar Spinal Stenosis Using Convolutional Neural Networks. Procedia Comput. Sci. 2023, 218, 2210–2223. [Google Scholar] [CrossRef]
Figure 1. Flow chart of a standard CAD system for diagnosis of lumbar spine stenosis.
Figure 1. Flow chart of a standard CAD system for diagnosis of lumbar spine stenosis.
Diagnostics 13 02975 g001
Figure 2. CNN architecture for image classification.
Figure 2. CNN architecture for image classification.
Diagnostics 13 02975 g002
Figure 3. Four-class manual segmentation mask.
Figure 3. Four-class manual segmentation mask.
Diagnostics 13 02975 g003
Figure 4. Training and validation loss and accuracy with 12.5 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Figure 4. Training and validation loss and accuracy with 12.5 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Diagnostics 13 02975 g004
Figure 5. Training and validation loss and accuracy for grading of LSS with 10 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Figure 5. Training and validation loss and accuracy for grading of LSS with 10 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Diagnostics 13 02975 g005aDiagnostics 13 02975 g005b
Figure 6. Training and validation loss for grading of LSS with 5 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Figure 6. Training and validation loss for grading of LSS with 5 k data samples. (a) Training and validation accuracy for multi-ROI dataset. (b) Training and validation loss for multi-ROI dataset. (c) Training and validation accuracy for single-ROI dataset. (d) Training and validation loss for single-ROI dataset.
Diagnostics 13 02975 g006
Figure 7. Comparative analysis of classification with different dataset values using the (a) Multi-ROI and (b) Single-ROI datasets.
Figure 7. Comparative analysis of classification with different dataset values using the (a) Multi-ROI and (b) Single-ROI datasets.
Diagnostics 13 02975 g007
Figure 8. Comparative analysis of accuracy. (a) Non-augmented multi-ROI vs. single-ROI dataset. (b) Augmented multi-ROI vs. single-ROI dataset.
Figure 8. Comparative analysis of accuracy. (a) Non-augmented multi-ROI vs. single-ROI dataset. (b) Augmented multi-ROI vs. single-ROI dataset.
Diagnostics 13 02975 g008
Table 1. Summary of the proposed CNN model.
Table 1. Summary of the proposed CNN model.
LayerFeature MapFilterStridesPaddingActivation
Convolution323 × 31 × 1SameReLU
Max Pooling322 × 2-SameRelu
Convolution643 × 31 × 1SameReLU
Max Pooling642 × 2-SameRelu
Convolution1283 × 31 × 1SameReLU
Max Pooling1282 × 2-SameRelu
Convolution2563 × 31 × 1SameReLU
Max Pooling2562 × 2-SameRelu
Convolution5123 × 31 × 1SameReLU
Max Pooling5122 × 2-SameReLU
Fully Connected256---ReLU
Fully Connected256---ReLU
Fully Connected256---ReLU
Fully Connected128---ReLU
Fully Connected4---Softmax
Table 2. Comparative results of multi-ROI and single-ROI with 12.5 k data.
Table 2. Comparative results of multi-ROI and single-ROI with 12.5 k data.
ROI TypeAccuracyClassPrecisionRecallF1 Score
Multi-ROI dataset0.97Normal1.000.990.99
Mild0.940.920.93
Moderate0.960.980.97
Severe0.970.970.97
Macro avg.0.970.960.96
Weighted avg.0.970.970.97
Single-ROI dataset0.98Normal0.970.990.98
Mild0.990.990.99
Moderate0.960.960.96
Severe0.990.980.98
Macro avg.0.980.980.98
Weighted avg.0.980.980.98
Table 3. Comparative results of multi-ROI and single-ROI with 10 k data.
Table 3. Comparative results of multi-ROI and single-ROI with 10 k data.
ROI TypeAccuracyClassPrecisionRecallF1 Score
Multi-ROI dataset0.96Normal0.981.000.99
Mild0.990.890.94
Moderate0.920.960.94
Severe0.940.990.97
Macro avg.0.960.960.96
Weighted avg.0.960.960.96
Single-ROI dataset0.98Normal0.970.990.98
Mild0.990.990.99
Moderate0.960.960.96
Severe0.990.980.98
Macro avg.0.980.980.98
Weighted avg.0.980.980.98
Table 4. Comparative results of multi-ROI and single-ROI datasets with 5 k data.
Table 4. Comparative results of multi-ROI and single-ROI datasets with 5 k data.
ROI TypeAccuracyClassPrecisionRecallF1 Score
Multi-ROI dataset0.85Normal0.960.840.90
Mild0.810.810.81
Moderate0.740.850.79
Severe0.860.890.87
Macro avg.0.840.850.84
Weighted avg.0.860.850.85
Single-ROI dataset0.92Normal0.990.860.92
Mild0.910.900.91
Moderate0.860.940.90
Severe0.890.990.94
Macro avg.0.910.920.92
Weighted avg.0.920.920.92
Table 5. Comparative results of multi-ROI and single-ROI datasets with different dataset sizes.
Table 5. Comparative results of multi-ROI and single-ROI datasets with different dataset sizes.
ROI TypeData ValuesPrecisionRecallF1 ScoreAccuracy
Multi-ROI dataset5 k0.860.850.850.85
10 k0.960.960.960.96
12.5 k0.970.970.970.97
Single-ROI dataset5 k0.920.920.920.92
10 k0.980.980.980.98
12.5 k0.980.980.980.98
Table 6. Comparison of overall dataset accuracy with different data values.
Table 6. Comparison of overall dataset accuracy with different data values.
Data TypeData ValuesMulti-ROI DatasetSingle-ROI Dataset
Augmented data500086.393.15
10,00095.8397.71
125,00097.0197.71
Non-Augmented Data154536.4735.66
Table 7. Performance comparison with state-of-the-art approaches.
Table 7. Performance comparison with state-of-the-art approaches.
Ref.ModelAccuracy
[22]CNN87.75%
[38]UNet94%
[40]DMML-Net84.5%
[35]CNN84.5%
[36]ResNet-20LDH—84.17%,LCCS—86.99%,LNRC—81.21%
[37]VGG1687.70%
[39]Different models95%
ProposedCNN97.71% for singl-ROI with 12.5 k data97.01% for multi-ROI with 12.5 k data
Table 8. Analytical overview of the discussed research works.
Table 8. Analytical overview of the discussed research works.
Ref.ModelDataResults
[20]SegNet2D T1- and T2-weighted MRI images with 320 × 320 resolution96.7% agreement with medical expert
[22]CNNHigh-resolution 3D MRI images87.75% accuracy with data augmentation
[35]CNN2D T1- and T2-weighted MRI images84.5% accuracy
[36]ResNet-202D T2-weighted MRI imagesLDH, LCCS, and LNRC accuracy of 84.17%, 86.99%, and 81.21%, respectively, for internal dataset.74.16%, 79.65%, and 81.21% accuracy for LDH, LCCS, and LNRC, respectively, using external dataset
[37]VGG162D MRI images87.70% accuracy
[38]U-NetHigh-resolution 2D MRI images with 512 × 512 resolution94% accuracy
[46]V-Net model2D T2-weighted MRI imagesFacet and neural foraminal stenosis AUC of 0.92 and 0.93, respectively
[39]Different modelsMRI T1-weighted, T2-weighted, short T1-inversion95% accuracy
[40]DMML-NetHigh-resolution 2D T1- and T2-weighted MRI images with 512 × 512 resolution84.5% accuracy
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shahzadi, T.; Ali, M.U.; Majeed, F.; Sana, M.U.; Diaz, R.M.; Samad, M.A.; Ashraf, I. Nerve Root Compression Analysis to Find Lumbar Spine Stenosis on MRI Using CNN. Diagnostics 2023, 13, 2975. https://doi.org/10.3390/diagnostics13182975

AMA Style

Shahzadi T, Ali MU, Majeed F, Sana MU, Diaz RM, Samad MA, Ashraf I. Nerve Root Compression Analysis to Find Lumbar Spine Stenosis on MRI Using CNN. Diagnostics. 2023; 13(18):2975. https://doi.org/10.3390/diagnostics13182975

Chicago/Turabian Style

Shahzadi, Turrnum, Muhammad Usman Ali, Fiaz Majeed, Muhammad Usman Sana, Raquel Martínez Diaz, Md Abdus Samad, and Imran Ashraf. 2023. "Nerve Root Compression Analysis to Find Lumbar Spine Stenosis on MRI Using CNN" Diagnostics 13, no. 18: 2975. https://doi.org/10.3390/diagnostics13182975

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop