Next Article in Journal
Deep-Learning-Based Visualization and Volumetric Analysis of Fluid Regions in Optical Coherence Tomography Scans
Previous Article in Journal
Detection of Diabetes through Microarray Genes with Enhancement of Classifiers Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images

1
Department of Computer Science, University of Okara, Okara 56310, Pakistan
2
Mathematics Research Centre, Academy of Athens, 10679 Athens, Greece
3
Department of Computer Science, Government College Women University, Sialkot 51310, Pakistan
4
Department of Information Systems, College of Computer and Information Science, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(16), 2658; https://doi.org/10.3390/diagnostics13162658
Submission received: 4 March 2023 / Revised: 9 April 2023 / Accepted: 20 April 2023 / Published: 12 August 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Spine image analysis is based on the accurate segmentation and vertebrae recognition of the spine. Several deep learning models have been proposed for spine segmentation and vertebrae recognition, but they are very computationally demanding. In this research, a novel deep learning model is introduced for spine segmentation and vertebrae recognition using CT images. The proposed model works in two steps: (1) A cascaded hierarchical atrous spatial pyramid pooling residual attention U-Net (CHASPPRAU-Net), which is a modified version of U-Net, is used for the segmentation of the spine. Cascaded spatial pyramid pooling layers, along with residual blocks, are used for feature extraction, while the attention module is used for focusing on regions of interest. (2) A 3D mobile residual U-Net (MRU-Net) is used for vertebrae recognition. MobileNetv2 includes residual and attention modules to accurately extract features from the axial, sagittal, and coronal views of 3D spine images. The features from these three views are concatenated to form a 3D feature map. After that, a 3D deep learning model is used for vertebrae recognition. The VerSe 20 and VerSe 19 datasets were used to validate the proposed model. The model achieved more accurate results in spine segmentation and vertebrae recognition than the state-of-the-art methods.

1. Introduction

The treatment and diagnosis of pathological diseases require adequate spine segmentation and vertebrae recognition, which help to plan better spine surgeries [1]. Vertebral segmentation and recognition are key components in the development of computer-aided diagnosis (CAD) applications for the diagnosis and treatment of spine-related diseases. A computer-based system extracts important features from medical images and creates a 3D model of the patient. The surgeon only needs to make minor adjustments to this model to obtain views of the volume from any angle at any depth. As a result, the surgeon may analyze the situation more thoroughly and provide a more precise diagnosis [2]. The accurate identification of vertebral boundaries and the recognition of individual vertebrae from medical images can improve efficiency and accuracy.
Spine segmentation refers to the process of identifying the vertebral boundaries from medical images such as computer tomography (CT), magnetic resonance imaging (MRI), or X-ray imaging. The segmentation of the spine is challenging due to the complex anatomical structure, the variability in the shape and size of the vertebrae, and the presence of surrounding tissues. Different approaches have been proposed in the literature to address these challenges. Some of the commonly used segmentation techniques are atlas-based segmentation [3], region-based segmentation [4], and active contour-based segmentation [5]. A spine segmentation procedure that creates anatomically correct 3D models can be hindered by certain factors such as the anatomic complexity of the spine, image noise, low intensity, and the partial volume effect [6]. Vertebrae recognition refers to the process of identifying and labeling individual vertebrae from segmented spine images. The accurate recognition of vertebrae is important for various clinical applications such as vertebral fracture detection, scoliosis diagnosis, and surgical planning. Different approaches have been proposed in the literature for vertebrae recognition. Some of the commonly used techniques are template matching [7], shape-based recognition [8], and machine-learning-based recognition [9].
The exact vertebral boundaries, however, cannot be easily defined due to the articulation of the vertebrae with one another, which can cause vertebral overlaps in segmentation. Despite growing interest in spine segmentation and vertebrae recognition, reliable and precise spine segmentation approaches are still absent. Furthermore, numerous techniques to appropriately segment spine patients with osteoporosis fractures fail because such patients frequently experience vertebral fractures in various phases and present spinal anomalies. Because the variability of a distinctive shape differs from the mean shape, model-dependent segmentation may not succeed with such fractures.
In this research, a novel deep learning model is introduced for automated and efficient spine segmentation and vertebrae recognition using CT images. This approach works in two phases. First, a cascaded hierarchical atrous spatial pyramid pooling residual attention U-Net (CHASPPRAU-Net) is used for the segmentation of the spine. The model contains residual and cascaded hierarchical atrous spatial pyramid pooling in the encoder part for feature extraction. The CHASPP module focuses on objects in the dataset, while the residual blocks extract deep features from the dataset. Attention modules are used in the decoder part, which helps improve the performance by focusing on regions of interest. The skip connections pass the information from residual blocks to decoder layers. After that, a novel 3D mobile residual U-Net (MRU-Net) deep learning model is used for vertebrae recognition. This model uses a modified version of MobileNetv2, with residual blocks and depthwise convolutions, as a decoder. Three modified MobileNetv2 architectures are used for feature extraction from axial, sagittal, and coronal views of 3D CT images. The features are combined to form a 3D feature map, which is given to the decoder part for final vertebrae recognition. The decoder part of the model is the expansion part of the standard 3D U-Net for final vertebrae recognition. The proposed models achieve better results in spine segmentation and vertebrae recognition compared to other models.
The main contributions of the proposed approach are given below:
  • A novel automated deep learning approach for spine segmentation with residual blocks, an attention module, a U-Net, and cascaded hierarchical atrous spatial pyramid pooling known as CHASPPRAU-Net, which uses CT images.
  • Efficient and robust automated spine vertebrae recognition using MobileNetv2, residual blocks, and a 3D U-Net architecture known as 3D MRU-Net, which uses CT images.
  • The proposed CHASPPRAU-Net model provides efficient and accurate spine segmentation compared to the state-of-the-art methods.
  • The 3D MRU-Net has an optimal number of trainable parameters that have lower computational costs compared to the other 3D deep learning models.

2. Related Work

Medical image analysis has been revolutionized through the innovation of deep learning models. Many machine learning models are available for the diagnosis and treatment of diseases, such as brain tumor segmentation [10], diabetic retinopathy detection [11], glaucoma detection [12], and COVID-19 detection [13]. Kim et al. [14] used deep-learning-based approaches, particularly the U-Net architecture [15], to create an application for segmenting the spine using CT images. The data processing tool was built in Python using the Keras library, while the web accessibility was built using the Flask server framework. The U-Net was trained on 310 CT images, validated on 20 images, and tested on 14 images. The model was able to obtain a dice score of 90.40% for binary segmentation.
An automated deep learning approach for spine segmentation was proposed by Zhang et al. [16] by improving the U-Net model’s architecture. MRI images were used for the segmentation of the spine. The proposed improved U-Net architecture reduced the image processing time of the MRI images compared to the FCN and the standard architecture of the U-Net model. The results of this approach were better than those of previous models; however, the test datasets were very small, and the models were tested on a single source. Christian et al. [17] proposed a four-step approach for automated spine segmentation. First, a basic U-Net model was used to segment the spines in MRI images. After that, a multi-class U-Net model was used to generate a fine segmentation, including labels for vertebrae and vertebral body landmarks. After that, a model-based segmentation was initialized to detect the vertebrae. The proposed model achieved better results using a dataset of 147 images.
A pixel-based model using deep learning was proposed for spine segmentation by Kiran et al. [18]. The proposed model was different from the conventional spine segmentation methods and achieved better results in terms of accuracy and precision. A multitask learning model was proposed by Tran et al. [19] for automated spine segmentation and spinal parameter inspection. This approach consisted of two parts: the first part was used for spine segmentation, while the second part was responsible for spine inspection. The model was evaluated using a dataset that was collected and annotated by radiologists. The results of the model were better for spine segmentation as well as spine inspection. Patch-based deep belief networks were created by Qadri et al. to automatically segment vertebrae in CT images [20]. DBNs are deep learning models made up of layered restricted Boltzmann machines (RBMs) [21]. The proposed methodology helps to quantify the differences between classes by automatically selecting features from image patches. Weight initialization is performed using unsupervised learning, whereas weight updates are conducted via supervised fine-tuning.
Binhui et al. [22] proposed a SePiCo (semantic guided pixel contrast) for semantic segmentation using a one-stage adaption network that used learning of class discriminative and class-balanced pixel representations with the increase in self-training approaches. First, the discriminative features were used for centroid-aware pixel contrast. After that, distribution-aware pixel contrast was also used to determine the correct distribution of each semantic category of a labeled dataset. Ban et al. [23] proposed a feature-based algorithm for the medical image registration of 2D and 3D images. Statistical features were extracted using a weighted spatial histogram of gradient directions. The proposed approach was tested on CT images and X-ray images, which improved the accuracy and efficiency of the model. A cascade residual dense network was developed by Duanet al. [24] for the construction of high-quality diffusion-weighted (DW) images using k-space data. The model achieved better results for lung disease using the DW-MRI dataset.
Diniz et al. [25] proposed an approach that was based on template matching and a CNN model with residual blocks. The model was evaluated using a CT database of 36 patients. The model achieved better accuracy results for spine segmentation. Chang et al. [26] proposed a deep learning approach for spine segmentation where the parameters were updated adaptively based on the input image. In seventeen 3D vertebral CT images of the lumbar and thoracic spine, the normal and pathological instances of both systems were calculated in terms of DSC. The results were compared to four distinct models. The robustness of the APCNN and MLPNN was also tested by introducing noise to the images. With a dice score of 95%, the APCNN outperformed the other approaches.
Having reviewed key works on the use of deep learning for segmentation and classification, it needs to be noted that there are still challenges that need to be improved. The biggest problem is the computational costs of the deep learning models, which require more hardware resources to process the training datasets. The testing images of most of the deep learning models are limited. The testing should contain at least 20% of the total dataset so that the performances of the models can be evaluated correctly.

3. Materials and Methods

Deep learning is becoming more popular in medical imaging for the diagnosis [27] and treatment of many diseases, mainly by performing segmentation and classification tasks [28]. For image-related problems, convolutional neural networks are widely used for segmentation and classification, and they usually involve a fully convolutional neural network [29] and some encoder–decoder-based architectures such as V-Net [30] and 3D U-Net [31]. There are many applications of semantic segmentation in medical imaging, such as glomeruli segmentation [32,33], autosomal dominant polycystic segmentation using magnetic resonance images (MRI) [34], brain tumor segmentation [10], and vessel segmentation [35].

3.1. MobileNetv2

MobileNetv2 [36] is a neural network architecture that contains 53 deep layers. The features are filtered using a lightweight depthwise convolution that extracts fewer trainable parameters. It is specially designed for devices with low computational resources such as mobile phones. This type of model can reduce the computational costs of machine learning models while maintaining accuracy. The complete architecture of MobileNetv2 is shown in Table 1.
It is based on a residual structure in which there are residual connections between the bottleneck layers. The features are filtered using lightweight depthwise convolutions by intermediate expansion layers. The complete architecture of MobileNetv2 contains fully convolutional layers with 32 filters and 19 residual bottleneck layers. MobileNetv2 is used for feature extraction in spine vertebrae recognition, which decreases the number of trainable parameters and subsequently decreases the computational cost.

3.2. Structure of Residual Blocks

Residual blocks [37] are very popular for learning deep features because they make another path to reach other layers of the neural network by skipping some layers. This acts like a stack, where the output of a layer is added to another layer of the network. The complete architecture of the residual blocks is shown in Figure 1.
Residual blocks are used in the encoder part of the proposed model for spine segmentation. The features from residual blocks are passed to the decoder layers using skip connections.

3.3. Attention Module

The attention module [38] is widely used in various deep learning tasks, such as image processing, natural language, and video processing, because it has an in-depth understanding of the input. There are two types of attention modules, namely soft attention and hard attention. In soft attention, all data pay attention and calculate the corresponding weight without setting conditions, whereas hard attention filters the unqualified part of the data after the calculations. The general attention module performs the following operations:
Each query vector (q = St − 1) is matched with database keys to calculate the value of the score. It is considered a dot product of a query by matching with each key vector (ki).
eq,ki = q.ki
The weighted sum of vectors (Vki), where every value is joined with a corresponding key, is computed to obtain the generalized attention:
Attention (q, K, V) = ∑iαq,kiVki
Each input word is attributed its own key, value, and query. Such vectors are generated by multiplying specific words with three different weight matrices that are generated during the training process. Basically, when the generalized attention mechanism is provided with a string of words, it evaluates each key in the database using the query vector assigned to a particular word in the string. This depicts the relationship between the word under examination and the other words in the sequence. The values are then scaled in accordance with the attention weights to maintain the focus on the query-relevant terms. As a result, the term under consideration receives an output of attention. A general mechanism of an attention module is shown in Figure 2.
To focus on the region of interest, the attention module is used in the decoder part of the proposed model for spine segmentation. The attention modules improve the accuracy of the model.

3.4. Atrous Spatial Pyramid Pooling Modules (ASPP)

ASPP [39] is a module used for semantic segmentation in which a feature layer can be resampled with different convolution rates. This helps to obtain useful image information at multiple scales and preserve more important information. Besides resampling features, it is implemented using multiple atrous convolutional layers with different sampling rates. The complete working of the ASPP module is shown in Figure 3.
With the help of different convolution rates, the local information from the feature map can be used to improve the performance of the algorithm.
Limitations: As ASPP provides better results using different convolution rates and extracts features at multiple resolutions, it has some limitations [40]. The sampling scope is limited due to the fact that the ASPP is not applicable in some conditions: (1) The target object is in a large distribution or is very disconnected. For objects in a limited scope, ASPP performs well and extracts key information. If the components of an object are split, the ASPP cannot extract global information from them. (2) The contextual information that can provide auxiliary information, which helps to discriminate local patches, is arbitrarily scattered in the image.

3.5. Cascaded Hierarchical Atrous Spatial Pyramid Pooling Module

To solve the problem of the ASPP, a new cascaded hierarchical ASPP was proposed [40] that increases the number of sampling points inside the receptive field. The structure of the ASPP was modified into a two-level hierarchical structure using a one-root atrous convolution and three-branch convolution layers with small convolution layers. The unique features of the sampling area can be extracted easily and can be used in the training process. The density of the sampling points of CHASPP is shown in Figure 4.
CHASPP is used in the encoder part of the proposed machine learning model for feature extraction. With its help, the local and global features are extracted, which increases the performance of the proposed model for spine segmentation. The limitations of the ASPP are also improved with the help of CHASPP.

3.6. U-Net Model for Semantic Segmentation

The most popular model for the segmentation of biomedical images is U-Net [15]. It is known as U-Net because its architecture is U-shaped. It consists of two parts: an encoder and a decoder. The encoder part is used to extract features from the given dataset, and the decoder part is used to predict the segmented mask. The model uses the concept of a fully convolutional network and extracted localization as well as context features. The standard architecture of U-Net is shown in Figure 5.
The encoder block reduces the size of the image using max-pooling layers with a stride of 2. The convolutional layers are used with an increasing number of filters. In the decoder part, the number of filters starts decreasing with gradual upsampling. Skip connections that preserve the losses from previous layers and connect those layers with the layers of the decoder blocks are also used.
In the proposed approach for spine segmentation, we modified the standard U-Net architecture by adding CHASPP and residual blocks to the encoder part and adding the attention module to the decoder part. The model achieved promising results in spine segmentation and provided better segmentation.

3.7. Proposed Cascaded Hierarchical Atrous Spatial Pyramid Pooling Attention Residual U-Net for Spine Segmentation

The proposed model is a modified version of the standard U-Net architecture for the automated segmentation of the spine and was named a cascaded hierarchical atrous spatial pyramid pooling attention residual U-Net (CHASPPARU). In the standard U-Net architecture, there are convolution layers and max-pooling layers for feature extraction. The proposed model is an encoder–decoder-based architecture in which three CHASPP layers are used after each max-pooling layer in the encoder part. The complete architecture of the proposed CHASPP is shown in Figure 6. As discussed in Section 3.5, this CHASPP enhances the performance of the machine learning model by extracting local and global contextual information using multiple scale rates. Residual blocks are inserted in each encoder part that preserve the information and pass it to the decoder layers with skip connections. The decoder part of the proposed model contains attention modules that focus on the area of interest.

3.8. Proposed 3D Mobile Residual U-Net for Spine Vertebrae Recognition

A 3D MRU-Net is introduced for the recognition of spine vertebrae using CT images. The complete structure of the proposed model is shown in Figure 7. The proposed model has an encoder–decoder architecture that is a combination of MobileNetV2 [36], a residual block [37], and 3D U-net [31]. MobileNetv2 is a lightweight network that can be used for low-resource devices such as mobile phones. It is difficult to process 3D images without more powerful computational resources such as GPUs. Moreover, there are a lot of trainable parameters in 3D images that further increase the computational cost. Due to these implementation challenges, a lightweight deep learning model, MobileNetv2, is used for feature extraction. The standard architecture of MobileNetv2 was modified by adding residual blocks that help in learning deep features. The new version of MobileNetv2, including residual blocks, is used as an encoder. The 3D U-net is used as a decoder for final vertebrae recognition. The input of the network is CT images of the spine, and the output is segmented spine vertebrae. The network is trained using a 2D neural network that reduces the computational cost. This approach is similar to [41], in which each orthogonal slice was trained on an individual CNN model. The three MobileNetv2 architectures are not connected to each other but are used to separately extract features from the three orthogonal slices. The output feature maps from all individual networks are similar in size and are concatenated to make a 3D feature map. This 3D feature map is given to the decoder part of the 3D-Unet for final vertebrae recognition. The evaluation of this model is conducted using evaluation metrics such as the dice score, intersection over union, precision, and recall.

3.9. Dataset Definition

The publicly available datasets VerSe 2020 and VerSe 2019 [42,43,44] were used in this research to evaluate the model. These challenging datasets allow researchers to adopt deep learning methods for the analysis of spines with multiple conditions, labeled vertebrae, and fields of view. The datasets can be downloaded from the OSF repository [42] and are available in the NIfTI format. The VerSe 2020 dataset consists of 300 CT images with labels. Typical anatomy such as transitional vertebrae, the sacralization of L5, and cervical ribs are included in the VerSe 2020 dataset. The VerSe 2019 data include patients with metallic implants or spinal fractures as well as combinations of isotropic and sagittal reformations. They consist of 160 CT images with centroids and segmented masks. The complete details of the VerSe 2020 and VerSe 2019 datasets are given in Table 2.

4. Results

4.1. Evaluation Metrics

The dice coefficient score, IoU, precision, and recall were used to validate the proposed models. The details of these metrics are given below:

4.1.1. Dice Score

The most common and useful evaluation metric for segmentation tasks is the dice coefficient score [45]. It compares the actual mask and the predicted mask using the following formula:
DSC = 2(P1 × P2)/(P1 + P2)
where P 1 is the predicted image and P 2 is the ground truth of the image.

4.1.2. Intersection over Union (IoU)

Intersection over union [45] is commonly used to compare a predicted mask with a known mask for semantic segmentation. The formula for the calculation of IoU is given below:
IoU = TP/(TP + FN + FP)

4.1.3. Precision

Precision [45] quantifies the total number of correct positive outcomes made by the proposed model. The mathematical formula for calculating precision is given below:
Precision = TP/(TP + FP)

4.1.4. Recall

Recall [45] is calculated as the total number of true positive outcomes divided by the sum of the true positive and false negative outcomes. The mathematical formula for recall is given below:
Recall = TP/(TP + FN)

4.2. Pre-Processing

The VerSe datasets were resized to 256 × 256 to reduce the computational cost of the training process. Different image normalization methods were used, as shown in Figure 8, which included zero-mean scaling, rescaling between 0 and 1, rescaling between −1 and 1, and rescaling between −1000 and +800.
The CT images from both VerSe datasets were pre-processed with smoothing, clamping, and reorienting as used by Payer et al. [26]. However, the clamping range used in this research was [−1000, 800] instead of [−1024, 8192] because the Hounsfield units (HU) of high atomic structures such as bones are in this range [43]. It is important to note that the appropriate HU threshold for bones can vary depending on the specific CT scanner and the protocol used as well as the patient’s age, sex, and other individual factors. The range of HUs in this study was selected by a radiologist with expertise in CT interpretation, who helped us to determine the most appropriate HU threshold for the dataset that was used.
The dataset was partitioned into two parts; eighty percent was used for training and twenty percent was used for testing. The proposed model was trained with 150 epochs, using 0.001 as the initial learning rate. The proposed model took 3.5 h for training and 27 s for a single prediction.

4.3. Experimental Results

This section shows the results of CHASPPRAU-Net and 3D MRU-Net on the VerSe 2020 and VerSe 2019 datasets for spine segmentation and vertebrae recognition. Different experiments were performed to test the performances of the proposed models. The complete details of the experiments conducted in this research are described, and the results are also compared with other methods. The results on the VerSe 2020 and VerSe 2019 datasets were obtained using the best model parameters.
The experimental results that were obtained after applying pre-processing methods are given in Table 3 and Table 4. It was concluded that the normalization range between −1000 to 800 provided the best results for both datasets when compared to other methods.

4.3.1. Data Augmentation

Data augmentation is particularly important because of the limited availability of large annotated datasets. By generating new images with different variations and distortions, data augmentation can help to reduce overfitting and improve the generalizability of machine learning models. Additionally, data augmentation can help to account for differences in imaging protocols, equipment, and conditions, which can all influence the appearance of medical images. In this research, scaling, rotation, and flip rotation were used to produce augmented images. The results are given in Table 5 and Table 6.
The segmentation results of the proposed model with normalization methods and data augmentation are shown in Figure 9. From the results, it was concluded that the proposed model provides accurate segmentation results.

4.3.2. Deep Feature Extraction with Residual Blocks

Residual blocks [37] are popular for extracting deep features and preserving contextual information that is lost due to convolution operations. Therefore, residual blocks were added to the encoder part of 3D MRU-Net for deep feature extraction, which improved the performance of spine vertebrae recognition. The blocks were added to MobileNetv2. The results for spine vertebrae recognition improved, as shown in Table 7.

4.3.3. Dropout Regularization to Overcome Overfitting Problem

Dropout regularization is a technique used in machine learning to prevent overfitting, which occurs when a model becomes too complex and starts to fit the noise in the training data instead of the underlying patterns. During the experiments, the original dropout value of 0.3 was used, which shows a substantial increase in the prediction of the model. The results are shown in Table 8.
The results of spine vertebrae recognition after performing all the experiments are shown in Figure 10. Qualitative comparisons of all the experiments that were conducted are shown in Table 9 and Table 10. The best results are highlighted in bold. From the results, it can be concluded that the proposed approach achieves better results for spine segmentation as well as spine vertebrae recognition, with efficient and accurate results, when compared to the state-of-the-art deep learning models.

5. Discussion

The results obtained using the proposed model were compared with state-of-the-art methods found in the literature using the dice score. The proposed model is efficient and robust and has low computational costs because of MobileNetv2 and the residual blocks. A comparison of the proposed model and other deep learning models is shown in Table 11, which was inspired by [32]. The model performed well with 60 test samples and achieved a dice score of 95.19%.
A 3D V-Net model was proposed by Altini et al. [44] for the automatic identification of vertebrae using k-nn, CNN, and k-means clustering. This method consists of two phases: a binary segmentation of the whole spine using a 3D network and using traditional machine learning models to find the centroids of the vertebrae. The dataset used for the testing of the model was VerSe 2020, in which 214 CT images were extracted for the training and testing of the proposed approach. This approach achieved an 89.17 percent dice score for binary segmentation and a 90.09 percent dice score for the multi-class segmentation. Kim et al. [14] proposed a web-based segmentation of the spine using a deep learning approach. The U-Net architecture was trained on a data format file for spine segmentation. There were 314 CT images in total, from which 300 images were used for training and 14 images were used for testing. The dice score achieved by this approach was 90.4 percent, and it can be used as a diagnostic tool for spine segmentation. However, the testing dataset consisted of a very small number of CT images.
Another network with a redundant class label was used for the automated segmentation of the spine [46]. This was a hybrid approach because a network was combined with a fully convolutional network to enhance the quality of model. The dataset used in this research was the spineweb dataset. The training process took 13 h to complete and achieved a dice score of 94 percent. Qadri et al. [20] introduced an automated deep learning approach using a patch-based method for learning deep features. This method selects features and measures classes. This model used only three CT images for testing and achieved a dice score of 86.1 percent. Zareie et al. [48] introduced a 3D pulse-coupled network for the segmentation of vertebrae using CT images. This model achieved a 95% score for a segmentation task validated using 17 CT images. However, more CT scan images should be used to enhance the performance of the model.
In this work, a novel deep learning approach was used for spine segmentation and vertebrae recognition. First, a CHASPPRAU-Net model was used for the segmentation of the spine. The model had an encoder–decoder architecture in which residual blocks and CHASPP modules were added to the encoder part and attention modules were added to the decoder part of the standard U-Net. After spine segmentation, a 3D MRU-Net was used for vertebrae recognition, which also had an encoder–decoder-based architecture. Three individual modified MobileNetv2 models were used on three different views of CT images (axial, coronal, and sagittal). The features from all three networks were concatenated, and a 3D feature map was given as an input to the decoder part of the 3D MRU-Net for vertebrae recognition. The VerSe 2020 and VerSe 2019 datasets were used to validate the proposed model. From the results, we can see that the proposed model achieved better results.

6. Conclusions

Spine segmentation and vertebrae recognition play important roles in the diagnosis process of patients. However, there is no accurate and efficient method available for the segmentation task. In this study, a novel deep learning model was proposed for the segmentation of the spine and spine vertebrae using CT images. The segmentation of the spine was performed using a CHASPPRAU-Net, while a 3D MRU-Net was used for the automated recognition of vertebrae. The performance of the proposed model was validated using different evaluation metrics. The VerSe 2020 and VerSe 2019 datasets were used for the evaluation of the model. From the results, we can see that the proposed CHASPPRAU-Net model achieved 93.72% and 94.58% dice scores for the VerSe 2020 and VerSe 2019 datasets, respectively, while the 3D MRU-Net model achieved a dice score of 95.19% for the VerSe 2019 dataset and a dice score of 93.89% for the VerSe 2020 dataset for vertebrae recognition.

Author Contributions

Methodology, M.U.S., N.D., A.D. and G.A.; Formal analysis, M.U.S., N.D. and F.H.; Writing—original draft, M.U.S.; Writing—review & editing, A.D. and M.H.; Data curation, M.U.S., G.A. and F.H.; Project administration, N.D., A.D. and F.H.; Resources, N.D. and A.D.; Supervision, N.D.; Investigation, G.A. and M.H.; Software, G.A. and M.H.; Validation, G.A. and M.H.; All authors have read and agreed to the published version of the manuscript.

Funding

Princess Nourah bint Abdulrahman University Researchers Supporting Project Number (PNURSP2023R236), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

OSF|VerSe2020, OSF|VerSe 2019.

Acknowledgments

Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R236), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pereañez, M.; Lekadir, K.; Castro-Mateos, I.; Pozo, J.M.; Lazáry, Á.; Frangi, A.F. Accurate segmentation of vertebral bodies and processes using statistical shape decomposition and conditional models. IEEE Trans. Med. Imaging 2015, 34, 1627–1639. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Shoham, M.; Lieberman, I.; Benzel, E.; Togawa, D.; Zehavi, E.; Zilberstein, B.; Roffman, M.; Bruskin, A.; Fridlander, A.; Joskowicz, L.J. Robotic assisted spinal surgery–from concept to clinical practice. Comput. Aided Surg. 2007, 12, 105–115. [Google Scholar] [PubMed]
  3. Forsberg, D. Atlas-based registration for accurate segmentation of thoracic and lumbar vertebrae in CT data. In Recent Advances in Computational Methods and Clinical Applications for Spine Imaging; Springer: Berlin/Heidelberg, Germany, 2015; pp. 49–59. [Google Scholar]
  4. Ahammad, S.; Rajesh, V. Image processing based segmentation techniques for spinal cord in MRI. Indian J. Public Health Res. Dev. 2018, 9, 317. [Google Scholar] [CrossRef]
  5. Zhang, X.; Li, Y.; Liu, Y.; Tang, S.-X.; Liu, X.; Punithakumar, K.; Shi, D. Automatic spinal cord segmentation from axial-view MRI slices using CNN with grayscale regularized active contour propagation. Comput. Biol. Med. 2021, 132, 104345. [Google Scholar] [CrossRef] [PubMed]
  6. Frost, B.A.; Camarero-Espinosa, S.; Foster, E.J. Materials for the Spine: Anatomy, Problems, and Solutions. Materials 2019, 12, 253. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Huang, Q.; Luo, H.; Yang, C.; Li, J.; Deng, Q.; Liu, P.; Fu, M.; Li, L.; Li, X.J. Anatomical prior based vertebra modelling for reappearance of human spines. Neurocomputing 2022, 500, 750–760. [Google Scholar] [CrossRef]
  8. Benjelloun, M.; Mahmoudi, S.; Lecron, F.J. A framework of vertebra segmentation using the active shape model-based approach. Int. J. Biomed. Imaging 2011, 2011, 621905. [Google Scholar] [CrossRef]
  9. Hornung, A.L.; Hornung, C.M.; Mallow, G.M.; Barajas, J.N.; Rush, A., III; Sayari, A.J.; Galbusera, F.; Wilke, H.-J.; Colman, M.; Phillips, F.M. Artificial intelligence in spine care: Current applications and future utility. Eur. Spine J. 2022, 31, 2057–2081. [Google Scholar] [CrossRef]
  10. Saeed, M.U.; Ali, G.; Bin, W.; Almotiri, S.H.; AlGhamdi, M.A.; Nagra, A.A.; Masood, K.; Amin, R.U. RMU-net: A novel residual mobile U-net model for brain tumor segmentation from MR images. Electronics 2021, 10, 1962. [Google Scholar] [CrossRef]
  11. Nguyen, Q.H.; Muthuraman, R.; Singh, L.; Sen, G.; Tran, A.C.; Nguyen, B.P.; Chua, M. Diabetic retinopathy detection using deep learning. In Proceedings of the 4th International Conference on Machine Learning and Soft Computing, Haiphong City, Vietnam, 17–19 January 2020; pp. 103–107. [Google Scholar]
  12. Christopher, M.; Nakahara, K.; Bowd, C.; Proudfoot, J.A.; Belghith, A.; Goldbaum, M.H.; Rezapour, J.; Weinreb, R.N.; Fazio, M.A.; Girkin, C.A.; et al. Effects of study population, labeling and training on glaucoma detection using deep learning algorithms. Transl. Vis. Sci. Technol. 2020, 9, 27. [Google Scholar] [CrossRef]
  13. Giri, B.; Pandey, S.; Shrestha, R.; Pokharel, K.; Ligler, F.S.; Neupane, B.B. Review of analytical performance of COVID-19 detection methods. Anal. Bioanal. Chem. 2021, 413, 35–48. [Google Scholar] [CrossRef] [PubMed]
  14. Kim, Y.J.; Ganbold, B.; Kim, K.G.J. Web-based spine segmentation using deep learning in computed tomography images. Healthc. Inform. Res. 2020, 26, 61–67. [Google Scholar] [CrossRef]
  15. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  16. Zhang, Q.; Du, Y.; Wei, Z.; Liu, H.; Yang, X.; Zhao, D. Spine medical image segmentation based on deep learning. J. Health Eng. 2021, 2021, 1917946. [Google Scholar] [CrossRef] [PubMed]
  17. Buerger, C.; von Berg, J.; Franz, A.; Klinder, T.; Lorenz, C.; Lenga, M. Combining Deep Learning and Model-Based Segmentation for Labeled Spine CT Segmentation; SPIE: Bellingham, WA, USA, 2020; Volume 11313. [Google Scholar]
  18. Kiran, K.U.; Mamidisetti, G.; shaker Pittala, C.; Vijay, V.; Vallabhuni, R.R. A PCCN-Based Centered Deep Learning Process for Segmentation of Spine and Heart: Image Deep Learning. In Handbook of Research on Technologies and Systems for E-Collaboration During Global Crises; IGI Global: Hershey, PA, USA, 2022; pp. 15–26. [Google Scholar]
  19. Lin, H.-Y.; Liu, H.-W. Multitask deep learning for segmentation and lumbosacral spine inspection. IEEE Trans. Instrum. Meas. 2022, 71, 1–10. [Google Scholar]
  20. Furqan Qadri, S.; Ai, D.; Hu, G.; Ahmad, M.; Huang, Y.; Wang, Y.; Yang, J. Automatic deep feature learning via patch-based deep belief network for vertebrae segmentation in CT images. Appl. Sci. 2018, 9, 69. [Google Scholar] [CrossRef] [Green Version]
  21. Hinton, G.E.; Osindero, S.; Teh, Y.-W.J.N.c. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef]
  22. Xie, B.; Li, S.; Li, M.; Liu, C.H.; Huang, G.; Wang, G. Sepico: Semantic-guided pixel contrast for domain adaptive semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9004–9021. [Google Scholar] [CrossRef]
  23. Ban, Y.; Wang, Y.; Liu, S.; Yang, B.; Liu, M.; Yin, L.; Zheng, W. 2D/3D Multimode Medical Image Alignment Based on Spatial Histograms. Appl. Sci. 2022, 12, 8261. [Google Scholar] [CrossRef]
  24. Duan, C.; Deng, H.; Xiao, S.; Xie, J.; Li, H.; Zhao, X.; Han, D.; Sun, X.; Lou, X.; Ye, C.; et al. Accelerate gas diffusion-weighted MRI for lung morphometry with deep learning. Eur. Radiol. 2022, 32, 702–713. [Google Scholar] [CrossRef]
  25. Diniz, J.O.B.; Ferreira, J.L.; Diniz, P.H.B.; Silva, A.C.; Paiva, A.C. A deep learning method with residual blocks for automatic spinal cord segmentation in planning CT. Biomed. Signal Process. Control. 2022, 71, 103074. [Google Scholar] [CrossRef]
  26. Chang, Q.; Shi, J.; Xiao, Z. A new 3d segmentation algorithm based on 3d pcnn for lung ct slices. In Proceedings of the 2nd International Conference on Biomedical Engineering and Informatics, Tianjin, China, 17–19 October 2009; pp. 1–5. [Google Scholar]
  27. Esteva, A.; Robicquet, A.; Ramsundar, B.; Kuleshov, V.; DePristo, M.; Chou, K.; Cui, C.; Corrado, G.; Thrun, S.; Dean, J. A guide to deep learning in healthcare. Nat. Med. 2019, 25, 24–29. [Google Scholar] [CrossRef]
  28. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2020. [Google Scholar]
  29. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  30. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  31. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, October 17–21 2016; pp. 424–432. [Google Scholar]
  32. Altini, N.; Cascarano, G.D.; Brunetti, A.; Marino, F.; Rocchetti, M.T.; Matino, S.; Venere, U.; Rossini, M.; Pesce, F.; Gesualdo, L. Semantic segmentation framework for glomeruli detection and classification in kidney histological sections. Electronics 2020, 9, 503. [Google Scholar] [CrossRef] [Green Version]
  33. Altini, N.; Cascarano, G.D.; Brunetti, A.; de Feudis, I.; Buongiorno, D.; Rossini, M.; Pesce, F.; Gesualdo, L.; Bevilacqua, V. A deep learning instance segmentation approach for global glomerulosclerosis assessment in donor kidney biopsies. Electronics 2020, 9, 1768. [Google Scholar] [CrossRef]
  34. Bevilacqua, V.; Brunetti, A.; Cascarano, G.D.; Guerriero, A.; Pesce, F.; Moschetta, M.; Gesualdo, L. A comparison between two semantic deep learning frameworks for the autosomal dominant polycystic kidney disease segmentation based on magnetic resonance images. BMC Med. Inform. Decis. Mak. 2019, 19, 244. [Google Scholar] [CrossRef] [Green Version]
  35. Altini, N.; Prencipe, B.; Brunetti, A.; Brunetti, G.; Triggiani, V.; Carnimeo, L.; Marino, F.; Guerriero, A.; Villani, L.; Scardapane, A. A Tversky loss-based convolutional neural network for liver vessels segmentation. In Proceedings of the International Conference on Intelligent Computing, Sanya, China, 4–6 December 2020; pp. 342–354. [Google Scholar]
  36. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  37. Hanif, M.S.; Bilal, M. Competitive residual neural network for image classification. ICT Express 2020, 6, 28–37. [Google Scholar] [CrossRef]
  38. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  39. Yurtkulu, S.C.; Şahin, Y.H.; Unal, G. Semantic segmentation with extended DeepLabv3 architecture. In Proceedings of the 27th Signal Processing and Communications Applications Conference (SIU), Sivas, Turkey, 24–26 April 2019; pp. 1–4. [Google Scholar]
  40. Lian, X.; Pang, Y.; Han, J.; Pan, J. Cascaded hierarchical atrous spatial pyramid pooling module for semantic segmentation. Pattern Recognit. 2021, 110, 107622. [Google Scholar] [CrossRef]
  41. Sekuboyina, A.; Rempfler, M.; Valentinitsch, A.; Menze, B.H.; Kirschke, J.S. Labeling vertebrae with two-dimensional reformations of multidetector CT images: An adversarial approach for incorporating prior knowledge of spine anatomy. Radiol. Artif. Intell. 2020, 2, e190074. [Google Scholar] [CrossRef]
  42. OSF Repository. Available online: https://osf.io/t98fz/ (accessed on 12 December 2022).
  43. Brant, W.E.; Helms, C.A. Fundamentals of Diagnostic Radiology; Wolters Kluwer: Alphen aan den Rijn, The Netherlands, 2012. [Google Scholar]
  44. Altini, N.; de Giosa, G.; Fragasso, N.; Coscia, C.; Sibilano, E.; Prencipe, B.; Hussain, S.M.; Brunetti, A.; Buongiorno, D.; Guerriero, A. Segmentation and identification of vertebrae in CT scans using CNN, k-means clustering and k-NN. Informatics 2021, 8, 40. [Google Scholar] [CrossRef]
  45. Qadri, S.F.; Lin, H.; Shen, L.; Ahmad, M.; Qadri, S.; Khan, S.; Khan, M.; Zareen, S.S.; Akbar, M.A.; Bin Heyat, M.B.; et al. CT-Based Automatic Spine Segmentation Using Patch-Based Deep Learning. Int. J. Intell. Syst. 2023, 2023, 2345835. [Google Scholar] [CrossRef]
  46. Vania, M.; Mureja, D.; Lee, D. Automatic spine segmentation from CT images using convolutional neural network via redundant generation of class labels. J. Comput. Des. Eng. 2019, 6, 224–232. [Google Scholar] [CrossRef]
  47. Li, D.; Dharmawan, D.A.; Ng, B.P.; Rahardja, S. Residual u-net for retinal vessel segmentation. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1425–1429. [Google Scholar]
  48. Zareie, M.; Parsaei, H.; Amiri, S.; Awan, M.S.; Ghofrani, M. Automatic segmentation of vertebrae in 3D CT images using adaptive fast 3D pulse coupled neural networks. Australas. Phys. Eng. Sci. Med. 2018, 41, 1009–1020. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Residual block architecture with skip connections.
Figure 1. Residual block architecture with skip connections.
Diagnostics 13 02658 g001
Figure 2. Attention module architecture used in this research.
Figure 2. Attention module architecture used in this research.
Diagnostics 13 02658 g002
Figure 3. Atrous spatial pyramid pooling module used in this research.
Figure 3. Atrous spatial pyramid pooling module used in this research.
Diagnostics 13 02658 g003
Figure 4. Sampling points of CHASPP. Red donates the receptive field, and blue denotes sampling points.
Figure 4. Sampling points of CHASPP. Red donates the receptive field, and blue denotes sampling points.
Diagnostics 13 02658 g004
Figure 5. The U-Net model’s architecture for medical image segmentation.
Figure 5. The U-Net model’s architecture for medical image segmentation.
Diagnostics 13 02658 g005
Figure 6. The proposed CHASPPARU-Net for automated spine segmentation.
Figure 6. The proposed CHASPPARU-Net for automated spine segmentation.
Diagnostics 13 02658 g006
Figure 7. The proposed 3D MRU-Net for automated vertebrae recognition.
Figure 7. The proposed 3D MRU-Net for automated vertebrae recognition.
Diagnostics 13 02658 g007
Figure 8. Image normalization results with rescaling between −1000 and +800, zero-mean scaling, rescaling between 0 and 1, and rescaling between −1 and 1.
Figure 8. Image normalization results with rescaling between −1000 and +800, zero-mean scaling, rescaling between 0 and 1, and rescaling between −1 and 1.
Diagnostics 13 02658 g008
Figure 9. The image segmentation results of the proposed CHASPPRAU-Net model for spine segmentation.
Figure 9. The image segmentation results of the proposed CHASPPRAU-Net model for spine segmentation.
Diagnostics 13 02658 g009
Figure 10. The image segmentation results of the proposed 3D MRU-Net model for spine segmentation.
Figure 10. The image segmentation results of the proposed 3D MRU-Net model for spine segmentation.
Diagnostics 13 02658 g010
Table 1. The standard architecture of MobileNetv2.
Table 1. The standard architecture of MobileNetv2.
InputOperatortcna
2242 × 3Conv 2D-3222
1122 × 32Bottleneck11611
1122 × 16Bottleneck62422
562 × 24Bottleneck63232
282 × 32Bottleneck66442
142 × 64Bottleneck69631
142 × 96Bottleneck616032
72 × 160Bottleneck632011
Table 2. Descriptions of VerSe 2020 and VerSe 2019 datasets.
Table 2. Descriptions of VerSe 2020 and VerSe 2019 datasets.
DatasetSpine TractTotal ImagesModalityAnnotation
VerSe 2020Whole Spine300CTCentroids and Masks
VerSe 2019Whole Spine160CTCentroids and Masks
Table 3. The results of the CHASPPRAU-Net model for spine segmentation using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods.
Table 3. The results of the CHASPPRAU-Net model for spine segmentation using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80090.4591.4894.3795.07
Zero mean89.9190.3495.1694.32
0 to 190.1894.0193.4194.08
−1 to 188.7592.1791.6291.43
VerSe 2019−1000 to 80091.6390.8995.8494.77
Zero mean88.4289.3793.2592.98
0 to 189.7990.4894.6293.10
−1 to 190.5189.1792.4891.79
Table 4. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods.
Table 4. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80087.9688.3992.7692.46
Zero mean82.4181.6488.4587.45
0 to 183.7584.7889.3788.17
−1 to 182.6383.7986.9785.39
VerSe 2019−1000 to 80086.5887.3590.4991.48
Zero mean84.7284.9888.2187.91
0 to 181.1582.4785.7986.48
−1 to 183.4984.0987.8986.12
Table 5. The results of the CHASPPRAU-Net model for spine segmentation using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods and data augmentation.
Table 5. The results of the CHASPPRAU-Net model for spine segmentation using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods and data augmentation.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80093.7292.2597.0496.87
Zero mean92.6393.1496.1394.53
0 to 190.5990.2893.0892.48
−1 to 193.1494.6094.0593.75
VerSe 2019−1000 to 80094.5895.9398.7197.10
Zero mean92.4991.6895.6295.78
0 to 193.6494.8496.3894.63
−1 to 194.3495.0897.1496.31
Table 6. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods and data augmentation.
Table 6. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods and data augmentation.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80089.5990.7895.2795.88
Zero mean84.6486.4790.3491.15
0 to 185.3784.3992.1390.72
−1 to 184.1985.4789.7888.34
VerSe 2019−1000 to 80090.1891.4896.0595.94
Zero mean86.1787.1990.7589.11
0 to 184.6585.9489.1989.78
−1 to 185.4884.9790.1288.95
Table 7. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods, data augmentation, and residual blocks.
Table 7. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods, data augmentation, and residual blocks.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80093.4693.6698.7198.47
Zero mean88.9588.1492.4993.75
0 to 189.1890.7995.6495.46
−1 to 188.2888.3491.9392.62
VerSe 2019−1000 to 80094.9695.6198.1597.91
Zero mean89.7589.7292.8691.85
0 to 187.4888.6392.4592.14
−1 to 188.9687.9891.4392.48
Table 8. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods, data augmentation, residual blocks, and dropout regularization.
Table 8. The results of the 3D MRU-Net model for spine vertebrae recognition using the VerSe 2020 and VerSe 2019 datasets after applying image normalization methods, data augmentation, residual blocks, and dropout regularization.
DatasetNormalizationDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020−1000 to 80093.8994.0199.1298.79
Zero mean89.4690.4892.8994.18
0 to 190.0291.3796.1595.97
−1 to 189.7188.5292.4593.14
VerSe 2019−1000 to 80095.1995.8199.4898.36
Zero mean89.9290.7293.6292.34
0 to 188.1687.2593.4893.03
−1 to 189.5288.1692.1993.14
Table 9. A comparison of the proposed spine segmentation model with all the experiments conducted in this research on the VerSe 2020 and VerSe 2019 datasets.
Table 9. A comparison of the proposed spine segmentation model with all the experiments conducted in this research on the VerSe 2020 and VerSe 2019 datasets.
DatasetExperimentsDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020Image Normalization90.4591.4894.3795.07
Data Augmentation93.7292.2597.0496.87
VerSe 2019Image Normalization91.6390.8995.8494.77
Data Augmentation94.5895.9398.7197.10
Table 10. A comparison of the proposed spine vertebrae recognition model with all the experiments conducted in this research on the VerSe 2020 and VerSe 2019 datasets.
Table 10. A comparison of the proposed spine vertebrae recognition model with all the experiments conducted in this research on the VerSe 2020 and VerSe 2019 datasets.
DatasetExperimentsDSC (%)IoU (%)Precision (%)Recall (%)
VerSe 2020Image Normalization87.9688.3992.7692.46
Data Augmentation89.5990.7895.2795.88
Residual Blocks93.4693.6698.7198.47
Dropout Regularization93.8994.0199.1298.79
VerSe 2019Image Normalization86.5887.3590.4991.48
Data Augmentation90.1891.4896.0595.94
Residual Blocks94.9695.6198.1597.91
Dropout Regularization95.1995.8199.4898.36
Table 11. Comparison of the proposed model and other deep learning models.
Table 11. Comparison of the proposed model and other deep learning models.
ReferenceMethodologyTest SamplesTrainable ParametersDSC (%)
[44]3D V-Net50-89.17
[14]U-Net141,941,10590.40
[46]CNN32-94.28
[20]PaDBN3-86.1
[47]3D U-Net2519,069,95584.6
[48]PCNN, APCNN, MLPNN, MLPNN1F, APCNN, MLPCNN17-65.7, 95, 91.1, 77.3, 94.3, 87.8
Proposed ModelCHASPPRAU-Net, 3D MRU-Net601,245,15594.58, 95.19
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Saeed, M.U.; Dikaios, N.; Dastgir, A.; Ali, G.; Hamid, M.; Hajjej, F. An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images. Diagnostics 2023, 13, 2658. https://doi.org/10.3390/diagnostics13162658

AMA Style

Saeed MU, Dikaios N, Dastgir A, Ali G, Hamid M, Hajjej F. An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images. Diagnostics. 2023; 13(16):2658. https://doi.org/10.3390/diagnostics13162658

Chicago/Turabian Style

Saeed, Muhammad Usman, Nikolaos Dikaios, Aqsa Dastgir, Ghulam Ali, Muhammad Hamid, and Fahima Hajjej. 2023. "An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images" Diagnostics 13, no. 16: 2658. https://doi.org/10.3390/diagnostics13162658

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop