Next Article in Journal
A System-Performance-Based Comparison of Sparse Regular and Irregular Antenna Arrays for Millimeter-Wave Multi-User MIMO Base Stations
Next Article in Special Issue
A Light-Weight CNN for Object Detection with Sparse Model and Knowledge Distillation
Previous Article in Journal
Model Predictive Control Scheme of a Four-Level Quasi-Nested Converter Fed AC-Drive, with dc-Link Voltage-Drift Compensation
Previous Article in Special Issue
A Two-Stage GAN for High-Resolution Retinal Image Generation and Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning and Kurtosis-Controlled, Entropy-Based Framework for Human Gait Recognition Using Video Sequences

by
Muhammad Imran Sharif
1,
Muhammad Attique Khan
1,
Abdullah Alqahtani
2,
Muhammad Nazir
1,
Shtwai Alsubai
2,
Adel Binbusayyis
2 and
Robertas Damaševičius
3,*
1
Department of Computer Science, HITEC University Taxila, Taxila 47080, Pakistan
2
College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
3
Faculty of Applied Mathematics, Silesian University of Technology, 44-100 Gliwice, Poland
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(3), 334; https://doi.org/10.3390/electronics11030334
Submission received: 28 December 2021 / Revised: 16 January 2022 / Accepted: 18 January 2022 / Published: 21 January 2022
(This article belongs to the Collection Image and Video Analysis and Understanding)

Abstract

:
Gait is commonly defined as the movement pattern of the limbs over a hard substrate, and it serves as a source of identification information for various computer-vision and image-understanding techniques. A variety of parameters, such as human clothing, angle shift, walking style, occlusion, and so on, have a significant impact on gait-recognition systems, making the scene quite complex to handle. In this article, we propose a system that effectively handles problems associated with viewing angle shifts and walking styles in a real-time environment. The following steps are included in the proposed novel framework: (a) real-time video capture, (b) feature extraction using transfer learning on the ResNet101 deep model, and (c) feature selection using the proposed kurtosis-controlled entropy (KcE) approach, followed by a correlation-based feature fusion step. The most discriminant features are then classified using the most advanced machine learning classifiers. The simulation process is fed by the CASIA B dataset as well as a real-time captured dataset. On selected datasets, the accuracy is 95.26% and 96.60%, respectively. When compared to several known techniques, the results show that our proposed framework outperforms them all.

Graphical Abstract

1. Introduction

Human gait recognition (HGR) [1] is a biometric application used to solve human recognition’s behavioral characteristics from a distance. A few other biometrics, such as handwriting, face [2], iris [3], ear [4], electroencephalography (EEG) [5], etc., are also useful for identifying an individual in a defined vicinity [6]. Gait recognition is critical in security systems. In this modern technological era, we require an innovative and up-to-date biometric application; thus, gait is an ideal approach for identifying individuals. The primary advantage of gait recognition over other biometric techniques is that it produces desirable results while avoiding identification from low-resolution videos [7]. Each individual has a few unique characteristics, such as walking style, speed, clothes variation, carrying conditions, and variation in angles [8].
Individual body gestures or walking styles are used to detect human gait features because each subject has a unique walking style. Each subject’s walking style varies depending on the situation and the type of clothing he is wearing [9]. Additionally, when an individual holds a bag, the features are changed [10]. Gait recognition is divided into two approaches: model-based and model-free, also known as the holistic model [11]. The model-based method requires extravagant computing costs. However, in the free-model-based technique, we can detect suspicious activity through preprocessing and segmentation techniques [12]. The major part of the process is to detect the same person in a different environment because a person’s body language varies in different situations (for example, carrying a bag and moving fast) [13].
Many gait-recognition techniques are presented in the literature using machine learning (ML) [14,15]. ML is an important research area; it is utilized in several applications, such as human action recognition [16,17,18], image processing [19,20], and recently, COVID-19 diagnostics [21]. A simple gait-recognition method involves a few essential steps, such as preprocessing original video frames, segmenting the region of interest (ROI), extracting features from ROI, and classifying extracted features using classification algorithms [22]. Researchers use thresholding techniques to segment the ROI after enhancing the contrast of video sequences in the preprocessing step. This step is critical in traditional approaches because the features are only extracted from these segmented regions. This procedure, however, is complicated and unauthentic.
Deep learning has had a lot of success with human gait recognition in recent years. The convolutional neural network (CNN) is a type of deep learning model that is used for several processes, such as gait recognition [23], action recognition [24], medical imaging [25], and others [26,27]. A simple CNN model consists of a few important layers, such as convolutional, pooling, batch normalization, ReLu, GAP, fully connected, and classification layers [28,29]. Figure 1 depicts a simple architecture. A group of filters is adapted in the convolutional layer to extract some important features of an image, such as edges and shape. The non-linear conversion is carried out by the ReLu layer, also known as the activation layer. The batch-normalization layer minimizes the number of training epochs, whereas the pooling layers solve the overflow problem. The fully connected layer extracts the image’s deep features, also known as high-level features, and classifies them in the final step using the softmax function [30,31].
Our major contributions in this work are:
  • A database captured in the real-time outdoor environment using more than 50 subjects. The captured videos include a high rate of noise and background complexity.
  • Refinement of the contrast of extracted video frames using the 3D box filtering approach and then fine-tuning of the ResNet101 model. The transfer-learning-based model is trained on real-time captured video frames and extracted features.
  • A kurtosis-based heuristic approach is proposed to select the best features and fuse them in one vector using the correlation approach.
  • Classification using multiclass one against all-SVM (OaA-SVM) and comparison of the performance of the proposed method on different feature sets.
This paper is organized as follows: previous techniques are discussed in Section 2. Section 3 describes the proposed method, such as frame refinement, deep learning, feature selection, and classification. Results of the proposed technique are discussed in Section 4. Finally, the conclusion and future directions are presented in Section 5.

2. Related Work

There are several methods for HGR using deep learning [32]. The authors focused on deep-learning-based methods in these methods, but the majority of them focused on selecting important features and feature fusion. The experimental process used several gait datasets, such as CASIA A, CASIA B, and CASIA C [33]. In this work, our focus is on the real-time recorded dataset and CASIA B datasets. CASIA B is a famous dataset and is mostly utilized for gait recognition.
Wang et al. [34] presented a novel gait-recognition method using a convolutional LSTM approach named Conv-LSTM. They performed a few steps to complete the gait-recognition process. At the start, they presented GEI frame by frame and then expanded its volume to relax the constraint of the gait cycle. Later on, they performed an experiment to analyze the cross-covariance of one subject. After that, they design a Conv-LSTM model for final gait recognition. The experiments were performed on the CASIA B dataset and OU-ISIR datasets. On the CASIA B dataset, they achieved an accuracy of 93%, and 95% on OU-ISIR, respectively.
Arshad et al. [6] presented a new approach for HGR. In this approach, two deep neural networks were used with the FEcS selection method. Two pre-trained models, VGG19 and AlexNet, were used for feature extraction. The extracted features were refined in the later step using entropy and skewness vectors. Four datasets were used for the experimental process, CASIA A, CASIA B, CASIA C, and AVAMVG. On these datasets, they achieved the accuracy of 99.8%, 99.7%, 93.3%, and 92.2%, respectively.
Mehmood et al. [22] presented an automated deep-learning-based system for HGR under multiple angles. Four key steps were performed, preprocessing, feature extraction, feature selection, and classification. They extracted deep-learning features and applied the firefly algorithm for feature optimization. The experiments were performed on a widely available CASIA B dataset and achieved notable accuracy.
Anusha et al. [35] presented a technique for HGR based on multiple features. They extracted low-level features through spatial, texture, and gradient information. Five databases were used for the experimental process, CASIA A, CASIA B, OU-ISIR D, CMU MoBo, and KTH video datasets. These all datasets were tested on different angles and achieved improved performance.
Sharif et al. [14] presented a method for HGR based on accurate ROI segmentation and multilevel features fusion. Several clothing and carrying conditions were considered for the experimental process and achieved an accuracy of 98.6%, 93.5%, and 97.3%, respectively. A PoseGait model was presented by Liao et al. [13] for HGR. They considered the problem of drastic variations of individuals. The 3D models were used for capturing data from different angles. The 3D image was defined as 3D coordinates of the human body joints.
Wu et al. [36] presented a graph-based approach for multiview HGR. The Spiderweb-graph-based approach was applied to capture the data in a single view and then connect to the other view of gait data concurrently. Memory and capsule modules were used for the trajectory view of each gait as well as STF extraction. The experiments were performed on three challenging gait datasets, SDUgait, OU-MVLP, and CASIA B, and achieved an accuracy of 98.54%, 96.91%, and 98.77%, respectively. Arshad et al. [37] focused on the feature-selection approach to improve the performance of HGR. They used the Bayesian model and quartile-deviation approaches for enhanced feature vectors.
In conclusion, the previous studies focused primarily on the selection of the most relevant features. They used CNN to extract features and then combined the results with data from a few other channels. They did not, however, focus on real-time gait recognition due to the complexity of the system design. In this paper, we focus on real-time gait recognition in an outdoor setting.

3. Proposed Methodology

The proposed method for real-time human gait recognition is presented in this section with detailed mathematical modeling and visual results. The following steps are involved in the proposed framework: video preprocessing, deep learning features extraction using transfer learning, kurtosis-based features selection, correlation-based features fusion, and finally, one-against-all-SVM (OaA-SVM)-based classification. The main architecture diagram is shown in Figure 2. The proposed method is executed in a sequence, and at the end of the execution, it returns a labeled output and numerical results in the form of recall rate, precision, accuracy, etc. The details of each step are given below in the following subsections.

3.1. Videos Preprocessing

One of the most prominent parts of digital image processing is to enhance the quality of visual information of an image. This technique helps to remove the noise and messy information from the image and make it more readable [38]. In the area of machine learning, the algorithm requires better information about an object to learn a good model. In this work, we initially process with videos, and in a later stage, convert them into frames and label them according to the actual class. Then, we resize images into a dimension of R × C × c h , where R denotes row pixels, C denotes column pixels, and c h represent number of channels, respectively. We set R = C = 224 and c h = 3 .
The real-time videos are recorded into four different angles, and each angle consists of two gait classes—walk while carrying a bag and normal walk without any bag. The four angles are 90°, 180°, 270°, and 360°. The original frames also captured some noise during the video recording. For this purpose, we used the 3D box filter, which is a perfect choice. The detail of 3D box filtering is given in Algorithm 1.
Algorithm 1: Data Augmentation Process.
Input: Original video frame ϕ ( r , c ) .
Output: Improved video frame ϕ ˜ ( r , c ) .
Step 1: Load all video frames Δ ( ϕ ) .
for: 1 : t o N f r a m e s
   Step 2: Calculate filter size.
F ˜ a b c h ˜ = r = 1 4 c = 1 4 c h ˜ = 1 3 ( F r c   c h ˜ )
   Step 3: Perform padding.
P a d d i n g = F ˜ a b c h ˜ 1 2
where horizontal padding is performed.
   Step 4: Update padding window.
U p d ( P a d d i n g ) =   ( P a d d i n g ) t
where t = 2
   Step 5: Perform normalization.
N o r m a l i z a t i o n = 1   ( U p d ( P a d d i n g ) )
  product of padding image.
end for
The visual effects are also shown in the main flow, Figure 2. Here, the frames are given before and after the filter processing. The improved frames are utilized in the next step for model learning.

3.2. Convolutional Neural Network

Deep learning emerged recently in the field of computer vision and has since spread to other fields. Convolutional Neural Network (CNN) is a deep learning approach that won a competition for image classification using ImageNet in 2012 [39]. In deep learning, images are directly passed in the model without segmentation; therefore, it is called image-based ML. A simple CNN consists of convolutional layer, ReLu layer, dense layer, pooling layer, and a softmax layer. Visually, simple architecture is shown in Figure 1. In a CNN, the input image is passed to the network in a fixed size. A convolutional operation is performed in the convolutional layer and in the output weights, and bias vectors are generated. Back-propagation is used in the CNN for training the weights. Mathematically, convolutional operation is formulated as follows:
Y ^ C = k = 1 C ψ k k x k ,   C ,
where Y ^ C is convolutional layer output, C denotes the convolutional kernels, x is composed of channels c h , * denotes the convolutional operation, ψ represents convolutional kernel and is defined by ψ k × k × C . To increase the nonlinearity of a network, ReLu layer is applied and defined as follows:
Y ^ r ˜ = m a x ( 0 , x ) ,
Another layer named maxpooling is employed to reduce the dimension of extracted features. This layer is formulated as follows:
Y ^ m a x ( i , j , c ) = max u , v N x + u ,   j + v ,   c ,   i , j , c ,
where Y ^ m a x denotes the max pooling operation, x denotes the input weight matrix, and x Y ^ C , u ,   j ,   and   v are static parameters. The filter of this layer is normally fixed as 3 × 3 , denoted by u , and stride is 2 , denoted by v . An important layer in a CNN is fully connected layer. In this layer, all neurons are connected to each other, and resultant output is obtained in a 1D matrix. Finally, the extracted features are classified in the softmax layer.

3.3. Deep Features Extraction

Feature extraction is the process of reducing original images into small number of important information groups. These groups are called features. The features have few important characteristics such as color, shape, texture, and location of an object. In this proposed technique, we utilized a pre-trained deep learning network named ResNet101. Originally, this model consists of total of 101 layers and was trained on ImageNet dataset. A visual flow is shown in Figure 3. Note that the filter size of first convolutional layer is 7 × 7 , stride is 2, and number of channels is 64 . In the next, a maxpooling layer is applied for filter size 3 × 3 , and stride is 2 . Five residual layers are included in this network, and lastly, global average pool layer and FC layer are added following a softmax layer.
Training Model using Transfer Learning: In this work, we use ResNet101 pre-trained CNN model for features extraction [40]. We trained this model using transfer learning, as visually shown in Figure 4 [41]. Note that parameters of original ResNet101 pre-trained model are transferred and trained on the real-time captured gait database. We remove the last layer from modified ResNet101 and extract features. The features are extracted from global average pool (GAP) and FC layers. Mathematically, TL can define as follows: given a source domain S D and learning task L t , a target domain T D ˜ and learning task L t ˜ , TL improves the learning of a target predictive function F t ( . ) in the T D ˜ using the knowledge in S D and L t , where S D T D ˜ or L t L t ˜ . After this process, we obtain two feature vectors from GAP and FC layers of dimensions N × 2048 and N × 1000 , respectively.

3.4. Kurtosis-Controlled, Entropy-Based Feature Selection

Feature selection (FS) is the process of selecting the most prominent features from the original feature vector. In the FS, the value of features is not updated, and features are selected in the original form [42]. The main motivation behind this step is to obtain the most optimal information of an image and discard the redundant and irrelevant data. In this work, we proposed a new feature selection technique named kurtosis-controlled OaA-SVM. The working of this technique is given in Algorithms 2 and 3 for both vectors.
Description: Consider that we have two extracted deep feature vectors denoted by V 1 and V 2 , where the dimension of each vector is N × 2048 and N × 1000 , respectively. In the first step, features are initialized and processed in the system until all features are passed. In the next step, kurtosis is computed for all features in the pair, where window size and stride were 2 × 2 and 1. A newly generated vector is obtained, which is evaluated in the next step by fitness function. In the fitness function, we select one vs. all SVM classifier that classifies the feature and, in the output, returns an error rate.
This process continues until the error rate is minimized and stops when error increases for the next iteration. By following this, we obtain two feature vectors in the output of dimension N × 600 , as discussed in Algorithms 1 and 2.
Finally, we fused both selected vectors by employing correlation-based approach. In this approach, features of both vectors are paired as i ( 1 ) and j ( 1 ) . After that, the correlation coefficient is computed, and higher-correlated feature pairs are added to the fused vector. Mathematically, it is defined as follows:
r v 1 v 2 = ( v 1 i v 1 ¯ ) ( v 2 j v 2 ¯ ) ( ( v 1 i v 1 ¯ ) 2 ( v 2 j v 2 ¯ ) 2 ) ,
where v 1 V 1 , v 2 V 2 , r v 1 v 2 is a correlation coefficient among two features v 1 i and v 2 j , v 1 i is i t h features of V 1 , and v 2 j is j t h feature of V 2 , respectively. The notation v 1 ¯ is mean value of feature vector V 1 and v 2 ¯ is mean value of feature vector V 2 .
In this approach, we choose those features for fused vector which have correlation value near to 1 or greater than 0. This means we only selected positively correlated features for final fused vector. This procedure is performed for all features in both vectors and in the output; we obtained a resultant features vector of length N × f k , where f k denotes the length of fused feature vector.
Finally, the features in the fused vector are classified using OaA-SVM.
Algorithm 2: Features selection for deep learning model 1.
Input: Feature vector V 1 of dimension N × 2048 .
Output: Selected feature V s of dimension N × 600 .
Step 1: Features initialization.
for  i 1   t o   N          // N = 2048
   Step 2: Compute kurtosis of each feature pair.
-
K u r t o s i s = i = 1 N ( V i V ¯ ) 4 N [ i = 1 N ( V i V ¯ ) 2 N ] 2   // V i is input features, V ¯ is mean feature value, and N denotes the total features.
-
Window size is 2 × 2       // σ is a standard deviation.
-
Stride is 1.
   Step 3: Newly generated kurtosis vector V ˜  // V ˜ is kurtosis vector.
   Step 4: Perform fitness function.
-
OaA-SVM classifier.
-
Evaluate features.
-
Compute error rate.
   Step 5: Repeat steps 2, 3, and 4 until error rate is minimized.
end for
Algorithm 3: Features selection for deep learning model 2.
Input: Feature vector V 2 of dimension   N × 1000 .
Output: Selected feature V s of dimension   N × 600 .
Step 1: Features initialization.
for  i 1   t o   N          // N = 1000
   Step 2: Compute kurtosis of each feature pair.
-
K u r t o s i s = i = 1 N ( V i V ¯ ) 4 N [ i = 1 N ( V i V ¯ ) 2 N ] 2   // V i is input features, V ¯ is mean feature value, and N denotes the total features.
-
Window size is 2 × 2 .
-
Stride is 1.
   Step 3: Newly generated kurtosis vector V ˜   // V ˜ = random   vector   size .
   Step 4: Perform fitness function.
-
OaA-SVM classifier.
-
Evaluate features.
-
Compute error rate.
Step 5: Repeat steps 2, 3, and 4 until error rate is minimized.
end for

3.5. Recognition

The one against all SVM (OaA-SVM) is utilized in this work for features classification [43]. The OaA-SVM approach is usually to determine the separation of one gait action with other listed gait action classes. Mostly, this classification is used for unknown patterns to generate the maximum results.
Here, we have N-class problem, and D represents the training samples: { f k 1 , l 1 } , , { f k D , l D } . Here, f k i   to R b represents b-dimensional feature vector, where l i to { 1 , 2 , , B } is the corresponding class label. This approach usually constructs the binary SVM classifiers. The i t h represent all the positive labels of training class and the other negative labels. Then, i t h   SVM solves the problems with the following yield-decision function:
f i ( s ) = h i T   ϕ ( f k i ) + β i ,
Here, f i ( s ) is the yield-decision function, h i T represents weight values, f k 1 denotes input features, and β i denotes the bias values. The minimization function is applied on weight values as:
Minimize :   M ( h , θ j i ) = 1 2 h i 2 + C l = 1 N θ j i ,
Subject   to :     l j ( h i T ϕ )   ( f k i ) + β i     1 θ j i , θ j i O ,
where M is a minimization parameter, C denotes the output class, θ j i is between-class distance, l j = 1 if   l j = i , and l j = 1   otherwise. A sample l is classified as in class i * .
i * = i = 1 , , B arg   m a x f k i = i = 1 , , B arg   m a x ( h T ϕ ) ( f k i ) + β i ) ,
Resultant visual results are also shown in Figure 5, in which the classifier returned a labeled output such as walking with bag and normal walk. The numerical results are also computed and described in Section 4.

4. Results

4.1. Datasets

The results of the proposed automated system for gait recognition are presented in this section. The proposed framework is tested on a real-time captured dataset and CASIA B dataset. In the real-time dataset, a total of 50 students are included, and for each student, we recorded 8 videos in 4 angles, 90°, 180°, 270°, and 360°. For each angle, two videos are recorded—one while wearing a bag and one without wearing a bag. A few samples are shown in Figure 6. For the CASIA B dataset [44], we only consider a 90° angle for evaluation results. In this dataset, three different covariant factors are included, normal walking, walking with a bag, and walking while wearing a coat, as shown in Figure 7.

4.2. Experimental Setup

We utilized 70% video frames for training the proposed framework, and the remaining 30% were utilized for the testing. In the training process, a group of hyper parameters was employed, where the learning rate was 0.001, epochs was 200, mini batch size was 64, the optimization algorithm was Adam, the activation function was sigmoid, the dropout factor was 0.5, momentum was 0.7, the loss function was cross entropy, and the learning rate schedule was Piecewise. Several classifiers were utilized for validation, such as SVM, K-Nearest Neighbor, decision trees, and Naïve Bayes. The following parameters were used for the analysis of selected classifiers: precision, recall, F1 Score, AUC, and classification time. All results were computed using K-Fold validation, whereas the value of K was 10. The simulations of this work were conducted in MATLAB2020a using a dedicated personal desktop computer with 256 GB SSD, 16 GB RAM, and 16 GB memory graphics card.

4.3. Real-Time Dataset Results

In this section, we present our proposed method’s results on real-time video sequences. The results are computed in three different scenarios—normal walking results for all selected angles, walking while carrying a bag for all three angles, and finally, classification of normal walking and walking while carrying a bag. Table 1 shows the results of a normal walk for the four selected angles of 90°, 180°, 270°, and 360°, respectively. The OaA-SVM classifier achieved superior performance compared to other listed classification algorithms. The achieved accuracy by OaA-SVM is 95.75%, whereas the precision rate is 95.25%, F1-Score is 95.98%, AUC is 1.00, and FPR is 0.0125, respectively. The Medium KNN also gives better performance and achieves an accuracy of 95%. The recall and precision rates are 95% and 95.50%, respectively. The worst accuracy noted for this experiment is 48.75% on the Kernel Naïve Bayes classifier. The change in accuracy among all classifiers shows the authenticity of this work. The performance of OaA-SVM can be further verified in Figure 8. The prediction accuracy for the 270° angle is maximum, whereas the correct prediction performance for 360° is 86%. In the latter, the recognition time is also tabulated in Table 1 and shows that the Fine Tree classifier executed much faster compared to the other listed classifiers. The execution time of Fine Tree is 19.465 (s). The variation in execution time is also shown in Figure 9.
Table 2 shows results of walking while carrying a bag at 90°, 180°, 270°, and 360° angles, respectively. The OaA-SVM classifier attained better accuracy compared to other classification algorithms. For OaA-SVM, the accuracy is 96.5%, whereas the precision rate is 97%, F1-Score is 96.7%, AUC is 1.00, and FPR is 0.01, respectively. The KNN series classifiers, such as Cubic KNN and Medium KNN, give the second-best accuracy of 95.1%. Naïve Bayes classifier achieves an accuracy of 83.7%. The recall and precision rates of the Naïve Bayes classifier are 83.7% and 86.7%, respectively. The Kernel Naïve Bayes, FG-SVM, and Coarse KNN classifiers do not perform well. They achieve accuracies of 48.6%, 61.3%, and 63%, respectively.
The performance of OaA-SVM can be further verified through Figure 10, which shows that the prediction accuracy for the 270° angle is the maximum (100%), whereas correct prediction performance for 360° is 89%. The execution time during the recognition process is also tabulated in Table 2 and shows that the Medium Tree classifier executed much faster compared to other listed classifiers. The execution time is also plotted in Figure 11, which shows a huge variation due to the classifier’s complexity.
Table 3 represents the classification results of a normal walk and walking while carrying a bag. The purpose of this experiment is to analyze the performance of the proposed algorithm for the binary class classification problem. Table 3 shows that the OaA-SVM attained an accuracy of 96.4%. The other calculated measures, such as recall rate, precision rate, f1-Score, AUC, and FPR values, are 96.5%, 96.5%, 96.5%, 1.00, and 0.03, respectively. Medium KNN gives the second-best classification performance of 96.1%, whereas the recall and precision rates are 96% and 96%, respectively. The following classifiers, Naïve Bayes, Coarse KNN, Kernel Naïve Bayes, and FG SVM, do not perform well and achieve an accuracy of 67.8%, 79.3%, 735, and 79.6%, respectively. The confusion matrix also provided and illustrated in Figure 12 shows that the correct prediction rate of both classes is above 90%. In the latter, the execution time is also plotted for this experiment and illustrated in Figure 13, which shows that the Fine Tree execution time is better compared to other algorithms.

4.4. CASIA B Dataset Results at a 90° Angle

CASIA B dataset results are presented in this subsection. For experimental results, we only selected a 90° angle because most of the recent studies used this angle. Numerical results are given in Table 4. The OaA-SVM gives better results among all other implemented classifiers. The results are computed with different feature vectors, such as original global average pool (GAP) layer features, fully connected layer features (FC), and the proposed approach. For GAP-layer features, the achieved accuracy is 90.22%, and the recall rate is 90.10%, whereas the FC layer gives an accuracy of 88.64%. The proposed method gives an accuracy of 95.26%, where the execution time is 176.4450 (s). Using Cubic KNN, the attained accuracy is 93.60%, which is the second-best performance after the OaA-SVM classifier. Table 5 presents the confusion matrix of the OaA-SVM classifier performance using the proposed scheme. In this table, it is illustrated that the corrected prediction accuracy of normal walking is 94%, walking while wearing a coat (W-Coat) is 95%, and walking while carrying a bag is 97%. In addition, the computational time of all classifiers is plotted in Figure 14, which shows that Medium KNN efficiency is better than all other classifiers in terms of the computational cost.

4.5. Discussion

A detailed discussion on results and comparison is conducted in this section. As shown in Figure 1, the proposed method has a few important steps, such as video preprocessing, deep-learning feature extraction, feature selection through the kurtosis approach, the fusion of selected features through the correlation approach, and finally, the OaA-SVM-based feature classification. The proposed method is evaluated on two datasets. One is recorded in the real-time environment, and the second is CASIA B. The real-time captured dataset results are given in Table 1, Table 2 and Table 3. In Table 1, results are presented for normal walking under four different angles, whereas in Table 2, results are given for walking while carrying a bag. The classification of the binary class problem is also conducted, and results are tabulated in Table 3. The OaA-SVM outperformed in all three experiments and can be verified through Table 1. Additionally, we used the CASIA B dataset for the validation of the proposed technique. The results are listed in Table 4 and verified through Table 5.
We also performed experiments on different feature sets to confirm the authenticity of the proposed heuristic-feature-selection approach. For this purpose, we select several feature vectors, such as 300 features, 400 features, 500 features, 600 features, 700 features, and a complete feature vector. Results are given in Table 6 for both datasets. This table shows that the results of 600 features are much better compared to other feature sets. The results increase initially, but after 600 features, accuracy is degraded by approximately 1%. For all features, the accuracy difference is almost 4%. Based on this table, we can say that the proposed technique achieves a significant performance on 600 features.
The comparison of the proposed method is also listed in Table 7. In this table, we only added those techniques in which the CASIA B dataset was used. For a real-time dataset, the comparison with recent techniques is not fair. This table shows that the previous best-reported accuracy on this dataset was 93.40% [6]. In this work, we have improved this accuracy by almost 2% and reached 95.26%.
Normally, researchers employ metaheuristic techniques, such as a genetic algorithm, PSO, ACO, and BCO [47]. These techniques have consumed too much time during the selection process. However, the proposed feature-selection approach is based on the single-kurtosis value activation function and is executed fast compared to the GA, PSO, BCO, Whale, and ACO on both the selected dataset, as shown by time plotted in Figure 15 and Figure 16. These figures show that the proposed feature-selection approach is more suitable for gait recognition in terms of computational time than the metaheuristic techniques.
A detailed analysis is also conducted through confidence intervals. For this purpose, the proposed framework is executed 500 times for both datasets. For the real-time dataset, we obtained three accuracy values—maximum (96.8%), minimum (95.3%), and average (96.05%). Based on these values, the standard deviation and standard error mean (SEM) are computed and obtain values of 0.75 and 0.5303. Similarly, for the CASIA B dataset, the standard deviation and standard error mean value are 0.56 and 0.3959, respectively. Using standard deviation and standard error mean values, the margin of error is computed, as plotted in Figure 17 and Figure 18. From these figures, it is shown that the proposed framework accuracy is consistent after the number of selected iterations. Moreover, the performance of the proposed framework (OvA-SVM) is also analyzed with a few other classifiers, such as Softmax, ELM, and KELM, as illustrated in Figure 19. From this figure, it can be noted that the proposed framework shows better performance.
The reason for the selected hyper parameters is as follows. Normally, the researchers employed gradient descent [48] as an optimization function, but due to the complex nature of the selected pre-trained model Resnet101 (residual blocks), the ADAM optimizer [49] can work better. The initial learning rate is normally 0.010, but in this work, the selected dataset dimension is high; therefore, the learning rate of 0.001 is suitable. Mini batch size is always selected based on the machine (desktop computer), but we do not have enough resources for the execution on a mini batch size of 128; therefore, we selected a value of 64. The maximum epochs are normally 70–100, but in this work, due to the higher number of video frames, we obtained a better training accuracy after 150 epochs. A few training results that were noted during the training process are given in Table 8 and Table 9 below.

5. Conclusions

Human gait recognition is an active research domain based on an important biometric application. Through gait, the human walking style can be easily determined in the video sequences. The major use of human gait recognition is in video surveillance, crime prevention, and biometrics. In this work, a deep-learning-based method is presented for human gait recognition in a real-time environment and for offline publicly available datasets. Transfer learning is applied for feature extraction and then selects robust features using a heuristic approach. The correlation formulation is applied for the best-selected feature fusion. In the end, multiclass OaA-SVM is applied for the final classification. The methodology is evaluated on a real-time captured database and CASIA B dataset, and it achieves an accuracy of 96.60% and 95.26%, respectively.
We can conclude that the preprocessing step before the model’s learning reduces the error rate. This step further shows strength in selecting the best features, but some features are discarded, which are essential for classification. The kurtosis-controlled entropy (KcE) is a new heuristic feature-selection technique that is executed in less time than the metaheuristic techniques. Another new technique named correlation-formulation-based fusion is used in this work for the best feature fusion. We compare the results of this method to existing methods such as PCA and LDA, and our newly proposed selection technique gives better results and has a shorter computational time for real-time video sequences. Moreover, the fusion process, through correlation formulation, increased the information of the human walking style that later helped the improved gait-recognition accuracy.
The drawback of this work is as follows: in the correlation-formulation-based feature fusion step, the method added some redundant features that later increased the computational time and reduced the recognition accuracy slightly. In the future, we shall improve the fusion approach and increase the database by adding data from more subjects.

Author Contributions

Conceptualization, M.I.S., M.A.K. and M.N.; methodology, M.I.S., M.A.K. and M.N.; software, M.I.S., M.A.K. and M.N.; validation, A.A., S.A. and A.B.; formal analysis, A.A., S.A. and A.B.; investigation, A.A., S.A. and A.B.; resources, A.A., S.A. and A.B.; data curation, M.A.K. and R.D.; writing—original draft preparation, M.A.K., M.I.S. and M.N.; writing—review and editing, R.D. and A.A.; visualization, R.D. and S.A.; supervision, R.D.; project administration, R.D. and A.B.; funding acquisition, R.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data used in this study are available from: https://drive.google.com/file/d/1g0UBj35Gu0HyjGhdid1NMobn2mPk8PyN/view?usp=sharing (accessed on 20 December 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Saleem, F.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Armghan, A.; Alenezi, F.; Choi, J.-I.; Kadry, S. Human Gait Recognition: A Single Stream Optimal Deep Learning Features Fusion. Sensors 2021, 21, 7584. [Google Scholar] [CrossRef] [PubMed]
  2. Bendjillali, R.I.; Beladgham, M.; Merit, K.; Taleb-Ahmed, A. Improved Facial Expression Recognition Based on DWT Feature for Deep CNN. Electronics 2019, 8, 324. [Google Scholar] [CrossRef] [Green Version]
  3. Adjabi, I.; Ouahabi, A.; Benzaoui, A.; Taleb-Ahmed, A. Past, Present, and Future of Face Recognition: A Review. Electronics 2020, 9, 1188. [Google Scholar] [CrossRef]
  4. Olanrewaju, L.; Oyebiyi, O.; Misra, S.; Maskeliunas, R.; Damasevicius, R. Secure ear biometrics using circular kernel principal component analysis, Chebyshev transform hashing and Bose–Chaudhuri–Hocquenghem error-correcting codes. Signal Image Video Process. 2020, 14, 847–855. [Google Scholar] [CrossRef]
  5. Rodrigues, J.D.C.; Filho, P.P.R.; Damasevicius, R.; de Albuquerque, V.H.C. EEG-based biometric systems. In Neurotechnology: Methods, Advances and Applications; The Institution of Engineering and Technology: London, UK, 2020; pp. 97–153. Available online: https://www.researchgate.net/publication/340455635_Neurotechnology_Methods_advances_and_applications (accessed on 17 January 2022).
  6. Arshad, H.; Khan, M.A.; Sharif, M.I.; Yasmin, M.; Tavares, J.M.R.S.; Zhang, Y.-D.; Satapathy, S.C. A multilevel paradigm for deep convolutional neural network features selection with an application to human gait recognition. Expert Syst. 2020, 20, 1–21. [Google Scholar] [CrossRef]
  7. Sokolova, A.; Konushin, A. Pose-based deep gait recognition. IET Biom. 2018, 8, 134–143. [Google Scholar] [CrossRef] [Green Version]
  8. Wang, L.; Li, Y.; Xiong, F.; Zhang, W. Gait Recognition Using Optical Motion Capture: A Decision Fusion Based Method. Sensors 2021, 21, 3496. [Google Scholar] [CrossRef]
  9. Kim, H.; Kim, H.-J.; Park, J.; Ryu, J.-K.; Kim, S.-C. Recognition of Fine-Grained Walking Patterns Using a Smartwatch with Deep Attentive Neural Networks. Sensors 2021, 21, 6393. [Google Scholar] [CrossRef]
  10. Hwang, T.-H.; Effenberg, A.O. Head Trajectory Diagrams for Gait Symmetry Analysis Using a Single Head-Worn IMU. Sensors 2021, 21, 6621. [Google Scholar] [CrossRef]
  11. Khan, M.H.; Li, F.; Farid, M.S.; Grzegorzek, M. Gait recognition using motion trajectory analysis. In Proceedings of the International Conference on Computer Recognition Systems, Polanica Zdroj, Poland, 22–24 May 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 73–82. [Google Scholar]
  12. Manssor, S.A.; Sun, S.; Elhassan, M.A. Real-Time Human Recognition at Night via Integrated Face and Gait Recognition Technologies. Sensors 2021, 21, 4323. [Google Scholar] [CrossRef]
  13. Liao, R.; Yu, S.; An, W.; Huang, Y. A model-based gait recognition method with body pose and human prior knowledge. Pattern Recognit. 2019, 98, 107069. [Google Scholar] [CrossRef]
  14. Sharif, M.; Attique, M.; Tahir, M.Z.; Yasmim, M.; Saba, T.; Tanik, U.J. A Machine Learning Method with Threshold Based Parallel Feature Fusion and Feature Selection for Automated Gait Recognition. J. Organ. End User Comput. 2020, 32, 67–92. [Google Scholar] [CrossRef]
  15. Priya, S.J.; Rani, A.J.; Subathra, M.S.P.; Mohammed, M.A.; Damaševičius, R.; Ubendran, N. Local Pattern Transformation Based Feature Extraction for Recognition of Parkinson’s Disease Based on Gait Signals. Diagnostics 2021, 11, 1395. [Google Scholar] [CrossRef] [PubMed]
  16. Khan, S.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Yong, H.-S.; Armghan, A.; Alenezi, F. Human Action Recognition: A Paradigm of Best Deep Learning Features Selection and Serial Based Extended Fusion. Sensors 2021, 21, 7941. [Google Scholar] [CrossRef]
  17. Khan, M.A.; Zhang, Y.-D.; Khan, S.A.; Attique, M.; Rehman, A.; Seo, S. A resource conscious human action recognition framework using 26-layered deep convolutional neural network. Multimed. Tools Appl. 2021, 80, 35827–35849. [Google Scholar] [CrossRef]
  18. Khan, M.A.; Javed, K.; Khan, S.A.; Saba, T.; Habib, U.; Khan, J.A.; Abbasi, A.A. Human action recognition using fusion of multiview and deep features: An application to video surveillance. Multimed. Tools Appl. 2020, 1–27. [Google Scholar] [CrossRef]
  19. Zebari, D.A.; Ibrahim, D.A.; Zeebaree, D.Q.; Haron, H.; Salih, M.S.; Damaševičius, R.; Mohammed, M.A. Systematic Review of Computing Approaches for Breast Cancer Detection Based Computer Aided Diagnosis Using Mammogram Images. Appl. Artif. Intell. 2021, 1–47. [Google Scholar] [CrossRef]
  20. Kassem, M.; Hosny, K.; Damaševičius, R.; Eltoukhy, M. Machine Learning and Deep Learning Methods for Skin Lesion Classification and Diagnosis: A Systematic Review. Diagnostics 2021, 11, 1390. [Google Scholar] [CrossRef]
  21. Kumar, V.; Singh, D.; Kaur, M.; Damaševičius, R. Overview of current state of research on the application of artificial intelligence techniques for COVID-19. PeerJ Comput. Sci. 2021, 7, e564. [Google Scholar] [CrossRef]
  22. Mehmood, A.; Khan, M.A.; Sharif, M.; Khan, S.A.; Shaheen, M.; Saba, T.; Riaz, N.; Ashraf, I. Prosperous Human Gait Recognition: An end-to-end system based on pre-trained CNN features selection. Multimed. Tools Appl. 2020, 11, 1–21. [Google Scholar] [CrossRef]
  23. Anusha, R.; Jaidhar, C.D. Clothing invariant human gait recognition using modified local optimal oriented pattern binary descriptor. Multimed. Tools Appl. 2020, 79, 2873–2896. [Google Scholar] [CrossRef]
  24. Khan, M.A.; Akram, T.; Sharif, M.; Muhammad, N.; Javed, M.Y.; Naqvi, S.R. Improved strategy for human action recognition; experiencing a cascaded design. IET Image Process. 2020, 14, 818–829. [Google Scholar] [CrossRef]
  25. Kadry, S.; Rajinikanth, V.; Taniar, D.; Damaševičius, R.; Valencia, X.P.B. Automated segmentation of leukocyte from hematological images—A study using various CNN schemes. J. Supercomput. 2021, 1–21. [Google Scholar] [CrossRef]
  26. Tanveer, M.; Rashid, A.H.; Ganaie, M.; Reza, M.; Razzak, I.; Hua, K.-L. Classification of Alzheimer’s disease using ensemble of deep neural networks trained through transfer learning. IEEE J. Biomed. Health Inform. 2021. [Google Scholar] [CrossRef]
  27. Khan, M.Z.; Saba, T.; Razzak, I.; Rehman, A.; Bahaj, S.A. Hot-Spot Zone Detection to Tackle Covid19 Spread by Fusing the Traditional Machine Learning and Deep Learning Approaches of Computer Vision. IEEE Access 2021, 9, 100040–100049. [Google Scholar] [CrossRef]
  28. Rehman, A.; Naz, S.; Razzak, M.I.; Hameed, I.A. Automatic Visual Features for Writer Identification: A Deep Learning Approach. IEEE Access 2019, 7, 17149–17157. [Google Scholar] [CrossRef]
  29. Alyasseri, Z.A.A.; Al-Betar, M.A.; Abu Doush, I.; Awadallah, M.A.; Abasi, A.K.; Makhadmeh, S.N.; Alomari, O.A.; Abdulkareem, K.H.; Adam, A.; Damasevicius, R.; et al. Review on COVID-19 diagnosis models based on machine learning and deep learning approaches. Expert Syst. 2021, e12759. [Google Scholar] [CrossRef]
  30. Castro, F.M.; Marín-Jiménez, M.J.; Guil, N.; De La Blanca, N.P. Multimodal feature fusion for CNN-based gait recognition: An empirical comparison. Neural Comput. Appl. 2020, 32, 14173–14193. [Google Scholar] [CrossRef] [Green Version]
  31. Rawat, W.; Wang, Z. Deep convolutional neural networks for image classification: A comprehensive review. Neural Comput. 2017, 29, 2352–2449. [Google Scholar] [CrossRef]
  32. Bari, A.S.M.H.; Gavrilova, M.L. Artificial Neural Network Based Gait Recognition Using Kinect Sensor. IEEE Access 2019, 7, 162708–162722. [Google Scholar] [CrossRef]
  33. Zheng, S.; Zhang, J.; Huang, K.; He, R.; Tan, T. Robust view transformation model for gait recognition. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 2073–2076. [Google Scholar]
  34. Wang, X.; Yan, W.Q. Human Gait Recognition Based on Frame-by-Frame Gait Energy Images and Convolutional Long Short-Term Memory. Int. J. Neural Syst. 2019, 30, 1950027. [Google Scholar] [CrossRef] [PubMed]
  35. Anusha, R.; Jaidhar, C.D. Human gait recognition based on histogram of oriented gradients and Haralick texture descriptor. Multimed. Tools Appl. 2020, 79, 8213–8234. [Google Scholar] [CrossRef]
  36. Zhao, A.; Li, J.; Ahmed, M. SpiderNet: A spiderweb graph neural network for multi-view gait recognition. Knowl. Based Syst. 2020, 206, 106273. [Google Scholar] [CrossRef]
  37. Arshad, H.; Khan, M.A.; Sharif, M.; Yasmin, M.; Javed, M.Y. Multi-level features fusion and selection for human gait recognition: An optimized framework of Bayesian model and binomial distribution. Int. J. Mach. Learn. Cybern. 2019, 10, 3601–3618. [Google Scholar] [CrossRef]
  38. Ferroukhi, M.; Ouahabi, A.; Attari, M.; Habchi, Y.; Taleb-Ahmed, A. Medical Video Coding Based on 2nd-Generation Wavelets: Performance Evaluation. Electronics 2019, 8, 88. [Google Scholar] [CrossRef] [Green Version]
  39. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  40. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. arXiv 2016, arXiv:1602.07261. [Google Scholar]
  41. Torrey, L.; Shavlik, J. Transfer learning. In Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods, and Techniques; IGI Global: Hershey, PA, USA, 2010; pp. 242–264. [Google Scholar]
  42. Naheed, N.; Shaheen, M.; Khan, S.A.; Alawairdhi, M.; Khan, M.A. Importance of Features Selection, Attributes Selection, Challenges and Future Directions for Medical Imaging Data: A Review. Comput. Model. Eng. Sci. 2020, 125, 315–344. [Google Scholar] [CrossRef]
  43. Liu, Y.; Zheng, Y.F. One-against-all multi-class SVM classification using reliability measures. In Proceedings of the 2005 IEEE International Joint Conference on Neural Networks, Montreal, QC, Canada, 31 July–4 August 2005; pp. 849–854. [Google Scholar]
  44. Arai, K.; Andrie, R. Gait recognition method based on wavelet transformation and its evaluation with chinese academy of sciences (casia) gait database as a human gait recognition dataset. In Proceedings of the 2012 Ninth International Conference on Information Technology-New Generations, Las Vegas, NV, USA, 16–18 April 2012; pp. 656–661. [Google Scholar]
  45. Arora, P.; Hanmandlu, M.; Srivastava, S. Gait based authentication using gait information image features. Pattern Recognit. Lett. 2015, 68, 336–342. [Google Scholar] [CrossRef]
  46. Castro, F.M.; Marín-Jiménez, M.J.; Mata, N.G.; Muñoz-Salinas, R. Fisher motion descriptor for multiview gait recognition. Int. J. Pattern Recognit. Artif. Intell. 2017, 31, 1756002. [Google Scholar] [CrossRef] [Green Version]
  47. Agrawal, P.; Abutarboush, H.F.; Ganesh, T.; Mohamed, A.W. Metaheuristic Algorithms on Feature Selection: A Survey of One Decade of Research (2009–2019). IEEE Access 2021, 9, 26766–26791. [Google Scholar] [CrossRef]
  48. Amari, S.-I. Backpropagation and stochastic gradient descent method. Neurocomputing 1993, 5, 185–196. [Google Scholar] [CrossRef]
  49. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. A simple CNN architecture for gait recognition.
Figure 1. A simple CNN architecture for gait recognition.
Electronics 11 00334 g001
Figure 2. Proposed architecture of real-time human gait recognition using deep learning.
Figure 2. Proposed architecture of real-time human gait recognition using deep learning.
Electronics 11 00334 g002
Figure 3. Original ResNet101 pre-trained CNN model for image classification.
Figure 3. Original ResNet101 pre-trained CNN model for image classification.
Electronics 11 00334 g003
Figure 4. Transfer-learning-based model retraining on real-time captured video frames for HGR.
Figure 4. Transfer-learning-based model retraining on real-time captured video frames for HGR.
Electronics 11 00334 g004
Figure 5. Proposed system predicted labeling results.
Figure 5. Proposed system predicted labeling results.
Electronics 11 00334 g005
Figure 6. Sample image frames of real-time video sequences.
Figure 6. Sample image frames of real-time video sequences.
Electronics 11 00334 g006
Figure 7. Sample frames collected from CASIA B dataset [44].
Figure 7. Sample frames collected from CASIA B dataset [44].
Electronics 11 00334 g007
Figure 8. Confusion matrix of OaA-SVM for normal walk on different selected angles.
Figure 8. Confusion matrix of OaA-SVM for normal walk on different selected angles.
Electronics 11 00334 g008
Figure 9. Variation in computational time for selected classifiers on Real time dataset.
Figure 9. Variation in computational time for selected classifiers on Real time dataset.
Electronics 11 00334 g009
Figure 10. Confusion matrix of OaA-SVM for walking while carrying a bag at different selected angles.
Figure 10. Confusion matrix of OaA-SVM for walking while carrying a bag at different selected angles.
Electronics 11 00334 g010
Figure 11. Variation in computational time for selected classifiers of 4 selected angles on Real time dataset.
Figure 11. Variation in computational time for selected classifiers of 4 selected angles on Real time dataset.
Electronics 11 00334 g011
Figure 12. Confusion matrix of OaA-SVM for normal walking and walking while carrying a bag classification.
Figure 12. Confusion matrix of OaA-SVM for normal walking and walking while carrying a bag classification.
Electronics 11 00334 g012
Figure 13. Execution time of proposed method on real-time dataset (normal walking and walking while carrying a bag).
Figure 13. Execution time of proposed method on real-time dataset (normal walking and walking while carrying a bag).
Electronics 11 00334 g013
Figure 14. Execution time of proposed method for CASIA B dataset (90° angle).
Figure 14. Execution time of proposed method for CASIA B dataset (90° angle).
Electronics 11 00334 g014
Figure 15. Time comparison for the real-time dataset.
Figure 15. Time comparison for the real-time dataset.
Electronics 11 00334 g015
Figure 16. Time comparison on the CASIA B dataset.
Figure 16. Time comparison on the CASIA B dataset.
Electronics 11 00334 g016
Figure 17. Confidence interval based analysis of proposed method on real-time dataset.
Figure 17. Confidence interval based analysis of proposed method on real-time dataset.
Electronics 11 00334 g017
Figure 18. Confidence-interval-based analysis of proposed method on CASIA B dataset.
Figure 18. Confidence-interval-based analysis of proposed method on CASIA B dataset.
Electronics 11 00334 g018
Figure 19. Comparison of OvS-SVM classifier with other classification algorithms for gait recognition.
Figure 19. Comparison of OvS-SVM classifier with other classification algorithms for gait recognition.
Electronics 11 00334 g019
Table 1. Proposed gait-recognition results using real-time captured dataset for normal walk on selected angles.
Table 1. Proposed gait-recognition results using real-time captured dataset for normal walk on selected angles.
ClassifiersPerformance Measures
Recall
(%)
Precision (%)FI Score
(%)
AUCFPRAccuracy
(%)
Time
(s)
OaA-SVM95.7596.2595.981.000.012596.0204.050
Cubic KNN94.2594.7594.481.000.017594.5366.480
Medium KNN95.0095.5095.241.000.017594.9175.870
Bagged Trees93.5093.7593.620.990.020093.5345.750
CG-SVM90.2590.5090.360.990.032590.1189.300
Fine Tree86.7586.5086.620.920.042586.719.565
Medium Tree85.5085.7585.620.920.047585.439.907
Naïve Bayes84.5087.2585.840.900.050084.334.035
Coarse KNN61.5072.5066.540.920.127561.5189.760
FG-SVM61.0084.7570.920.910.130060.9132.050
Kernel Bayes48.7563.5050.540.730.172548.5582.710
Table 2. Proposed gait-recognition results using real-time captured dataset for walking while carrying a bag at selected angles.
Table 2. Proposed gait-recognition results using real-time captured dataset for walking while carrying a bag at selected angles.
ClassifiersPerformance Measures
Recall (%)Precision (%) FI Score (%)AUCFPRAccuracy (%) Time (s)
OaA-SVM96.597.096.71.000.0196.6189.850
Cubic KNN95.095.295.11.000.0195.1274.850
Medium KNN95.095.295.11.000.0195.19.219
Bagged Trees94.795.094.80.990.0194.9202.010
CG-SVM90.591.090.70.990.0390.5137.480
Fine Tree87.781.084.20.930.0487.614.190
Medium Tree86.787.086.80.930.0486.813.590
Naïve Bayes83.786.785.20.890.0583.720.920
Coarse KNN63.074.068.00.920.1263.0169.521
FG-SVM61.084.770.90.910.1361.327.030
Kernel Bayes48.760.253.80.750.1748.6483.800
Table 3. Recognition results for all datasets with kurtosis function on 600 predictions on original deep-feature vector extracted from ResNet 101.
Table 3. Recognition results for all datasets with kurtosis function on 600 predictions on original deep-feature vector extracted from ResNet 101.
ClassifiersPerformance Measures
Recall
(%)
Precision
(%)
FI Score
(%)
AUCFPRAccuracy
(%)
Time
(s)
OaA-SVM96.596.596.51.000.0396.4129.830
Cubic KNN95.096.095.71.000.0495.6278.980
Medium KNN96.096.096.01.000.0496.1106.850
Bagged Trees90.590.590.50.970.0990.3209.610
CG-SVM82.082.582.20.910.1882.2102.260
Fine Tree80.080.080.00.810.2079.811.954
Medium Tree76.076.576.20.800.2476.330.888
Naïve Bayes68.069.068.40.780.3267.818.767
Coarse KNN79.079.579.20.880.2179.3117.990
FG-SVM79.585.582.30.980.4179.674.026
Kernel Bayes73.074.573.70.840.2773.0254.580
Table 4. Proposed recognition results on CASIA B dataset, where selected angle is 90°.
Table 4. Proposed recognition results on CASIA B dataset, where selected angle is 90°.
ClassifierFeaturesMeasures
GAPFCProposedRecall (%)Accuracy (%)Time (s)
OaA-SVM 90.1090.22242.4426
88.5288.64176.4450
95.1095.26114.2004
Cubic KNN 84.4284.54165.5994
83.6083.98111.2011
93.6093.6082.1460
Medium KNN 83.4083.48151.0014
84.8084.76104.1446
93.4093.4664.2914
Baggage Tree 85.1085.16256.1130
84.1484.33201.0148
87.5087.45117.1106
Naïve Bayes 71.1071.04171.2540
74.9474.82104.3360
79.3079.3076.3114
Table 5. Confusion matrix of OaA-SVM for CASIA B dataset.
Table 5. Confusion matrix of OaA-SVM for CASIA B dataset.
Gait NameGait Name
Normal WalkW-CoatW-Bag
Normal Walk94%4%2%
W-Coat3%95%2%
W-Bag1%2%97%
Table 6. Analysis of proposed feature-selection framework on numerous feature sets. *The OvA-SVM is employed as a classifier for this table.
Table 6. Analysis of proposed feature-selection framework on numerous feature sets. *The OvA-SVM is employed as a classifier for this table.
DatasetAccuracy (%) on Feature Sets
300 Features400 Features500 Features600 Features700 FeaturesAll Features
Real-time (normal walking)93.7094.2495.3596.0095.7093.04
Real-time (walking while carrying a bag)94.1094.9095.8096.6096.3292.10
Real-time (normal walking vs. walking while carrying a bag)92.9093.7295.3096.4096.1493.50
CASIA B Dataset92.9693.4093.8595.2695.1092.64
Table 7. Comparison with existing techniques for CASIA B dataset.
Table 7. Comparison with existing techniques for CASIA B dataset.
ReferenceYearDatasetAccuracy (%)
[45]2015CASIA B86.30
[46]2017CASIA B90.60
[37]2019CASIA B87.7
[6]2020CASIA B93.40
ProposedCASIA B95.26
ProposedReal-time96.60
Table 8. Training accuracy on selected hyper parameters using real-time collected dataset.
Table 8. Training accuracy on selected hyper parameters using real-time collected dataset.
EpochsAccuracy (%)Error (%)Time (min)
2083.516.5221.6784
4087.912.1375.7994
6090.29.8588.7834
8092.65.4792.5673
10093.95.1875.1247
15096.81.2988.0045
20098.10.41105.5683
Table 9. Training accuracy on selected hyper parameters using CASIA B dataset.
Table 9. Training accuracy on selected hyper parameters using CASIA B dataset.
EpochsAccuracy (%)Error (%)Time (min)
2081.418.6174.8957
4084.615.4292.0645
6088.012411.4756
8090.29.8581.8322
10091.68.4695.4570
15094.35.7808.5334
20097.52.5981.6873
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Sharif, M.I.; Khan, M.A.; Alqahtani, A.; Nazir, M.; Alsubai, S.; Binbusayyis, A.; Damaševičius, R. Deep Learning and Kurtosis-Controlled, Entropy-Based Framework for Human Gait Recognition Using Video Sequences. Electronics 2022, 11, 334. https://doi.org/10.3390/electronics11030334

AMA Style

Sharif MI, Khan MA, Alqahtani A, Nazir M, Alsubai S, Binbusayyis A, Damaševičius R. Deep Learning and Kurtosis-Controlled, Entropy-Based Framework for Human Gait Recognition Using Video Sequences. Electronics. 2022; 11(3):334. https://doi.org/10.3390/electronics11030334

Chicago/Turabian Style

Sharif, Muhammad Imran, Muhammad Attique Khan, Abdullah Alqahtani, Muhammad Nazir, Shtwai Alsubai, Adel Binbusayyis, and Robertas Damaševičius. 2022. "Deep Learning and Kurtosis-Controlled, Entropy-Based Framework for Human Gait Recognition Using Video Sequences" Electronics 11, no. 3: 334. https://doi.org/10.3390/electronics11030334

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop