Next Article in Journal
A Hybrid State/Disturbance Observer-Based Feedback Control of Robot with Multiple Constraints
Next Article in Special Issue
Brickognize: Applying Photo-Realistic Image Synthesis for Lego Bricks Recognition with Limited Data
Previous Article in Journal
Assessment of Intestinal Ischemia–Reperfusion Injury Using Diffuse Reflectance VIS-NIR Spectroscopy and Histology
Previous Article in Special Issue
A Study on Fast and Low-Complexity Algorithms for Versatile Video Coding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Gait-Based Real-Time Gender Classification System Using Whole Body Joints

1
Department of Computer Science & IT, University of Malakand, Chakdara 18800, Pakistan
2
Department of Software Engineering, University of Malakand, Chakdara 18800, Pakistan
3
School of Computing, Gachon University, 1342 Seongnam-daero, Sujeong-gu, Seongnam-si 13120, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(23), 9113; https://doi.org/10.3390/s22239113
Submission received: 16 September 2022 / Revised: 17 November 2022 / Accepted: 20 November 2022 / Published: 24 November 2022
(This article belongs to the Special Issue Applications of Video Processing and Computer Vision Sensor II)

Abstract

:
Gait-based gender classification is a challenging task since people may walk in different directions with varying speed, gait style, and occluded joints. The majority of research studies in the literature focused on gender-specific joints, while there is less attention on the comparison of all of a body’s joints. To consider all of the joints, it is essential to determine a person’s gender based on their gait using a Kinect sensor. This paper proposes a logistic-regression-based machine learning model using whole body joints for gender classification. The proposed method consists of different phases including gait feature extraction based on three dimensional (3D) positions, feature selection, and classification of human gender. The Kinect sensor is used to extract 3D features of different joints. Different statistical tools such as Cronbach’s alpha, correlation, t-test, and ANOVA techniques are exploited to select significant joints. The Coronbach’s alpha technique yields an average result of 99.74%, which indicates the reliability of joints. Similarly, the correlation results indicate that there is significant difference between male and female joints during gait. t-test and ANOVA approaches demonstrate that all twenty joints are statistically significant for gender classification, because the p-value for each joint is zero and less than 1%. Finally, classification is performed based on the selected features using binary logistic regression model. A total of hundred (100) volunteers participated in the experiments in real scenario. The suggested method successfully classifies gender based on 3D features recorded in real-time using machine learning classifier with an accuracy of 98.0% using all body joints. The proposed method outperformed the existing systems which mostly rely on digital images.

1. Introduction

The term “gait” describes a person’s manner of walking. Every person has a unique gait that can be used as a behavioral trait to identify them. Since gait capture may be performed from a distance, the system does not require any assistance from the person. Gait recognition has therefore become more common in recent years and has a variety of applications, including forensics, security systems, surveillance, and the investigation of criminal activity. As CCTV is currently implemented in nearly all public areas [1,2,3], it unobtrusively records the person’s gait.
In the existing research work, the majority of the literature has focused on gait recognition [4,5,6,7,8,9,10,11,12,13,14], whereas gender classification based on gait offers a lot of potential for real-time applications [15]. The difficulty of discovering lost people, especially youngsters, at train stations, airports, bus stops, and other public locations can be solved with the help of gait-based gender classification, which can act as a clue in an investigation. The gender classification based on gait can also be applied in commercial activities. For instance, relevant advertisements can be shown on the digital screen if the gender of the person entering the store is known. A talking robot that has been installed in supermarkets can recognize customers through cameras and assist them with gender-specific needs including finding washrooms, food stores, men’s wear stores, and women’s wear stores.
The majority of gender classification studies make use of a different number of biometric features for gender classification, e.g., Pries et al. [16] used 13 features while Alharbi et al. [17] used 7 different joints. In the literature, few joints are often used to achieve gait-based gender recognition; therefore, little focus has been given to studying the gender differences based on all joints while walking. Unfortunately, very few studies in the recent research literature have used a self-created dataset for gender classification, which may be thought of as the main motivation for this work.

1.1. Research Gap and Motivation

  • The existing gait recognition approaches have compared limited joints of male and female during gait and less focus has been paid for complete study/analysis of joints.
  • Gender has been classified based on gait in recent research [18,19,20,21,22], although the accuracy of this classification mostly depends on 2D camera images.
  • Many researchers have recognized the gender of walking participants [23,24,25,26], but they have analyzed the results using a comparatively limited dataset, which may adversely affect the system’s accuracy.

1.2. Main Contributions

The following are the key innovations and contributions of our work:
  • Using information obtained from MS Kinect, the authors created their own data set for gender classification.
  • Applying a straightforward binary logistic regression model to the challenging issue of gender classification with better accuracy.
  • The suggested system provides better accuracy as compared to the state-of-the-art system using the same data set.
In short, there is little information on gender-related differences in joints during casual walk; therefore, the purpose of this study is to compare/investigate every joint difference that contributes during walking. Finally, the gender of the walking is classified based on the complete set of joints positions. This paper is organized as follows. Section 2, presents related work, Section 3 describes proposed work, and Section 4 describes experiments and results. Finally, Section 5 is about conclusion and future work.

2. Related Work

Gender classification is performed in a number of ways using gait. According to the format of the gait, the existing gait identification techniques can be classified into 2D-based and 3D-based techniques [27,28,29]. The 2D-based gait identification techniques depend on a human silhouette, and is often collected by a single RGB camera in video surveillance. RGB-image-based methods are mostly used in this area of gait recognition. These methods are often categorized as model-free and model-based [18,23,30,31,32].
Model-free methods are often referred to as appearance-based techniques. The silhouettes, which are removed from the video sequences via background subtraction, are used to directly generate gait signatures. A gait energy image (GEI) is the most common representation of gait based on appearance [6], whereas 3D-based approaches typically make use of numerous calibrated cameras or cameras with depth sensors to extract 3D gait data. Moreover, 2D-based methods are typically simpler to execute because they simply require a normal camera. The 3D skeleton model depicted was suggested by Zhao et al. [33] and is based on 10 joints and 24 degrees of freedom (DOFs) recorded by various cameras. Yamauchi et al. [34] captured the dense 3D range gait, which may be used to identify people in various orientations. A system with four calibrated and synchronized cameras was developed by Krzeszowski et al. [35] to predict 3D gait motion from video sequences and identify view-variant gaits.
In terms of identifying gaits, both 2D-based and 3D-based approaches offer significant advantages; however, 3D-based approaches are more realistic as compared to 2D-based/image-based approaches because their robustness against the variation of observation perspective [36], walking speed [37], and clothing [38] is better than a 2D-based system.
Recently, the majority of researchers have focused on the view-invariant Kinect-based gait detection method because it is challenging to extract viewpoint information from only silhouettes [39]. In order to identify persons, Kale et al. [40] introduced a gait recognition algorithm based on static body metrics that are collected from the walking across multiple views. A framework was presented by Jean et al. [41] to compute and assess the view-normalized trajectories of the head and feet as seen in monocular videos. Hu et al. [42] suggested using a unitary linear projection to reduce the dimensions of the original gait parameters that were taken from any view(s) and increase their capacity to differentiate between different views. Because of its ability to collect skeletons with less distortion from clothes and lighting than prior depth sensors, the Kinect has become a popular tool for gait detection. Sivapalan et al. [43] used the Kinect sensor to convert the GEI from a 2D to a 3D representation. The lengths of body parts were calculated from joint positions and used as static anthropometric features by Araujo et al. [44] for gait recognition. The coordinates of all the joints recorded by the Kinect were used to produce an RGB image, which was then combined into a video to represent the walking sequence. This method was used by Milos et al. [45] to detect the gait. Price et al. [16] evaluated 11 skeletal features, such as step length and speed, which were used by Kinect as dynamic features; both static and dynamic features were combined for recognition. Yang et al. [46] developed relative distance-based gait features as an alternative to anthropometric measurements, which are better able to retain the periodic patterns of gait. Ahmed et al. [47] developed a gait signature using the total number of joint relative angles (JRA) sequences throughout a walking cycle. In order to determine the distance between two JRA sequences, they also created a dynamic temporal warping (DTW)-based kernel. Kastaniotis et al. [48] proposed a strategy for leveraging the Kinect for gait-based recognition. A person can be identified with 90% accuracy using a Kinect-based gait recognition system, as suggested by [16], which uses 13 biometric parameters. Similarly, X et al. [49] used 18 dynamic features that covered the dynamic angles in the lower limp joints. The system’s accuracy was still just 43.6%. A similar approach [50] used 14 features to identify objects, mostly based on the horizontal distance feature (HDF) and vertical distance feature (VDF). The system’s recognition accuracy was 92% when using the KNN method. An SVM-based method for gender classification using the Kinect obtained an average accuracy of 87% [51]. Azhar et al. used a limited set of joints in [52] for gender classification in real-time using Kinect sensor. With five subjects, another human identification system that used skeleton data from the Kinect sensor showed an accuracy of 86% [53]. Similarly, the researchers in [47] used joint relative angles (JRA) technique for 3D skeleton Kinect-based gait identification system with a 93.3% recognition rate. A different gait recognition and gender classification method based on the Kinect was suggested in [54]. The KNN classifier produced results with a reasonable accuracy of 96%. A gait-based recognition system for identifying individuals in baggy and lengthy clothing was proposed in [17]. Using nine joints the system used SVM and claimed a 97% accurate result.
The existing gait recognition approaches have compared limited joints of male and female during gait and less focus has been paid for complete study/analysis of joints. Furthermore, the majority of researchers have created systems that classify both age and gender in a single system, but they have typically divided age into various groups (classes), creating the issue of a classification system with a restricted age range. The key innovations and contributions of our work is the creation of our own dataset, comparison of all body joints of male and female using different statistical tools, comparison of the existing systems with the state-of-the-art systems and applying a straight forward binary logistic model to the challenging issue of gender classification with better accuracy.

3. Proposed System

This paper presents a comparative method for calculating the contribution of male and female body joints in walking. The contribution of male body joints is then compared to the contribution of female body joints using statistical methods such as correlation, t-test and ANOVA techniques. Then, Cronbach’s alpha technique is applied to find the reliability of joints. The findings also include a comprehensive list of all human body joints and their roles in gender classification. Furthermore, the proposed system model has two working stages i.e., training and testing. In the training phase, the 3D position of the gender are recorded through MS Kinect to find the best fitted model for gender classification. Finally, in the testing phase, the gender of the walking subject is recorded using percentage of the most contributed body joints via the binary logistic-regression-based machine learning model. The proposed system model is shown in Figure 1.

3.1. Dataset Generation

Since no other datasets provided both skeleton joint position and sufficient gender information, we had to develop our own. In this study, 373 participants ranging in age from 7 to 70 took part in the study. Inside the Kinect recognition space, the authors instructed each participant to walk in a free position while facing various directions (see Figure 2).
A cheap MS Kinect V1 camera is utilized to record the three dimensional (3D) joint movements of subjects. This sensor not only captures the traditional RGB image of the object being shown, but also the depth information between the objects and the camera. The Microsoft Kinect sensor is made up of a 3D depth sensor, RGB camera, tilt motor, and a Microsoft array. The Kinect sensor extracted the human skeleton 3D joints using the Kinect software development kit (SDK). The Kinect SDK uses Kinect skeletal tracking algorithm for extracting the 20 human joints. As seen in Figure 3, MS Kinect offers a full set of 20 usable joints, ranging from the head to the foot. Multiple records for each individual were taken in order to evaluate MS Kinect’s accuracy. The measurement was calculated from several distances, and 6 feet was found to be the ideal distance. The MS Kinect had a height of 3.4 feet and an angle of 0 degrees at 30 frames per second (fps). Every volunteer was free to adopt any behavior or stance at any time. There was no time restriction because the feature extraction process used the average of all frames.

3.2. Pre-Processing of Data

Our system for recording human gait consists of a single Kinect depth camera installed on a tripod and placed in a naturally lighted room. Humans are seen moving smoothly along the length of the room as the camera records their frontal motion and walking direction. Because the field of vision is 3 m and the depth field is 4.5 m, longer sequences can be captured. The Kinect makes it possible to capture a person at 30 fps and extract their 20-joint skeleton in real time; however, as our focus is on developing a reliable and effective gait processing approach, simple and generic low-cost acquisition setups are sufficient. A motion sequence in time t [ 1 T ] is represented by a collection of skeletons, each of which has 20 joints. We use technique to represent the j th joint at time t.
P i t = x j t y j t z j t
Equation (1) indicates that data can be recorded in 3D i.e., x, y, and z axis. We use the observation that the Euclidean distance between two joints should remain constant while the subject moves to create a scale-invariant skeleton. As a result, we normalize the skeleton joints in relation to how far they are from each other. Euclidean distance is a logical approach to compare two joints. The Euclidean distance between two joints of a body of a person during walk is defined as
d r s 2 = x ̲ r x ̲ s 2

3.2.1. Variance Covariance Matrix

In order to describe our gait data, we encode body joint data using a covariance matrix which calculates variance and covariance between various joints. We describe joints using a local relative coordinate system to achieve efficiency and consistency of body positions. In order to measure efficiency of each joint the sample variance–covariance matrix or covariance matrix is used as
S = s 11 s 12 s 1 p s 21 s 22 s 2 p s p 1 s p 2 s p p
The main diagonal values are referred to as the variances of particular joints and the remaining joints are the co-variances. The sample covariance matrix is calculated using the followingformula.
S = 1 n X . X X ̲ ¯ . X ̲ ¯
where S is covariance matrix, X ̲ ¯ is the mean vector and X is a data matrix of joints, and where
X ̲ ¯ = x 1 ¯ x 2 ¯ x p ¯

3.2.2. Correlation Matrix

The correlation coefficient between i th and j th joints is given by
r i j = s i j s i i s j j
The p × p matrix of correlation for all joints is defined as
R = r 11 r 12 r 1 p r 21 r 22 r 2 p r p 1 r p 2 r p p
where R = R , r i j = r j i for i j and 1 r i j + 1 . If R = I , then all joints are uncorrelated.
Alternately, some of the fundamental statistical techniques are used to assess the contribution and effectiveness of joints. In order to identify all skeletons as having the same component length and compute an average value for each of its skeletal parts, we also compute the joints’ mean, variance, and coefficient of variation in x, y, and z coordinates, respectively.
X ¯ H i p L = 1 n i = 1 n ( X i )
where i = 1 , 2 , 3 n and X ¯ H i p L represents arithmetic mean of hip-left joint. The arithmetic mean for other axis i.e., y and z axis can be determined in the same way using Equation (4). Further, to compute a single value of various joints i.e., combined mean of x, y, and z coordinates of individual person for comparison of male and female joints during gait. The combined mean is statistically defined for joint Hip left as follows:
X ¯ C = 1 3 ( X ¯ H i p L + Y ¯ H i p L + Z ¯ H i p L )
where X ¯ C is the combined mean of Hip left joint and X ¯ H i p L , Y ¯ H i p L and Z ¯ H i p L are the means for joint Hip left of x, y and z coordinates.

3.2.3. Variance of and Pooled Variance of Joints

To capture variation of joint features from person to person of male and female joints during gait, the authors calculated variances and standard deviation. In this phase, variance of all joints are calculated in the 3D axis e.g., variance of the left hip joint in x-axis is measured using Equation (10).
σ X H i p L 2 = 1 N i = 1 N ( X i μ X ) 2
where N is the size or number of observation in the population, σ 2 is the variance of population, X i is the variables, and μ X is the population mean at x coordinate. Similarly, the variance of hip left joint in y and z coordinates can be computed in the same way. The authors calculated the variances of other joints using Equation (11). After computing the variance of each joint, a pooled variance of all joints is calculated as
δ P 2 = 1 N ( N 1 δ 1 2 + N 2 δ 2 2 + N 3 δ 3 2 + + N k δ k 2 )
where δ P 2 is the pooled variance while δ 1 2 , δ 2 2 , δ 3 2 δ k 2 , are the separate variances of joints. Similarly N = N 1 + N 2 + N 3 , where N represents size of all joints and N 1 , N 2 , and N 3 are the sizes of x, y, and z axis. The pooled variances for joints H i p L may be computed as
δ P H i p L 2 = 1 N ( N 1 δ X H i p L 2 + N 2 δ Y H i p L 2 + N 3 δ Z H i p L 2 )
where N 1 , N 2 , and N 3 are the sizes of x, y, and z axis respectively. δ P H i p L 2 is the pooled variances of joints H i p L . Similarly, pooled variances of other joints can also be computed in the same way.

3.2.4. Coefficient of Variation of Joints

The coefficient of variation (C.V) is used to assess the consistent joints out of the twenty joints for both male and female. The coefficient of variation is defined as
C . V ( X H i p L ) = S . D ( X H i p L ) m e a n ( X H i p L ) 100
where S.D represents the standard deviation of the hip left joint and the mean corresponds to the mean of the hip left joint.

3.2.5. Average Variation in Particular Joints of Human during Gait

Mean, standard deviation, and coefficients of variation of male and female joints were computed by using various statistical software’s.
The results show comparatively a minimum score (10.31 and 9.05) for the coefficient of variation for the head joint of males and females. It suggests that the joint head is consistent joint for both male and female during the walk but more consistent for females as compared to males. On the other hand, C.V for the joint foot left of a male is a maximum of 17.58 as compared to the other joints, while C.V for the joint foot right of a female is also a maximum of 17.82 compared to other joints. It proves that the joints foot left and foot right of a male and a female vary more from individual to individual. C.V for the joints hand left and knee right of a male are same and also the C.V for joints spine and hip center of a female are the same, which show that these joints play the same role during gait. Table 1 shows the mean, standard deviation (S.D), and coefficient of variation (C.V) for various joints of male and female.

3.3. Comparison of Male and Female Joints

To assess the mean difference among all joints of male and female during the walk, different approaches are used.

3.3.1. Comparison of Male and Female Joints by Correlation

To identify the relationship between various joints and various directions during walk, the author used the Pearson correlation method. Pearson correlation is a statistical method used to measure the relationship between any two variables. The correlation is always lies between −1 and +1, the correlation near to +1 indicates that there is strong and direct relationship between variables and correlation near to −1 indicates that there is strong and indirect relationship. When the correlation is near to zero indicates that there is weak relationship between variables. The Pearson correlation is defined as
r = ( n x y x y ) [ ( n x 2 ( x ) 2 ) ( n y 2 ( y ) 2 ) ]
To know the difference between male and female joints during walk, the authors check the correlation between various joints. The significant difference between male and female joints was obtained from the correlation and there is high difference in the male and female joints during walk.
According to Table 2, there are strong and weak relationships between different joints of males and females during gait. During a walk, the relationship between x and y directions at HIP-center is 84% for males, while for females is 14%. Males in the HIP-center have a relationship between y and z directions of 88%, while females have 20%. Males at HIP-Center have 90% relationship between x and z directions, while females have 80%. The relationship between different joints in men and women is different. The relationship between x and z directions at joint Wrist left for males is 94%, while for females it is 79%. The relationship between x and z directions at joint Hand left for males is 93%, while for females is 72%. Similarly, the relationship between x and z directions at joint Knee right for males is 91%, while for females is 79%. The relationship between x and z directions at joint Ankle right is 93% for males, while 82% for females. The relationship between x and z directions at joint foot right is 93% for male joints, while for female joints, it is 88%. At joint HIP-Left, the correlation for males is 94% while for females it is 88%, which is a strong and significant relationship. The correlation results in Table 2 indicate that there is a significant difference between male and female joints during gait.

3.3.2. Comparison of Male and Female Joints Using t-test

To compare various joints of male and female during gait, the t-test is applied to identify the significant difference, i.e., at joint hip center there is significant difference between male and female walking. As each human can walk in three directions such as x, y, and z, therefore, we calculate mean of x, y, and z directions, respectively, and compute their combine mean. The combined means were compared by using t-test. Consider t-test of equal variances and not equal variances provide the same conclusion. The results of two samples t-test are shown in Table 3. The formulation of our hypothesis is given below.
Hypothesis 1. 
There is no difference at joint hip left of males and females during walk, i.e.,
μ H i p L ( M ) = μ H i p L ( F )
Hypothesis 2. 
There is significant difference at joint hip left for males and females during walk, i.e.,
μ H i p L ( M ) μ H i p L ( F )
where μ H i p L ( M ) is the mean of hip left for males and μ H i p L ( F ) is the mean of joint hip left for females. To compare each joint of males and females pair wise during gait, the t-test and ANOVA technique are used for all comparisons of all joints at the same time. Assume that variances are the same for both male and female joint contribution for t-test. The t-test is defined for analysis as
T = ( X ¯ 1 i X ¯ 2 i ) s p 1 n 1 + 1 n 2
where X ¯ 1 is the combined mean of each joint of male, X ¯ 2 is combined mean of each joint for female. n 1 is the size of males and n 2 is the size of females. The degree of freedom for t-test is
ν = n 1 + n 2 2
where ν represents the degree of freedom. S p is the pooled standard deviation for both male and female joints and is defined as
S p = ( n 1 1 ) s 1 2 + ( n 2 1 ) s 2 2 n 1 + n 2 2
S 1 = ( x i ( M ) x ¯ ( M ) ) 2 n 1 1
S 2 = ( x i ( F ) x ¯ ( F ) ) 2 n 2 1
where s 1 2 and s 2 2 are the variances of various joints for males and females.
From Table 3, all twenty joints are statistically significant at 1% level of significance, because p-value of all joints is zero and less than 1%—this suggests that there is clearly difference between males and females at each joint during a casual walk. With the help of the t-test, it is concluded that all joints of males and females play different roles during gait.

3.3.3. Comparison of Gender Joints Based on ANOVA Technique

t-test is used for the comparison of two joints simultaneously. To compare more than two joints at the same time, the authors used analysis of variance (ANOVA) technique. Table 4 shows the result of ANOVA for particular joints during gait of both males and females—Twenty joints of males compared with twenty joints of females; k = 40 and d.f = 39. Further, = 1440 and k = 40, therefore, an error d.f = n − k = 1400, sum of square (SS) and mean square (MS), respectively. The p-value is less than 5% and difference between various joints is significant, i.e., there is significant difference between males and females at joint hip center. It suggests that, at joint hip center, the female contribution is higher than males, which further concludes that the contribution of the hip center during walking is different for both males and females, as shown in Table 3. The t-test and ANOVA technique provides us the same conclusion. The constant term k denotes the total numbers of joints, i.e., male = 20 and female = 20.

3.4. Feature Selection

We first calculate the adjusted mean and adjusted standard deviation to determine the Cronbach’s alpha value. By evaluating a feature’s consistency, or how close together a set of objects is, we can use Cronbach’s alpha to assess it.
The Cronbach’s alpha yields an average result of 99.74% as shown in Table 5. It indicates that the data can be analyzed very effectively.

3.5. Classification Using Machine Learning

Machine learning classification can often capture some key information to represent subjects’ identities, which other classifiers cannot capture, by iteratively feeding gait features and correlating labels into machine learning models in the training phase. Regression models are a well-known machine learning classifications for gait biometrics. The relationship between the categorical target variable and one or more independent factors is measured using logistic regression. When there are just two alternative outcomes for a goal variable, it can be helpful (in other words, it is binary). In order to predict the target variable classes, binary logistic regression classification uses one or more predictor variables that may be continuous or categorical. With the aid of this technique, it is possible to pinpoint crucial variables (Xi) that have an impact on the dependent variable (Y) as well as the nature of their interactions. Binary logistic regression model can be represented mathematically as
P 1 = e ( a + b X ) 1 + e ( a + b X )
P 0 = 1 1 + e ( a + b X )
The main advantage of the proposed system is to identify the gender (male/female) of the walking subject on the basis of gait 3D joints positions detected via MS Kinect controller. The proposed system calculates the values of P 1 and P 0 using Equations (22) and (23). When the value of P 1 exceeds the threshold value, the walking person will be a male, and when the value of P 0 exceeds the threshold value, the walking person will be a female. To compare the results of the fitted binary logistic regression model, we used the threshold value of the binary logistic regression model for gender classification.

4. Experimental Result

To evaluate our proposed method, we used a dataset of 373 different individuals wearing their typical shoes and clothes with different ages and gender. Our setup placed the Kinect camera on a tripod at a height of 1.0 m in a typical room measuring 6 m × 8 m with normal daylight. On a PC with a 64 GB RAM and an Intel Core i3 312 GHZ processor, the recorded data were processed. The training computations take the walking of subjects for only a few seconds to complete. The MS Kinect tracks the subjects’ 3D joints positions at a rate of 30 frame per seconds (fps). We recorded people walking toward and away from the camera in front of the sensor in order to capture lengthy sequences. As a result, each walking sequence had between 30 and 150 frames. Different step sizes and speeds among different participants caused variations in sequence duration. As for 3D joint detection, the visibility of the full body joints is mandatory in front of the MS Kinect; therefore, the maximum distance between the MS Kinect and walking subject is defined as 6 feet (see Figure 4).
The main factor to optimize the distance to be 6 feet was the “height of the person” and “resolution (640 × 480) of the camera”. Because when a tall person walks near to the camera, due to their height, some of the body joints are occluded, which affects the accuracy due to the detection of few joints. In contrast, a person walking beyond the working area of the MS Kinect can result in the false detection of 3D joints. The detection of the walking person has no effect on the distance between subject and MS Kinect, but just for the detection of full body joints. For a better depiction of the data, the data also include additional details, such as participants’ age and gender. Figure 5 and Figure 6 show screenshots of the gender detected in real time from their gait in different direction. For experimental analysis, 100 volunteer subjects (50 male and 50 female) participated in the experimental study.

4.1. Classification Accuracy

We have evaluated several parameters and features of our method in order to fully assess it. We assess the efficiency of joint combinations. The accuracy’s detected by our system, using different set of joints such as 9, 12, 14, 16, 18, and 20 were 93.0%, 95.0%, 90.0%, 96.0%, 97.0%, and 98.0%, respectively. The best accuracy was 98.0% when all 20 joints in our suggested system were used. We consider several combinations of walking sequences for training and testing in order to choose the optimal model for gender classification. For each set, we maintained a minimum difference of two to three joints. By contrasting all of the male and female joints, we find a natural overall difference. The complete set of joints greatly enhanced the output; however, the variance is minimal and the rate of increase is moderate, demonstrating the robustness of our classifier to the training data. To measure the accuracy of classification, we utilize an accuracy definition as
a c c u r a c y = T P + T N T P + F P + T N + F N
Our classification accuracy for each of the classes is summarized in Table 6. The proposed technique achieves the same accuracy for males as well as females, but comparatively it achieves better accuracy than other systems.
The covariates are the combined averages of all human joints. The walking person’s gender (y = 1, 0) is the dependent variable. Additionally, we evaluate that all of the fitted binary logistic regression model’s coefficients are equal to zero using the G-statistic of 71.463, the DF of 20, and the p-value of 0.000. Further, we determine the statistical significance of each coefficient parameter at the 5% level. The predicted values of the best-fit binary logistic regression model are categorized in confusion matrix (see Figure 7). The model successfully predicted 49 out of 50 males, and 49 females out of 50 females. Only one man and one woman had incorrect predictions. By comparing threshold values (0.5), the model predicts gender status 98.0% accurately.
Accuracy will decrease with larger standard errors, and will increase with lower standard errors. These joints provide better gender detection because the its standard errors are minimum. Moreover, the p-value for each joints is less than 0.05, which indicates that the joints are statistically significant in prediction of a gender as shown in Table 7. The standard error of joints can be computed using formula
S . E ( X ¯ j ) = σ j ( x ) n

4.1.1. Comparison of the Proposed System with Different Conventional Algorithms

We tested the result of the proposed system using a number of conventional algorithms, such as naive Bayes, multi-layer perception (MLP), random forest (RF), and k-nearest neighbor (KNN). The comparison results are shown in Table 8. Performance evaluation parameters are used to generate the experimental results of gender classification as well as the parameters for performance evaluation. Classification accuracy and true positive rate (recall) for male and female individuals have been considered as evaluation criteria. In the conventional algorithms, the best performance is shown by the KNN algorithm, with a classification accuracy of 92.0%. In comparison to the conventional algorithms, the classification accuracy of the proposed system is 98.0%, which is better than all the existing systems. In the conventional methods, most of the systems showed unbalanced recall for males and females, which indicates that the data are more difficult to classify. On the other hand, the proposed system is better because it shows a balanced recall for both males and females.
To the best of the author’s knowledge, very limited datasets are publicly available for gait-based gender recognition captured by Microsoft Kinect. After an extensive literature review, we found one public dataset (UPCVgait) of gait-based data using Microsoft Kinect for gender recognition. We applied the proposed model on the above dataset for gender classification and compared it with the other conventional methods [55]. As shown in Table 9, the existing system produced an accuracy of 96.67% [55]. The proposed system outperformed the existing system with an accuracy of 98.5% by using the same dataset (see Table 8).

4.1.2. Discussion

The statistical findings of this research reveal that coefficient of variations for both male and female joints during gait shows significant differences. The coefficient of variation at joint head for males is 10.31 (see Table 1), which shows less variability as compared to others joints of male. Therefore, the joint head of males is more consistent during walking. On the contrary, the coefficient of variation at the joint head of female is 9.05, which also shows less variability compared to the remaining joints of females. So, the joint head is also more consistent for females during casual walk. Similarly, the coefficient of variation at joint foot-left of males is 17.58, which shows high variability compared to the remaining joints of males. Therefore, the male performance at joint foot-left varies from person to person. For foot-left of females, the coefficient of variation is 17.01, which shows less variability as compared to joint foot left of males. Therefore, the female performance at joint foot-left is consistent compared to males. The coefficient of variations at joint foot-right for males and females are 16.34 and 17.82, respectively, which shows that male performance is more consistent than female at joint foot-right.
To compare all joints using t-test, there is a significant difference between male and female joints. p-values for all joints are zero (see Table 3) which indicates that male and female joints play different roles during gait.
According to analysis of correlation, both males and females widely move in the X and Z coordinates, as the correlation is high and highly significant at X and Z coordinates, i.e., for both males and females at X and Z-coordinates because the p-values for each joints is zero. On the contrary, the correlation is weak and less than 50% for most joints of males and females at X and Y-coordinates as well as Y and Z-coordinates. The correlation of joints elbow-left and hip-left for male and head of female are more than 90%, which means that these joints are highly significant (see Table 2). According to analysis of correlation there is quit difference between male and female joints. Outcomes of t-test show that all joints of male and female are significantly different from each other. ANOVA technique was applied and provided the same conclusion for reduction in type-I error and for comparison of all joints of male and female. It showed that all joints of males and females are statistically significant at 1% and 5% level of significance.
Analysis and results of t-test, ANOVA, and correlation suggest that male and female joints have different contributions during walking. Further more, the Cronbach’s alpha yields a result of 99.75%, which suggests the reliability of data. Finally, the gender of the walking subjects are detected with an accuracy of 98.0% using a machine learning approach.

5. Conclusions and Future Work

In this paper, a machine learning approach for gait-based gender classification using all body joints is proposed. The proposed method included several stages, such as gait feature extraction based on three dimensional (3D) features, feature selection, and human gender classification. The proposed system extracted the 3D joints positions of all human joints using Kinect sensor, which were compared for the identification of gender during walk. Different statistical tools such as Cronbach’s alpha, correlation, t-test, and ANOVA techniques were exploited to select significant joints. The Cronbach’s alpha technique yielded an average result of 99.74%, which indicates the reliability of joints. Similarly, the correlation results indicated that there is significant difference between male and female joints during gait. t-test and ANOVA approaches demonstrated that all twenty joints are statistically significant for gender classification, because the p-value for each joint was zero and less than 1%. It was observed that all male joints contributed differently than female joints during gait. The results showed that the head is consistent for both male and female during walk. Furthermore, the results proved that the joints foot-left and foot-right of males and females vary more from individual to individual. The joints hand-left, knee-right, hip-center, and spine of all males played the same role during gait. Further, the joint spine, shoulder-center, head, and shoulder-right played the same contribution for all female during gait. According to the coefficient of variation of male and female joints, male joints are more consistent than female joints during walking, which showed that there is a large variance in female joints as compared to male joints. There was high variation in males at joint knee left during walking, which suggested that contribution of joint knee-left of males varies from person to person. There was high variation in joints foot-right, ankle-right, and foot-left of females, which indicates that female contribution varies from female to female. Finally, the gender of the walking person was correctly identified with an accuracy of 98.0% using all body joints. To demonstrate the improved performance of the proposed system, experimental results are compared with deep learning as well as with conventional machine learning techniques. The comparative studies demonstrated the robustness of the proposed machine learning approach, which outperforms the state-of-the-art models. The majority of research on gait-based gender recognition that have been published in the literature have relied on very small sample sizes, which may have an impact on gender identification.
In future work, the researchers would strive further to strengthen the results by studding a much larger database. The proposed method will investigate which joints of males and females are more significant in running, jumping, and sports. Furthermore, these results can be used for modeling human identification.

Author Contributions

Writing—original draft, M.A.; Writing—review & editing, K.U., I.S. and J.C.; Supervision, S.U. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. NRF-2020R1A2C1013308).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bouchrika, I.; Goffredo, M.; Carter, J.; Nixon, M. On using gait in forensic biometrics. J. Forensic Sci. 2011, 56, 882–889. [Google Scholar] [CrossRef]
  2. Iwama, H.; Muramatsu, D.; Makihara, Y.; Yagi, Y. Gait verification system for criminal investigation. Inf. Media Technol. 2013, 8, 1187–1199. [Google Scholar] [CrossRef] [Green Version]
  3. Upadhyay, J.; Gonsalves, T. Robust and Lightweight System for Gait-Based Gender Classification toward Viewing Angle Variations. AI 2022, 3, 538–553. [Google Scholar] [CrossRef]
  4. Upadhyay, J.L.; Gonsalves, T.; Katkar, V. A Lightweight System Towards Viewing Angle and Clothing Variation in Gait Recognition. Int. J. Big Data Intell. Appl. (IJBDIA) 2021, 2, 1–18. [Google Scholar] [CrossRef]
  5. Saleem, F.; Khan, M.A.; Alhaisoni, M.; Tariq, U.; Armghan, A.; Alenezi, F.; Choi, J.I.; Kadry, S. Human gait recognition: A single stream optimal deep learning features fusion. Sensors 2021, 21, 7584. [Google Scholar] [CrossRef]
  6. Han, J.; Bhanu, B. Individual recognition using gait energy image. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 28, 316–322. [Google Scholar] [CrossRef] [PubMed]
  7. Chtourou, I.; Fendri, E.; Hammami, M. Person re-identification based on gait via Part View Transformation Model under variable covariate conditions. J. Vis. Commun. Image Represent. 2021, 77, 103093. [Google Scholar] [CrossRef]
  8. Wu, Z.; Huang, Y.; Wang, L.; Wang, X.; Tan, T. A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 209–226. [Google Scholar] [CrossRef]
  9. Makihara, Y.; Suzuki, A.; Muramatsu, D.; Li, X.; Yagi, Y. Joint intensity and spatial metric learning for robust gait recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5705–5715. [Google Scholar]
  10. Li, X.; Makihara, Y.; Xu, C.; Yagi, Y.; Ren, M. Joint intensity transformer network for gait recognition robust against clothing and carrying status. IEEE Trans. Inf. Forensics Secur. 2019, 14, 3102–3115. [Google Scholar] [CrossRef]
  11. Takemura, N.; Makihara, Y.; Muramatsu, D.; Echigo, T.; Yagi, Y. On input/output architectures for convolutional neural network-based cross-view gait recognition. IEEE Trans. Circuits Syst. Video Technol. 2017, 29, 2708–2719. [Google Scholar] [CrossRef]
  12. Chao, H.; He, Y.; Zhang, J.; Feng, J. Gaitset: Regarding gait as a set for cross-view gait recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8126–8133. [Google Scholar]
  13. Xu, C.; Makihara, Y.; Li, X.; Yagi, Y.; Lu, J. Cross-view gait recognition using pairwise spatial transformer networks. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 260–274. [Google Scholar] [CrossRef]
  14. Li, X.; Makihara, Y.; Xu, C.; Yagi, Y.; Ren, M. Gait recognition via semi-supervised disentangled representation learning to identity and covariate features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 13309–13319. [Google Scholar]
  15. Xu, C.; Makihara, Y.; Ogi, G.; Li, X.; Yagi, Y.; Lu, J. The OU-ISIR gait database comprising the large population dataset with age and performance evaluation of age estimation. IPSJ Trans. Comput. Vis. Appl. 2017, 9, 1–14. [Google Scholar] [CrossRef]
  16. Preis, J.; Kessel, M.; Werner, M.; Linnhoff-Popien, C. Gait recognition with kinect. In Proceedings of the 1st International Workshop on Kinect in Pervasive Computing, New Castle, UK, 18–22 June 2012; pp. 1–4. [Google Scholar]
  17. Alharbi, A.; Alharbi, F.; Kamioka, E. Skeleton based gait recognition for long and baggy clothes. MATEC Web Conf. 2019, 277, 03005. [Google Scholar] [CrossRef]
  18. Bei, S.; Deng, J.; Zhen, Z.; Shaojing, S. Gender recognition via fused silhouette features based on visual sensors. IEEE Sens. J. 2019, 19, 9496–9503. [Google Scholar] [CrossRef]
  19. Xu, C.; Makihara, Y.; Liao, R.; Niitsuma, H.; Li, X.; Yagi, Y.; Lu, J. Real-time gait-based age estimation and gender classification from a single image. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 202; pp. 3460–3470.
  20. Zhang, S.; Wang, Y.; Li, A. Gait-based age estimation with deep convolutional neural network. In Proceedings of the 2019 International Conference on Biometrics (ICB), Crete, Greece, 4–7 June 2019; pp. 1–8. [Google Scholar]
  21. Zhang, Y.; Huang, Y.; Wang, L.; Yu, S. A comprehensive study on gait biometrics using a joint CNN-based method. Pattern Recognit. 2019, 93, 228–236. [Google Scholar] [CrossRef]
  22. Tang, C.; Lu, J.; Liu, J. Non-contact heart rate monitoring by combining convolutional neural network skin detection and remote photoplethysmography via a low-cost camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1309–1315. [Google Scholar]
  23. Kwon, B.; Lee, S. Joint swing energy for skeleton-based gender classification. IEEE Access 2021, 9, 28334–28348. [Google Scholar] [CrossRef]
  24. Guffanti, D.; Brunete, A.; Hernando, M. Non-invasive multi-camera gait analysis system and its application to gender classification. IEEE Access 2020, 8, 95734–95746. [Google Scholar] [CrossRef]
  25. Lee, M.; Lee, J.H.; Kim, D.H. Gender recognition using optimal gait feature based on recursive feature elimination in normal walking. Expert Syst. Appl. 2022, 189, 116040. [Google Scholar] [CrossRef]
  26. More, S.A.; Deore, P.J. Gait-based human recognition using partial wavelet coherence and phase features. J. King Saud-Univ.-Comput. Inf. Sci. 2020, 32, 375–383. [Google Scholar] [CrossRef]
  27. Barra, P.; Bisogni, C.; Nappi, M.; Freire-Obregón, D.; Castrillón-Santana, M. Gait analysis for gender classification in forensics. In Proceedings of the International Conference on Dependability in Sensor, Cloud, and Big Data Systems and Applications, Guangzhou, China, 12–15 November 2019; Springer: Singapore, 2019; pp. 180–190. [Google Scholar]
  28. Dung Do, T.; Kim, H.; Huan Nguyen, V. Real-time and robust multiple-view gender classification using gait features in video surveillance. arXiv 2019, arXiv:1905.01013. [Google Scholar]
  29. Russel, N.S.; Selvaraj, A. Gender discrimination, age group classification and carried object recognition from gait energy image using fusion of parallel convolutional neural network. IET Image Process. 2021, 15, 239–251. [Google Scholar] [CrossRef]
  30. Sudha, L.R.; Bhavani, R. Gait based gender identification using statistical pattern classifiers. Int. J. Comput. Appl. 2012, 40, 30–35. [Google Scholar] [CrossRef]
  31. Hu, M.; Wang, Y. A new approach for gender classification based on gait analysis. In Proceedings of the 2009 Fifth International Conference on Image and Graphics, Xi’an, China, 20–23 September 2009; pp. 869–874. [Google Scholar]
  32. Hassan, O.M.S.; Abdulazeez, A.M.; TİRYAKİ, V.M. Gait-based human gender classification using lifting 5/3 wavelet and principal component analysis. In Proceedings of the 2018 International Conference on Advanced Science and Engineering (ICOASE), Duhok, Iraq, 9–11 October 2018; pp. 173–178. [Google Scholar]
  33. Zhao, G.; Liu, G.; Li, H.; Pietikainen, M. 3D gait recognition using multiple cameras. In Proceedings of the 7th International Conference on Automatic Face and Gesture Recognition (FGR06), Southampton, UK, 10–12 April 2006; pp. 529–534. [Google Scholar]
  34. Yamauchi, K.; Bhanu, B.; Saito, H. Recognition of walking humans in 3D: Initial results. In Proceedings of the 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Miami Beach, FL, USA, 20–25 June 2009; pp. 45–52. [Google Scholar]
  35. Krzeszowski, T.; Michalczuk, A.; Kwolek, B.; Switonski, A.; Josinski, H. Gait recognition based on marker-less 3D motion capture. In Proceedings of the 2013 10th IEEE International Conference on Advanced Video and Signal Based Surveillance, Krakow, Poland, 27–30 August 2013; pp. 232–237. [Google Scholar]
  36. Kusakunniran, W.; Wu, Q.; Zhang, J.; Li, H. Gait recognition under various viewing angles based on correlated motion regression. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 966–980. [Google Scholar] [CrossRef]
  37. Tsuji, A.; Makihara, Y.; Yagi, Y. Silhouette transformation based on walking speed for gait identification. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 717–722. [Google Scholar]
  38. Hossain, M.A.; Makihara, Y.; Wang, J.; Yagi, Y. Clothing-invariant gait identification using part-based clothing categorization and adaptive weight control. Pattern Recognit. 2010, 43, 2281–2291. [Google Scholar] [CrossRef]
  39. Goffredo, M.; Bouchrika, I.; Carter, J.N.; Nixon, M.S. Self-calibrating view-invariant gait biometrics. IEEE Trans. Syst. Man, Cybern. Part B (Cybern.) 2009, 40, 997–1008. [Google Scholar] [CrossRef] [Green Version]
  40. Kale, A.; Sundaresan, A.; Rajagopalan, A.; Cuntoor, N.P.; Roy-Chowdhury, A.K.; Kruger, V.; Chellappa, R. Identification of humans using gait. IEEE Trans. Image Process. 2004, 13, 1163–1173. [Google Scholar] [CrossRef]
  41. Jean, F.; Bergevin, R.; Albu, A.B. Computing view-normalized body parts trajectories. In Proceedings of the Fourth Canadian Conference on Computer and Robot Vision (CRV’07), Montreal, QC, Canada, 28–30 May 2007; pp. 89–96. [Google Scholar]
  42. Iwama, H.; Okumura, M.; Makihara, Y.; Yagi, Y. The ou-isir gait database comprising the large population dataset and performance evaluation of gait recognition. IEEE Trans. Inf. Forensics Secur. 2012, 7, 1511–1521. [Google Scholar] [CrossRef] [Green Version]
  43. Sivapalan, S.; Chen, D.; Denman, S.; Sridharan, S.; Fookes, C. Gait energy volumes and frontal gait recognition using depth images. In Proceedings of the 2011 International Joint Conference on Biometrics (IJCB), Washington, DC, USA, 11–13 October 2011; pp. 1–6. [Google Scholar]
  44. Araujo, R.M.; Graña, G.; Andersson, V. Towards skeleton biometric identification using the microsoft kinect sensor. In Proceedings of the 28th Annual ACM Symposium on Applied Computing, Coimbra, Portugal, 18–22 March 2013; pp. 21–26. [Google Scholar]
  45. Milovanovic, M.; Minovic, M.; Starcevic, D. Walking in colors: Human gait recognition using kinect and cbir. IEEE MultiMedia 2013, 20, 28–36. [Google Scholar] [CrossRef]
  46. Yang, K.; Dou, Y.; Lv, S.; Zhang, F.; Lv, Q. Relative distance features for gait recognition with Kinect. J. Vis. Commun. Image Represent. 2016, 39, 209–217. [Google Scholar] [CrossRef] [Green Version]
  47. Ahmed, F.; Polash Paul, P.; Gavrilova, M.L. Kinect-based gait recognition using sequences of the most relevant joint relative angles. J. WSCG 2015, 23, 147–156. [Google Scholar]
  48. Kastaniotis, D.; Theodorakopoulos, I.; Theoharatos, C.; Economou, G.; Fotopoulos, S. A framework for gait-based recognition using Kinect. Pattern Recognit. Lett. 2015, 68, 327–335. [Google Scholar] [CrossRef]
  49. Ball, A.; Rye, D.; Ramos, F.; Velonaki, M. Unsupervised clustering of people from’skeleton’data. In Proceedings of the Seventh Annual ACM/IEEE International Conference on Human-Robot Interaction, Boston, MA, USA, 5–8 March 2012; pp. 225–226. [Google Scholar]
  50. Ahmed, M.; Al-Jawad, N.; Sabir, A.T. Gait recognition based on Kinect sensor. In Proceedings of the Real-Time Image and Video Processing 2014. International Society for Optics and Photonics, Brussels, Belgium, 13–17 April 2014; Volume 9139, p. 91390B. [Google Scholar]
  51. Eltaher, M.; Yang, Y.; Lee, J. Gait-Based Gender Classification Using Kinect Sensor. In Proceedings of the 2015 ASEE Annual Conference & Exposition, Seattle, WA, USA, 14–17 June 2015; pp. 26–808. [Google Scholar]
  52. Azhar, M.; Ullah, S.; Raees, M.; Rahman, K.U.; Rehman, I.U. A real-time multi view gait-based automatic gender classification system using kinect sensor. Multimed. Tools Appl. 2022, 1–24. [Google Scholar] [CrossRef]
  53. Sinha, A.; Chakravarty, K.; Bhowmick, B. Person identification using skeleton information from kinect. In Proceedings of the International Conference on Advances in Computer-Human Interactions, Nice, France, 24 February–1 March 2013; pp. 101–108. [Google Scholar]
  54. Ahmed, M.H.; Sabir, A.T. Human gender classification based on gait features using kinect sensor. In Proceedings of the 2017 3rd IEEE International Conference on Cybernetics (Cybconf), Exeter, UK, 21–23 June 2017; pp. 1–5. [Google Scholar]
  55. Kastaniotis, D.; Theodorakopoulos, I.; Economou, G.; Fotopoulos, S. Gait-based gender recognition using pose information for real time applications. In Proceedings of the 2013 18th International Conference on Digital Signal Processing (DSP), Fira, Greece, 1–3 July 2013; pp. 1–6. [Google Scholar]
Figure 1. Detail system model of the proposed system.
Figure 1. Detail system model of the proposed system.
Sensors 22 09113 g001
Figure 2. Walking of subjects in different directions.
Figure 2. Walking of subjects in different directions.
Sensors 22 09113 g002
Figure 3. Various human joints.
Figure 3. Various human joints.
Sensors 22 09113 g003
Figure 4. Experimental setup using MS Kinect.
Figure 4. Experimental setup using MS Kinect.
Sensors 22 09113 g004
Figure 5. Walking of different subjects from different directions.
Figure 5. Walking of different subjects from different directions.
Sensors 22 09113 g005
Figure 6. Walking of more than one subjects at the same time from different directions.
Figure 6. Walking of more than one subjects at the same time from different directions.
Sensors 22 09113 g006
Figure 7. Confusion matrix.
Figure 7. Confusion matrix.
Sensors 22 09113 g007
Table 1. The mean, standard deviation (S.D), and coefficient of variation (C.V) for various joints of male and female.
Table 1. The mean, standard deviation (S.D), and coefficient of variation (C.V) for various joints of male and female.
MaleFemale
JointsMeanS.DC.VMeanS.DC.V
Hip Center0.99180.115411.631.19060.148912.50
SPINE1.02330.115411.631.19060.148912.50
Shoulder Center1.16930.123210.531.22370.143911.76
HEAD1.20230.124010.311.38330.12529.05
Shoulder Left1.13820.127211.181.28830.151511.76
Elbow Left1.05040.126612.061.22200.178514.61
Wrest Left0.96300.126413.121.17240.184115.70
Hand Left0.93430.127813.671.15530.189416.39
Shoulder Right1.11140.120410.831.29810.130110.02
Elbow Right1.00430.119011.851.20840.144811.98
Wrest Right0.90490.118013.041.12450.155813.85
Hand Right0.87700.117613.411.10110.157614.32
HIP Left0.97380.117212.031.16730.152913.10
Knee Left0.81300.141517.401.04580.160215.31
Ankle Left0.74000.122416.530.97870.161516.50
Foot Left0.70540.124017.580.94990.161617.01
HIP Right0.97240.115511.881.17540.148112.60
Knee Right0.84860.116013.671.08040.157614.59
Ankle Right0.76070.117215.401.01200.173317.13
Foot Right0.72170.118016.340.98270.175117.82
Table 2. The Pearson correlation between various joints in 3D coordinates system (x, y, and z).
Table 2. The Pearson correlation between various joints in 3D coordinates system (x, y, and z).
MaleFemale
JointsXandYYandZXandZXandYYandZXandZ
HIP-C0.84 (0.00)0.88 (0.00)0.90 (0.00)0.14 (0.00)0.20 (0.00)0.80 (0.00)
HIP-L0.81 (0.00)0.88 (0.00)0.94 (0.00)0.01 (0.79)0.01 (0.68)0.84 (0.00)
HIP-R0.82 (0.00)0.76 (0.00)0.93 (0.00)0.24 (0.00)0.19 (0.00)0.60 (0.00)
Spine0.86 (0.00)0.84 (0.00)0.90 (0.00)0.22 (0.00)0.16 (0.00)0.85 (0.00)
Shoulder-C0.65 (0.00)0.52 (0.00)0.92 (0.00)0.29 (0.00)0.16 (0.00)0.88 (0.00)
Shoulder-L0.60 (0.00)0.49 (0.00)0.95 (0.00)0.04 (0.27)−0.04 (0.24)0.86 (0.00)
Shoulder-R0.57 (0.00)0.16 (0.00)0.92 (0.00)0.02 (0.61)0.31 (0.00)0.70 (0.00)
Head0.70 (0.00)0.66 (0.00)0.90 (0.00)0.32 (0.00)0.27 (0.00)0.90 (0.00)
Elbow-L0.40 (0.00)0.31 (0.00)0.95 (0.00)0.18 (0.00)0.04 (0.26)0.87 (0.00)
Elbow-R0.39 (0.00)0.54 (0.00)0.93 (0.00)−0.47 (0.00)−0.25 (0.00)0.82 (0.00)
Wrest-L0.16 (0.00)0.25 (0.00)0.94 (0.00)0.16 (0.00)0.52 (0.00)0.79 (0.00)
Wrest-R0.56 (0.00)0.45 (0.00)0.96 (0.00)−0.75 (0.00)−0.22 (0.00)0.77 (0.00)
Hand-L0.08 (0.03)0.15 (0.00)0.93 (0.00)0.08 (0.03)0.49 (0.00)0.72 (0.00)
Hand-R0.49 (0.00)0.32 (0.00)0.91 (0.00)−0.72 (0.00)−0.29 (0.00)0.80 (0.00)
Knee-L0.35 (0.00)0.49 (0.00)0.90 (0.00)−0.15 (0.00)−0.37 (0.00)0.90 (0.00)
Knee-R0.41 (0.00)0.44 (0.00)0.91 (0.00)−0.06 (0.09)−0.13 (0.00)0.79 (0.00)
Ankle-L−0.49 (0.00)−0.27 (0.00)0.94 (0.00)−0.44 (0.00)−0.45 (0.00)0.87 (0.00)
Ankle-R−0.49 (0.00)0.03 (0.49)0.93 (0.00)−0.47 (0.00)−0.25 (0.00)0.82 (0.00)
Foot-L−0.59 (0.00)−0.46 (0.00)0.94 (0.00)−0.54 (0.00)−0.51 (0.00)0.88 (0.00)
Foot-R−0.53 (0.00)0.03 (0.44)0.93 (0.00)−0.39 (0.00)−0.32 (0.00)0.88 (0.00)
Table 3. The results of the two-sample t-test used for comparison between male and female joints during walk.
Table 3. The results of the two-sample t-test used for comparison between male and female joints during walk.
JointsDFPooled S.DT-Valuep-Value
Hip Center700.12226.610.000
SPINE700.1214−6.710.000
Shoulder Center700.1222−5.830.000
HEAD700.1225−6.120.000
Shoulder Left700.13204.810.000
Elbow Left700.1391−5.100.000
Wrest Left700.14106.030.000
Hand Left700.1437−6.210.000
Shoulder Right700.12106.400.000
Elbow Right700.1238−6.810.000
Wrest Right700.1261−7.120.000
Hand Right700.1262−7.210.000
HIP Left700.1246−6.300.000
Knee Left700.14326.630.000
Ankle Left700.1304−7.380.000
Foot Left700.13157.500.000
HIP Right700.1220−6.750.000
Knee Right700.1252−7.530.000
Ankle Right700.1307−7.820.000
Foot Right700.1317−8.030.000
Table 4. The result of analysis of variance (ANOVA).
Table 4. The result of analysis of variance (ANOVA).
Source of VariationDFSSMSF-Valuep-Value
Joints3942.11051.079864.660.000
Error140023.38020.0167
Total143965.4906
Table 5. The results of adjusted mean (Adj.mean), adjusted standard deviation (Adj.SD), and Cronbach’s alpha for reliability of human body joints.
Table 5. The results of adjusted mean (Adj.mean), adjusted standard deviation (Adj.SD), and Cronbach’s alpha for reliability of human body joints.
JointsAdj.MeanAdj.SDCronbach’s Value
X ¯ C ( H i p C ) 18.9412.9860.9973
X ¯ C ( S p i n e ) 18.9092.9860.9973
X ¯ C ( S h C ) 18.8013.0210.9979
X ¯ C ( H e a d ) 18.7352.9940.9974
X ¯ C ( S h L ) 18.8072.9930.9974
X ¯ C ( E L ) 18.8892.9820.9974
X ¯ C ( W L ) 18.9672.9720.9974
X ¯ C ( H L ) 18.9922.9680.9974
X ¯ C ( S h R ) 18.8242.9920.9973
X ¯ C ( E R ) 18.9272.9840.9973
X ¯ C ( W R ) 19.0222.9790.9973
X ¯ C ( H R ) 19.0492.9770.9973
X ¯ C ( H i p L ) 18.9602.9850.9973
X ¯ C ( K L ) 19.1112.9680.9976
X ¯ C ( A L ) 19.1822.9710.9973
X ¯ C ( F L ) 19.2152.9690.9973
X ¯ C ( H i p R ) 18.9592.9850.9972
X ¯ C ( K R ) 19.0742.9760.9973
X ¯ C ( A R ) 19.1582.9670.9974
X ¯ C ( F R ) 19.1942.9640.9974
Table 6. Classification accuracy of the proposed system with different set of joints.
Table 6. Classification accuracy of the proposed system with different set of joints.
Gender09 Joints12 Joints14 Joints16 Joints18 Joints20 Joints
Male92.094.090.096.098.098.0
Female94.096.090.096.096.098.0
Both93.095.090.096.097.098.0
Table 7. Standard error, coefficient, and p-Value of binary logistic regression model.
Table 7. Standard error, coefficient, and p-Value of binary logistic regression model.
Predictors (Joints)CoefficientsStandard Errorp-Value
Constant18.3023.65110.013
X ¯ C ( H i p C ) 20.242.17860.043
X ¯ C ( S p i n e ) 55.536.14210.021
X ¯ C ( S h C ) 45.347.00120.032
X ¯ C ( H e a d ) 10.7555.56090.040
X ¯ C ( S h L ) −12.030.98000.026
X ¯ C ( E L ) 22.719.34510.029
X ¯ C ( W L ) 33.329.65550.032
X ¯ C ( H L ) 28.942.77640.035
X ¯ C ( S h R ) 36.023.22130.026
X ¯ C ( E R ) 29.149.32440.029
X ¯ C ( W R ) 50.432.92120.034
X ¯ C ( H R ) 44.640.66650.038
X ¯ C ( H i p L ) −11.560.65420.020
X ¯ C ( K L ) 20.855.87210.033
X ¯ C ( A L ) 26.944.81110.036
X ¯ C ( F L ) 45.234.54440.038
X ¯ C ( H i p R ) 52.333.21110.040
X ¯ C ( K R ) 37.835.54330.041
X ¯ C ( A R ) 39.145.44560.046
X ¯ C ( F R ) 10.622.54670.045
Table 8. Comparison of the proposed system with the conventional systems based on classification accuracy.
Table 8. Comparison of the proposed system with the conventional systems based on classification accuracy.
AlgorithmRecall Male (%)Recall Female (%)Accuracy (%)
Naive Bayes80.070.075.0%
MLP84.070.077.0%
RF92.088.090.0%
KNN94.090.092.0%
Our98.098.098.0%
Table 9. Comparison of classification accuracy between the proposed system and the conventional system.
Table 9. Comparison of classification accuracy between the proposed system and the conventional system.
ClassifierDatasetAccuracy (%)
SVM [55]UPCVgait96.67%
Proposed MethodUPCVgait98.5%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Azhar, M.; Ullah, S.; Ullah, K.; Syed, I.; Choi, J. A Gait-Based Real-Time Gender Classification System Using Whole Body Joints. Sensors 2022, 22, 9113. https://doi.org/10.3390/s22239113

AMA Style

Azhar M, Ullah S, Ullah K, Syed I, Choi J. A Gait-Based Real-Time Gender Classification System Using Whole Body Joints. Sensors. 2022; 22(23):9113. https://doi.org/10.3390/s22239113

Chicago/Turabian Style

Azhar, Muhammad, Sehat Ullah, Khalil Ullah, Ikram Syed, and Jaehyuk Choi. 2022. "A Gait-Based Real-Time Gender Classification System Using Whole Body Joints" Sensors 22, no. 23: 9113. https://doi.org/10.3390/s22239113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop