Next Article in Journal
Seismic Stability for 3D Two-Step Slope Governed by Non-Linearity in Soils Using Modified Pseudo-Dynamic Approach
Previous Article in Journal
Scale-Adaptive Deep Matching Network for Constrained Image Splicing Detection and Localization
Previous Article in Special Issue
Extrinsic Behavior Prediction of Pedestrians via Maximum Entropy Markov Model and Graph-Based Features Mining
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier

1
Department of Computer Science, Air University, Islamabad 44100, Pakistan
2
Information Systems Department, College of Computer and Information Science, Princess Nourah Bint Abdulrahman University, Riyadh 11564, Saudi Arabia
3
Department of Computer Science, College of Computers and Information Technology, Taif University, Taif 21944, Saudi Arabia
4
Department of Computer Science, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
5
Department of Computer Engineering, Tech University of Korea, 237 Sangidaehak-ro, Siheung-si 15073, Gyeonggi-do, Korea
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2022, 12(13), 6481; https://doi.org/10.3390/app12136481
Submission received: 25 May 2022 / Revised: 17 June 2022 / Accepted: 24 June 2022 / Published: 26 June 2022
(This article belongs to the Special Issue Advanced Intelligent Imaging Technology Ⅲ)

Abstract

:

Featured Application

The proposed system is an image processing module that monitors, tracks, and recognizes hand gestures and has been evaluated over publicly available benchmark datasets. However, this technique can be used over automated home appliances as well as security systems to control surrounding environments and classify their events.

Abstract

In the past few years, home appliances have been influenced by the latest technologies and changes in consumer trends. One of the most desired gadgets of this time is a universal remote control for gestures. Hand gestures are the best way to control home appliances. This paper presents a novel method of recognizing hand gestures for smart home appliances using imaging sensors. The proposed model is divided into six steps. First, preprocessing is done to de-noise the video frames and resize each frame to a specific dimension. Second, the hand is detected using a single shot detector-based convolution neural network (SSD-CNN) model. Third, landmarks are localized on the hand using the skeleton method. Fourth, features are extracted based on point-based trajectories, frame differencing, orientation histograms, and 3D point clouds. Fifth, features are optimized using fuzzy logic, and last, the H-Hash classifier is used for the classification of hand gestures. The system is tested on two benchmark datasets, namely, the IPN hand dataset and Jester dataset. The recognition accuracy on the IPN hand dataset is 88.46% and on Jester datasets is 87.69%. Users can control their smart home appliances, such as television, radio, air conditioner, and vacuum cleaner, using the proposed system.

1. Introduction

Over the past few years, intelligent human-computer interaction recognition in the smart home environment is getting more attention in many fields, including architecture, robotics, biomedical, and smart home appliances [1,2,3]. The easiest way of controlling smart home appliances is through hand gestures [4,5,6]. The appliances that you use every day are an important part of your home. Consumers today are very careful about their comfort and safety and are more interesting in smart appliances [7,8,9]. They will facilitate their daily life by controlling the lights, speakers, air conditioners, and similar robots through hand gestures [10,11,12]. By making gestures, you can control all the devices in your home. A motion sensor is one of the most important tools in your smart home [13,14,15].
Different models proposed for controlling smart home appliances using hand gesture recognition can be divided into two main streams. The first approach is based on the recognition of hand gestures using motion sensors embedded in smart home appliances [16,17,18]. In motion-based sensors, one inertial sensor or an array of sensors is used. These sensors are responsible for tracking the acceleration, velocity, and position of the hand. Such type of motion features help to control smart appliances such as television, radio, and lighting of rooms [19,20,21], however, the drawback of using motion-based sensors in smart home appliances is the high sensitivity. The second approach is the use of image sensors [22,23,24] or cameras to obtain the commands from hand gestures [25]; the sensors are trained on image features which include color, shape, texture, position contours, and motion of the hands. Our proposed model is based on the second approach and recognizes the hand gestures using imagery sensors or cameras [26].
In this research article, we propose a robust method for recognizing hand gestures for controlling smart home appliances. For this, we use the IPN hand dataset and Jester dataset. Initially, preprocessing of the video samples for frame conversion, motion blur noise reduction, and resizing is performed. The next step is hand detection via SSD-CNN. After that, the hand skeleton is extracted to process these data sources by various algorithms for features extraction, optimization of the extracted features, and recognition of the hand gestures. The main contributions of this paper are:
  • For hand motion and position analysis, we propose a method for extracting hand skeletons;
  • For the recognition of image-based hand gestures, we have extracted novel features based on point-based trajectories, frame differencing, orientation histogram, and 3D point clouds.
The article is subdivided as follows: we start with the related works section, which is followed by our system methodology. Then, the detailed experimental setup is discussed, and finally, an overview of the paper is presented in the Conclusions Section.

2. Related Work

Hand gesture recognition can help computers translate and interpret specific motions to control smart home appliances. With the advancement of technology, various hand gesture recognition systems have been developed via smart tools and various classification approaches. In this section, we will discuss the detailed description of various HGR models developed in the past few years. Table 1 includes a comprehensive review of recent research in this area.

3. Material and Methods

In this section, a detailed description of the proposed model is given. First, the video input is converted into RGB frames, then the frames are resized to a fixed dimension. Noise is removed and the quality of images is enhanced and sharpened. The second step is hand detection by first removing the background and extracting the foreground. A hand skeleton is extracted for localizing the points on the entire hand. Then the point-based and texture-based features are extracted. These features are optimized by using an optimization algorithm. At last, a classification algorithm is used for the classification of hand gestures for controlling smart home appliances. Figure 1 illustrates the proposed HGR system model structural design.

3.1. Preprocessing of the Input Videos

Before the localization of the hand points, some preprocessing techniques are applied to save computational cost and time. Initially, video data is converted into frames (images). These frames are set to a fixed dimension of 452 × 357. After that, the frames are denoised using the median filtering algorithm. The median filtering is used for detecting the distorted pixels in the images and replacing the corrupted pixels values with the median values. A 5 × 5 window is used to de-noise the image [31,32,33,34,35,36,37]. The median filter is defined in Equations (1)–(3);
M e d   ( I ) = M e d { I p )
= I p ( k + 1 ) 2 ; k   i s   o d d
= 1 2 [ I p ( k 2 ) + I p ( k 2 ) + 1 ] ,  
where I1, I2, I3,…, Ik is the sequence of neighboring pixels. Before applying the filter, all pixels of the images should be arranged in ascending or descending order. After sorting the pixels, the sequence of the pixels will be I p 1 < I p 2 < I p 3 < I p k , where k is usually odd. Figure 2 shows the results of preprocessing on video frames.

3.2. Hand Detection Using Single Shot MultiBox Detector

The ridge detection of the human silhouette comprises two steps that are binary edge extraction and ridge data generation [38,39,40]. In the binary edge extraction step, the binary edges are extracted from the RGB and depth silhouettes are obtained in the described preprocessing stage. The distance maps are produced on the edges by using distance transform (See Figure 3). While, in the ridge data generation step, the local maximal is obtained from the pre-computed maps, which produces ridge data within the binary edges [41,42]. A further description of binary edge detection and generation of ridge data is described below in Algorithm 1:
Algorithm 1: Hand detection using a single shot multibox detector.
Input:        Optimized feature vectors.
Output:     Hand gesture classification.
Step 1:       Check the length of the hash table (say n).
          Check the number of entries in the hash table by setting a fixed threshold
          (say T = 40).
If (n > T) then
Find the correlation matching or minimum distance between the vectors by the following equation:
γ K = x = 1 n | | α x | | β | |   w h e r e   K = 1 , 2 , 3 , 4
where α represents the centroid of the vectors stored in the hash table, β represents the new vectors of the test image, γ represents the distance between the stored values of the hash table and the new vectors.
 Now, finding the sum of the vectors
θ = γ 1 + γ 2 + γ 3 + γ 4
End
Step 2: /*Check the correlation γ K of the new entry*/
   If ( γ K 0.98 )
   Match exists
   Else
   Match does not exist
End

3.3. Hand Landmarks Localization Using Skeleton Method

To localize the hand landmarks points, the first step is to localize the palm region. For doing this, we selected the palm area via a single shot multibox detector and removed the fingers; with this help, a bounding box appears on the palm. Then the extreme top-left, top-right, bottom-left and bottom-right points are calculated and marked with the 4 points [43,44,45,46].
The next step is to localize the finger points. For this, the palm region is removed and only the fingers are left. The extreme points are identified with the help of a scanning window that moves from top to bottom, identifying all the extreme points. As a result, the extreme top points of all fingers are marked with the 5 points. Similarly, the extreme bottom points are identified using the scanning window that moves from bottom to top marking 5 points on the bottom of the fingers [47]. Figure 4 shows the results of hand point localization.

3.4. Features Extraction

For features extraction, we have extracted both points-based and appearance-based features for better classification of the hand gestures. For point-based features, we have used Bezier curves and frame differencing. For appearance-based features, 3D point clouds are mapped on the hands.

3.4.1. Bezier Curves

The landmark points localized on the entire hand are utilized for Bezier curves fitting for analyzing the trajectories of the hand in different gestures. For this, we have taken three control points to represent a curve using Equation (4);
C u r v e ( x ) = i = 0 n Q i , n ( x ) Y i
The points along the curve are determined by x, where 0 x 1 . The degree of the curve is denoted by n, which is one less than the control points. Yi is the i-th control point where Y(0) = Curve(0) and Y(n) = Curve(1). Q i , n is the Bernstein polynomial and is calculated in Equation (5).
Q i , n = n ! i ! ( n i ) ! n i ( 1 x ) n i
We have used Bezier curves with three control points describing the quadratic curve. Therefore, the Bernstein polynomial with n = 2 is calculated as in Equations (6)–(8) [48,49,50,51].
Q 0 , 2 = ( 1 x ) 2
Q 1 , 2 = 2 x ( 1 x )
Q 2 , 2 = x 2
Therefore, the equation of a Bezier curve with three control points is simplified as in Equation (9). Figure 5 shows the results of the Bezier curves fitting on the hand.
C u r v e ( x ) = ( 1 x ) 2 + 2 x ( 1 x ) + x 2

3.4.2. Frame Differencing

Keyframes are the representation of the elements in the image sequences. In this model, the keyframes have been extracted that exhibit the dynamic hand gestures. Each hand shows a gesture that is localized by a set of points. To find the difference in the positions of the landmarks, the first and the pause frame sequences are taken as the keyframes. Figure 6 illustrates the change in the position of the pixel between different frames. The first frame is well established. To find the motion of the hand gestures, we have adopted the following method of frame difference as defined in Equation (10) [52,53,54,55,56].
  D i f f k ( x , y ) = | F r a m e k ( x , y ) F r a m e k 1 ( x , y ) |
where F r a m e k ( x , y ) and F r a m e k 1 ( x , y ) are the two consecutive frames in which the hand is not moving and D i f f k is defined as in Equation (11).
D i f f k ( x , y ) = | F r a m e k ( x , y ) F r a m e k 1 ( x , y ) | 0
where F r a m e k ( x , y ) is the pause keyframe. The two continuous frames are impossible to be the keyframes. Thus, if F r a m e k 1 ( x , y ) is the keyframe, then F r a m e k ( x , y ) is not the keyframe. Therefore, for each frame sequence, F r a m e k ( x , y ) with N number of frames: The following approach should be outlined as:
  • Initialize the keyframe number with n = 1. So, the keyframes are marked as M 1 ( x ) = 1 ,   M k ( x ) = 0 ,   k = ( 2 , ,   N ) . After that, compute the difference frame D i f f k between F r a m e k ( x , y ) and F r a m e k 1 ( x , y ) , k = ( 2 , ,   N ) . Compute the valid pixels of N.
  • If N > T h r e s h 1 and M k 1 ( x ) = 0 , set n = n + 1, set M k ( x ) = 1 .
  • Set k = k + 1 if the value of k is less than N. Then repeat the steps, otherwise end the procedure.
  • After calculating the frame difference, each key point L in the first keyframe and L points in the other keyframes positions are calculated using the distance formula defined in Equation (12).
    D i s t a n c e = ( L 1 L 1 ) 2 + ( L 2 L 2 )

D Point Clouds

For appearance-based features, extracted the 3D point clouds. The following are the main steps of extracting the appearance-based features [57,58,59].
  • First, a central point on the palm is taken and the maximum distance d between the central point and the edges Ei point of the gesture region is calculated. After that, ten different lengths of the radius are defined as X = n × d 10   w h e r e   n = 1 , 2 , 3 , 10 . Next, the center of the rhombus is defined as C and the radiuses as rn. We have drawn 10 rhombuses (innermost is the first rhombus and outmost is the tenth rhombus) as shown in Figure 7. To highlight the effect of changing hand gestures, the color of the rhombus changes on the hand.
  • In Figure 7, it is visible that every rhombus has a different number of intersections with hand gesture regions. For finding the number of stretched fingers S, we have taken the sixth rhombus (according to the thumb rule). In the sixth rhombus, we extracted those points whose colors vary from green to yellow and yellow to green. We define Gi as the point whose color changes from green to yellow and Yi as the point whose values changes from yellow to green.
  • Now, for midpoint identification, we define it as Mi which is the midpoint of Gi and Yi. Then, each midpoint Mi and the central point C can be connected through a line and the angles between the adjacent lines are calculated. The angles are represented as Anj (j = 1, 2, 3…I − 1).
  • Using the thumb rule, the fifth rhombus is taken as a boundary line to divide the hand gesture into two parts. For instance, we have taken the first part as P1 and the second part as P2, where P1 lies inside the rhombus and P2 is the outside area of the rhombus. Then the ratio R of P1 and P2 is calculated. The R is the gesture region area distribution feature as shown in Algorithm 2. Figure 7 shows the appearance features using 3D point clouds.
Algorithm 2: Feature extraction.
Input: Hands Point based and texture-based data (x, y, z).
Output: Feature Vectors (v1, v2, …, vn).
    featureVectors [   ]
    window_size Get   windowSize (   )
     Overlap Get   Overlapping   Time (   )
For HandComponent in [x,y,z] do
    Hand window Getwindow ( HandComponent )
  /* Extracting features */
  BezierCurves ExtractBezierCurvesFeatures (Hand_window)
  Frame Differencing ExtractFrameDifferencingFeatures (Hand_window)
  3D Point Clouds Extract3DPointCloudsFeatures (Hand_window)
  featureVectors GetFeatureVectors [BezierCurves, Frame Differencing, 3D Point Clouds]
  featureVectors.append (featureVectors)
  End for
featureVectors Normalize (featureVectors)
  return featureVectors

3.5. Feature FOptimization Using Fuzzy Logic

The objective of using fuzzy logic optimizer is to recognize the hand gestures based on the information obtained through different feature descriptors. Each feature descriptor value is labeled with a specific variable and is mapped to their respective fuzzy sets. For instance, we have five fingers labeled as; F1 (thumb), F2 (index finger), F3 (middle finger), F4 (ring finger), and F5 (little finger). The joints of the fingers are labeled as J1, J2, J3, and J4. Similarly, the distance between the fingers is denoted by Di,j showing the distance between the fingers Fi and Fj.
Since any movement of the hand shows the variation of the position of the hand in a sequence of images (frames), to simulate the data transfer, a hand configuration is generated by a tuple of angles. For each tuple of angles, the data is represented using a set of linguistic variables such as curve, straight, and bent. The separation of the fingers is represented as open, closed, crossed, and semi-open. By these notations, the set of features is optimized, helping to reduce the overall computational time and complexity [60,61,62,63]. Figure 8 shows the result of the fuzzy logic optimization.

3.6. Hand Gestures Recognition

For hand gesture recognition, K-Ary Tree Hashing (KATH) classifier is used for the first time in our proposed model [64,65,66,67]. The KATH classifier takes the feature descriptors values of each image corresponding to the hand gesture and projects in a common space without the subtree pattern prior knowledge. Then similar pattern feature descriptors are kept in the traversal table. The unique patterns are specified by passing through recursive indexing N numbers to generate (n − 1). After that, the hand gesture is classified by the sub-patterns created by MinHash. The experimental results show that KATH classified different hand gestures more accurately than many other state-of-the-art methods, i.e., ANN and decision tree, as shown in Figure 9. In our proposed model, the graph g = (v, ɛ, l) is given input with the number of iterations I and F representing the feature space. To assign a new label l, the nodes v are relabeled considering the neighboring nodes WV. The traversal table T is generated and stored. After the traversal table, MinHash classifies data. For dimensional reduction, PCA is used to plot results in 3D feature space.

4. Experimental Setting and Results

4.1. Datasets Descriptions

The IPN hand dataset [68] is a large-scale hand gesture video dataset. It contains 13 gestures, including pointing with one finger, pointing with two fingers, click with one finger, click with two fingers, throw up, throw down, throw left, throw right, open twice, double click with one finger, click with two fingers, zoom in and zoom out. The IPN dataset contains RGB videos with a resolution of 640 × 480 at 30 fps. Figure 10 shows the example images of the IPN hand dataset.
The Jester dataset [69] contains a large collection of labeled hand gestures video clips collected by webcam. The dataset contains 148,092 videos, and each video frame is converted into a jpg image at the rate of 12 frames per second. There are 27 classes of hand gestures named: swiping down, swiping left, swiping right, swiping up, thumb down, thumb up, zooming in with full hand, zooming out with full hand, stop, and so on. Figure 11 shows the example images of the Jester dataset.

4.2. Performance Parameters and Evaluations

4.2.1. Experiment I: The Hand Detection Accuracies

In this experiment, the hand detection accuracies on different hand gestures over IPN Hand dataset and Jester dataset are shown in Table 2 and Table 3, respectively. Table 2 represents the results of the hand gestures of the IPN Hand dataset on both plain and complex backgrounds. We took 30 samples of each hand gesture in the plain and complex background and obtained 97.1% accuracy on plain background samples and 94.3% accuracy on complex background samples.
Table 3 represents the results of the 13 hand gestures of the Jester dataset on both plain and complex backgrounds. We took 30 samples of each hand gesture in the plain and complex background and obtained 95.6% accuracy on plain background samples and 88.6% accuracy on complex background samples.

4.2.2. Experiment II: Hand Gestures Classification Accuracies

For hand gestures classification, we used a KATH classifier. The design method was evaluated using the leave one subject out (LOSO) cross-validation method. In Table 4, the results over the IPN hand video dataset show 88.46% hand gestures classification accuracy. Table 5 represents the confusion matrix for the Jester dataset with 87.69% mean accuracy for hand gestures classification.

4.2.3. Experiment III: Comparison with Other Classification Algorithms

In this segment, we compared the recall, precision, and F1-measure over the IPN hand dataset and the Jester dataset. For the classification of hand gestures, we used a decision tree, an artificial neural network, and we associated the consequences with the KATH classifier. Figure 12 shows the results over the IPN hand dataset and Figure 13 shows the results over the Jester dataset.

4.2.4. Experiment IV: Comparison of our Proposed System with State-of-the-Art Techniques

In this section, we have compared the proposed model with other well-known techniques using the same datasets. Table 6 shows the comparative results between the proposed model and other state-of-the-art techniques.

5. Conclusions

This article is based on a hand gestures recognition system for controlling smart home appliances. Two benchmark datasets were selected for experiments: the IPN hand dataset and Jester dataset. Initially, images are acquired, in which hands are detected and landmarks are localized on the palm and the fingers. After that, the textures-based and point-based features are extracted. The hand skeletons are used for extracting the point-based features, whereas the full hand is used for extracting the texture-based features. For feature reduction and optimization, the fuzzy logic is adopted, and finally, the K-ary classification algorithm is used for classifying the hand gestures for operating smart home appliances. For the IPN hand dataset, we achieved the mean accuracy of 88.46% and for the Jester dataset, a mean accuracy of 87.69% was achieved. The proposed system’s performance shows a significant improvement compared to existing state-of-the-art frameworks. The limitation of the proposed framework is due to the complexity in the videos, such as cluttered backgrounds and various illumination conditions, which make it difficult to achieve more accurate results.

Author Contributions

Conceptualization, H.A. and A.K.; methodology, H.A. and A.J.; software, H.A. and A.K.; validation, H.A., M.S., S.A.A. and A.A.; formal analysis, A.K., S.A.A. and J.P.; resources, M.S., A.K., A.A. and J.P.; writing—review and editing, H.A., A.K., S.A.A. and J.P.; funding acquisition, A.A., S.A.A. and J.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-2018-0-01426) supervised by the IITP (Institute for Information & Communications Technology Planning and Evaluation). In addition, the authors would like to thank the support of the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University. This work was partially supported by the Taif University Researchers Supporting Project number (TURSP-2020/115), Taif University, Taif, Saudi Arabia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, M.A.; Javed, K.; Khan, S.A.; Saba, T.; Habib, U.; Khan, J.A.; Abbasi, A.A. Human action recognition using fusion of multiview and deep features: An application to video surveillance. Multimed. Tools Appl. 2020, 19, 1–27. [Google Scholar] [CrossRef]
  2. Zou, Y.; Shi, Y.; Shi, D.; Wang, Y.; Liang, Y.; Tian, Y. Adaptation-Oriented Feature Projection for One-shot Action Recognition. IEEE Trans. Multimed. 2020, 99, 10. [Google Scholar] [CrossRef]
  3. Ghadi, Y.; Akhter, I.; Alarfaj, M.; Jalal, A.; Kim, K. Syntactic model-based human body 3D reconstruction and event classification via association based features mining and deep learning. PeerJ Comput. Sci. 2021, 7, e764. [Google Scholar] [CrossRef] [PubMed]
  4. Van der Kruk, E.; Reijne, M.M. Accuracy of human motion capture systems for sport applications; state-of-the-art review. Eur. J. Sport Sci. 2018, 18, 6. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, Y.; Mori, G. Multiple tree models for occlusion and spatial constraints in human pose estimation. In Proceedings of the European Conference on Computer Vision, Marseille, France, 12–18 October 2008; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  6. Amft, O.; Tröster, G. Recognition of dietary activity events using on-body sensors. Artif. Intell. Med. 2008, 42, 121–136. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Sun, S.; Kuang, Z.; Sheng, L.; Ouyang, W.; Zhang, W. Optical Flow Guided Feature: A Fast and Robust Motion Representation for Video Action Recognition. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  8. Zhu, Y.; Zhou, K.; Wang, M.; Zhao, Y.; Zhao, Z. A comprehensive solution for detecting events in complex surveillance videos. Multimed. Tools Appl. 2019, 78, 1. [Google Scholar] [CrossRef]
  9. Akhter, I.; Jalal, A.; Kim, K. Adaptive Pose Estimation for Gait Event Detection Using Context-Aware Model and Hierarchical Optimization. J. Electr. Eng. Technol. 2021, 16, 2721–2729. [Google Scholar] [CrossRef]
  10. Jalal, A.; Lee, S.; Kim, J.T.; Kim, T.S. Human activity recognition via the features of labeled depth body parts. In Proceedings of the International Conference on Smart Homes and Health Telematics, Artiminio, Italy, 12–15 June 2012; Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  11. Ghadi, Y.; Manahil, W.; Tamara, S.; Suliman, A.; Jalal, A.; Park, J. Automated parts-based model for recognizing human-object interactions from aerial imagery with fully convolutional network. Remote Sens. 2022, 14, 1492. [Google Scholar] [CrossRef]
  12. Jalal, A.; Sarif, N.; Kim, J.T.; Kim, T.S. Human activity recognition via recognized body parts of human depth silhouettes for residents monitoring services at smart home. Indoor Built Environ. 2013, 22, 271–279. [Google Scholar] [CrossRef]
  13. Jalal, A.; Kim, Y.; Kim, D. Ridge body parts features for human pose estimation and recognition from RGB-D video data. In Proceedings of the Fifth International Conference on Computing, Communications and Networking Technologies (ICCCNT), Hefei, China, 11–14 July 2014; pp. 1–6. [Google Scholar]
  14. Akhter, I.; Jalal, A.; Kim, K. Pose estimation and detection for event recognition using Sense-Aware features and Adaboost classifier. In Proceedings of the 2021 International Bhurban Conference on Applied Sciences and Technologies (IBCAST), Islamabad, Pakistan, 12–16 January 2021. [Google Scholar]
  15. Jalal, A.; Kamal, S.; Kim, D. Depth Map-based Human Activity Tracking and Recognition Using Body Joints Features and Self-Organized Map. In Proceedings of the 5th International Conference on Computing, Communications and Networking Technologies (ICCCNT), Hefei, China, 11–13 July 2014. [Google Scholar]
  16. Ghadi, Y.; Akhter, I.; Suliman, A.; Tamara, S.; Jalal, A.; Park, J. Multiple events detection using context-intelligence features. IASC 2022, 34, 3. [Google Scholar]
  17. Jalal, A.; Kamal, S. Real-time life logging via a depth silhouette-based human activity recognition system for smart home services. In Proceedings of the 2014 11th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Seoul, Korea, 26–29 August 2014; pp. 74–80. [Google Scholar]
  18. Jalal, A.; Kamal, S.; Kim, D. A depth video sensor-based life-logging human activity recognition system for elderly care in smart indoor environments. Sensors 2014, 14, 11735–11759. [Google Scholar] [CrossRef] [PubMed]
  19. Ghadi, Y.Y.; Akhter, I.; Aljuaid, H.; Gochoo, M.; Alsuhibany, S.A.; Jalal, A.; Park, J. Extrinsic Behavior Prediction of Pedestrians via Maximum Entropy Markov Model and Graph-Based Features Mining. Appl. Sci. 2022, 12, 5985. [Google Scholar] [CrossRef]
  20. Gochoo, M.; Tahir, S.B.U.D.; Jalal, A.; Kim, K. Monitoring Real-Time Personal Locomotion Behaviors Over Smart Indoor-Outdoor Environments Via Body-Worn Sensors. IEEE Access 2021, 9, 70556–70570. [Google Scholar] [CrossRef]
  21. Pervaiz, M.; Ghadi, Y.Y.; Gochoo, M.; Jalal, A.; Kamal, S.; Kim, D.-S. A Smart Surveillance System for People Counting and Tracking Using Particle Flow and Modified SOM. Sustainability 2021, 13, 5367. [Google Scholar] [CrossRef]
  22. Jalal, A.; Akhtar, I.; Kim, K. Human Posture Estimation and Sustainable Events Classification via Pseudo-2D Stick Model and K-ary Tree Hashing. Sustainability 2020, 12, 9814. [Google Scholar] [CrossRef]
  23. Khalid, N.; Ghadi, Y.Y.; Gochoo, M.; Jalal, A.; Kim, K. Semantic Recognition of Human-Object Interactions via Gaussian-Based Elliptical Modeling and Pixel-Level Labeling. IEEE Access 2021, 9, 111249–111266. [Google Scholar] [CrossRef]
  24. Trong, K.N.; Bui, H.; Pham, C. Recognizing hand gestures for controlling home appliances with mobile sensors. In Proceedings of the 2019 11th International Conference on Knowledge and Systems Engineering (KSE), Da Nang, Vietnam, 24–26 October 2019; pp. 1–7. [Google Scholar]
  25. Senanayake, R.; Kumarawadu, S. A robust vision-based hand gesture recognition system for appliance control in smart homes. In Proceedings of the 2012 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2012), Hong Kong, China, 12–15 August 2012; pp. 760–763. [Google Scholar]
  26. Chong, Y.; Huang, J.; Pan, S. Hand Gesture recognition using appearance features based on 3D point cloud. J. Softw. Eng. Appl. 2016, 9, 103–111. [Google Scholar] [CrossRef] [Green Version]
  27. Solanki, U.V.; Desai, N.H. Hand gesture based remote control for home appliances: Handmote. In Proceedings of the 2011 World Congress on Information and Communication Technologies, Mumbai, India, 11–14 December 2011; pp. 419–423. [Google Scholar]
  28. Jamaludin, N.A.N.; Fang, O.H. Dynamic Hand Gesture to Text using Leap Motion. Int. J. Adv. Comput. Sci. Appl. 2019, 10, 199–204. [Google Scholar] [CrossRef] [Green Version]
  29. Chellaswamy, C.; Durgadevi, J.J.; Srinivasan, S. An intelligent hand gesture recognition system using fuzzy logic. In Proceedings of the IET Chennai Fourth International Conference on Sustainable Energy and Intelligent Systems (SEISCON 2013), Chennai, India, 12–14 December 2013; pp. 326–332. [Google Scholar]
  30. Yang, P.-Y.; Ho, K.-H.; Chen, H.-C.; Chien, M.-Y. Exercise training improves sleep quality in middle-aged and older adults with sleep problems: A systematic review. J. Physiother. 2012, 58, 157–163. [Google Scholar] [CrossRef] [Green Version]
  31. Farooq, A.; Jalal, A.; Kamal, S. Dense RGB-D Map-Based Human Tracking and Activity Recognition using Skin Joints Features and Self-Organizing Map. KSII Trans. Internet Inf. Syst. 2015, 9, 1856–1869. [Google Scholar]
  32. Jalal, A.; Kamal, S.; Kim, D. Depth silhouettes context: A new robust feature for human tracking and activity recognition based on embedded HMMs. In Proceedings of the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI 2015), Goyang City, Korea, 28 October 2015. [Google Scholar]
  33. Jalal, A.; Kamal, S.; Kim, D. Individual detection-tracking-recognition using depth activity images. In Proceedings of the 2015 12th International Conference on Ubiquitous Robots and Ambient Intelligence (URAI), Goyangi, Korea, 28–30 October 2015; pp. 450–455. [Google Scholar]
  34. Kamal, S.; Jalal, A. A hybrid feature extraction approach for human detection, tracking and activity recognition using depth sensors. Arab. J. Sci. Eng. 2016, 41, 1043–1051. [Google Scholar] [CrossRef]
  35. Jalal, A.; Kim, Y.-H.; Kim, Y.-J.; Kamal, S.; Kim, D. Robust human activity recognition from depth video using spatiotemporal multi-fused features. Pattern Recognit. 2017, 61, 295–308. [Google Scholar] [CrossRef]
  36. Kamal, S.; Jalal, A.; Kim, D. Depth images-based human detection, tracking and activity recognition using spatiotemporal features and modified HMM. J. Electr. Eng. Technol. 2016, 11, 1857–1862. [Google Scholar] [CrossRef] [Green Version]
  37. Gochoo, M.; Akhter, I.; Jalal, A.; Kim, K. Stochastic Remote Sensing Event Classification over Adaptive Posture Estimation via Multifused Data and Deep Belief Network. Remote Sens. 2021, 13, 912. [Google Scholar] [CrossRef]
  38. Jalal, A.; Kamal, S.; Kim, D. Facial Expression recognition using 1D transform features and Hidden Markov Model. J. Electr. Eng. Technol. 2017, 12, 1657–1662. [Google Scholar]
  39. Jalal, A.; Kamal, S.; Kim, D. A Depth Video-based Human Detection and Activity Recognition using Multi-features and Embedded Hidden Markov Models for Health Care Monitoring Systems. Int. J. Interact. Multimed. Artif. Intell. 2017, 4, -62–62. [Google Scholar] [CrossRef] [Green Version]
  40. Jalal, A.; Kamal, S.; Kim, D.-S. Detecting complex 3D human motions with body model low-rank representation for real-time smart activity monitoring system. KSII Trans. Internet Inf. Syst. 2018, 12, 1189–1204. [Google Scholar]
  41. Jalal, A.; Kamal, S. Improved Behavior Monitoring and Classification Using Cues Parameters Extraction from Camera Array Images. Int. J. Interact. Multimed. Artif. Intell. 2019, 5, 71. [Google Scholar] [CrossRef]
  42. Jalal, A.; Quaid, M.A.K.; Hasan, A.S. Wearable sensor-based human behavior understanding and recognition in daily life for smart environments. In Proceedings of the 2018 International Conference on Frontiers of Information Technology (FIT), Islamabad, Pakistan, 17–19 December 2018; pp. 105–110. [Google Scholar]
  43. Mahmood, M.; Jalal, A.; Sidduqi, M.A. Robust spatio-Temporal features for human interaction recognition via artificial neural network. In Proceedings of the 2018 International Conference on Frontiers of Information Technology (FIT 2018), Islamabad, Pakistan, 17–19 December 2018. [Google Scholar]
  44. Jalal, A.; Quaid, M.A.K.; Sidduqi, M.A. A Triaxial acceleration-based human motion detection for ambient smart home system. In Proceedings of the 2019 16th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 8–12 January 2019; pp. 353–358. [Google Scholar]
  45. Jalal, A.; Mahmood, M.; Hasan, A.S. Multi-features descriptors for human activity tracking and recognition in Indoor-outdoor environments. In Proceedings of the 2019 16th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 8–12 January 2019; pp. 371–376. [Google Scholar]
  46. Jalal, A.; Mahmood, M. Students’ behavior mining in e-learning environment using cognitive processes with information technologies. Educ. Inf. Technol. 2019, 24, 2797–2821. [Google Scholar] [CrossRef]
  47. Jalal, A.; Nadeem, A.; Bobasu, S. Human Body Parts Estimation and Detection for Physical Sports Movements. In Proceedings of the 2019 2nd International Conference on Communication, Computing and Digital Systems (C-CODE 2019), Islamabad, Pakistan, 6–7 March 2019. [Google Scholar]
  48. Jalal, A.; Quaid, M.A.K.; Kim, K. A wrist worn acceleration based human motion analysis and classification for ambient smart home system. J. Electr. Eng. Technol. 2019, 14, 1733–1739. [Google Scholar] [CrossRef]
  49. Ahmed, A.; Jalal, A.; Kim, K. Region and decision tree-based segmentations for multi-objects detection and classification in outdoor scenes. In Proceedings of the 2019 International Conference on Frontiers of Information Technology (FIT 2019), Islamabad, Pakistan, 16–18 December 2019. [Google Scholar]
  50. Rafique, A.A.; Jalal, A.; Kim, K. Statistical multi-objects segmentation for indoor/outdoor scene detection and classification via depth images. In Proceedings of the 2020 17th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 14–18 January 2020; pp. 271–276. [Google Scholar]
  51. Ahmed, A.; Jalal, A.; Kim, K. RGB-D images for object segmentation, localization and recognition in indoor scenes using feature descriptor and Hough voting. In Proceedings of the 2020 17th International Bhurban Conference on Applied Sciences and Technology (IBCAST), Islamabad, Pakistan, 14–18 January 2020; pp. 290–295. [Google Scholar]
  52. Tamara, S.; Akhter, I.; Suliman, A.; Ghadi, Y.; Jalal, A.; Park, J. Pedestrian Physical Education Training over Visualization Tool. CMC 2022, 73, 2389–2405. [Google Scholar]
  53. Quaid, M.A.K.; Jalal, A. Wearable sensors based human behavioral pattern recognition using statistical features and reweighted genetic algorithm. Multimed. Tools Appl. 2020, 79, 6061–6083. [Google Scholar] [CrossRef]
  54. Nadeem, A.; Jalal, A.; Kim, K. Human Actions Tracking and Recognition Based on Body Parts Detection via Artificial Neural Network. In Proceedings of the 3rd International Conference on Advancements in Computational Sciences (ICACS 2020), Lahore, Pakistan, 17–19 February 2020. [Google Scholar]
  55. Badar Ud Din Tahir, S.; Jalal, A.; Batool, M. Wearable Sensors for Activity Analysis using SMO-based Random Forest over Smart home and Sports Datasets. In Proceedings of the 3rd International Conference on Advancements in Computational Sciences (ICACS 2020), Lahore, Pakistan, 17–19 February 2020. [Google Scholar]
  56. Rizwan, S.A.; Jalal, A.; Kim, K. An Accurate Facial Expression Detector using Multi-Landmarks Selection and Local Transform Features. In Proceedings of the 2020 3rd International Conference on Advancements in Computational Sciences (ICACS), Lahore, Pakistan, 17–19 February 2020; pp. 1–6. [Google Scholar]
  57. Ud din Tahir, S.B.; Jalal, A.; Kim, K. Wearable inertial sensors for daily activity analysis based on adam optimization and the maximum entropy Markov model. Entropy 2020, 22, 579. [Google Scholar] [CrossRef] [PubMed]
  58. Alam; Abduallah, S.; Akhter, I.; Suliman, A.; Ghadi, Y.; Tamara, S.; Jalal, A. Object detection learning for intelligent self automated vehicles. In Proceedings of the 2019 IEEE International Conference on Vehicular Electronics and Safety (ICVES), Cairo, Egypt, 4–6 September 2019. [Google Scholar]
  59. Jalal, A.; Khalid, N.; Kim, K. Automatic recognition of human interaction via hybrid descriptors and maximum entropy markov model using depth sensors. Entropy 2020, 22, 817. [Google Scholar] [CrossRef]
  60. Batool, M.; Jalal, A.; Kim, K. Telemonitoring of Daily Activity Using Accelerometer and Gyroscope in Smart Home Environments. J. Electr. Eng. Technol. 2020, 15, 2801–2809. [Google Scholar] [CrossRef]
  61. Jalal, A.; Batool, M.; Kim, K. Stochastic Recognition of Physical Activity and Healthcare Using Tri-Axial Inertial Wearable Sensors. Appl. Sci. 2020, 10, 7122. [Google Scholar] [CrossRef]
  62. Jalal, A.; Quaid, M.A.K.; Kim, K.; Tahir, S.B.U.D. A Study of Accelerometer and Gyroscope Measurements in Physical Life-Log Activities Detection Systems. Sensors 2020, 20, 6670. [Google Scholar] [CrossRef]
  63. Rafique, A.A.; Jalal, A.; Kim, K. Automated Sustainable Multi-Object Segmentation and Recognition via Modified Sampling Consensus and Kernel Sliding Perceptron. Symmetry 2020, 12, 1928. [Google Scholar] [CrossRef]
  64. Ansar, H.; Jalal, A.; Gochoo, M.; Kim, K. Hand Gesture Recognition Based on Auto-Landmark Localization and Reweighted Genetic Algorithm for Healthcare Muscle Activities. Sustainability 2021, 13, 2961. [Google Scholar] [CrossRef]
  65. Nadeem, A.; Jalal, A.; Kim, K. Automatic human posture estimation for sport activity recognition with robust body parts detection and entropy markov model. Multimed. Tools Appl. 2021, 80, 21465–21498. [Google Scholar] [CrossRef]
  66. Akhter, I. Automated Posture Analysis of Gait Event Detection via a Hierarchical Optimization Algorithm and Pseudo 2D Stick-Model. Master’s Thesis, Air University, Islamabad, Pakistan, December 2020. [Google Scholar]
  67. Ud din Tahir, S.B. A Triaxial Inertial Devices for Stochastic Life-Log Monitoring via Augmented-Signal and a Hierarchical Recognizer. Master’s Thesis, Air University, Islamabad, Pakistan, December 2020. [Google Scholar]
  68. Benitez-Garcia, G.; Olivares-Mercado, J.; Sanchez-Perez, G.; Yanai, K. IPN hand: A video dataset and benchmark for real-time continuous hand gesture recognition. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 4340–4347. [Google Scholar]
  69. Materzynska, J.; Berger, G.; Bax, I.; Memisevic, R. The jester dataset: A large-scale video dataset of human gestures. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  70. Yamaguchi, O.; Fukui, K. Image-set based Classification using Multiple Pseudo-whitened Mutual Subspace Method. In Proceedings of the 11th International Conference on Pattern Recognition Applications and Methods, Vienna, Austria, 3–5 February 2022. [Google Scholar]
  71. Zhou, B.; Andonian, A.; Oliva, A.; Torralba, A. Temporal relational reasoning in videos. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 803–818. [Google Scholar]
  72. Gammulle, H.; Denman, S.; Sridharan, S.; Fookes, C. TMMF: Temporal Multi-Modal Fusion for Single-Stage Continuous Gesture Recognition. IEEE Trans. Image Process. 2021, 30, 7689–7701. [Google Scholar] [CrossRef] [PubMed]
  73. Shi, L.; Zhang, Y.; Hu, J.; Cheng, J.; Lu, H. Gesture recognition using spatiotemporal deformable convolutional representation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1900–1904. [Google Scholar]
  74. Kopuklu, O.; Kose, N.; Rigoll, G. Motion fused frames: Data level fusion strategy for hand gesture recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2103–2111. [Google Scholar]
  75. Benitez-Garcia, G.; Prudente-Tixteco, L.; Castro-Madrid, L.C.; Toscano-Medina, R.; Olivares-Mercado, J.; Sanchez-Perez, G.; Villalba, L.J.G. Improving real-time hand gesture recognition with semantic segmentation. Sensors 2021, 21, 356. [Google Scholar] [CrossRef] [PubMed]
  76. Jalal, A.; Batool, M.; Kim, K. Sustainable wearable system: Human behavior modeling for life-logging activities using K-Ary tree hashing classifier. Sustainability 2020, 12, 10324. [Google Scholar] [CrossRef]
Figure 1. The proposed HGR system model structural design.
Figure 1. The proposed HGR system model structural design.
Applsci 12 06481 g001
Figure 2. Results of preprocessing on video frames over IPN dataset.
Figure 2. Results of preprocessing on video frames over IPN dataset.
Applsci 12 06481 g002
Figure 3. Results of the hand detection using a single-shot multibox detector.
Figure 3. Results of the hand detection using a single-shot multibox detector.
Applsci 12 06481 g003
Figure 4. Hand landmarks localization results in hand gestures using the skeleton method.
Figure 4. Hand landmarks localization results in hand gestures using the skeleton method.
Applsci 12 06481 g004
Figure 5. The results of the Bezier curve.
Figure 5. The results of the Bezier curve.
Applsci 12 06481 g005
Figure 6. Illustration of the change in the direction between different frames’ extreme points.
Figure 6. Illustration of the change in the direction between different frames’ extreme points.
Applsci 12 06481 g006
Figure 7. The appearance features using 3D point clouds.
Figure 7. The appearance features using 3D point clouds.
Applsci 12 06481 g007
Figure 8. The result of the fuzzy logic optimization.
Figure 8. The result of the fuzzy logic optimization.
Applsci 12 06481 g008
Figure 9. KATH classifier of optimized data over IPN hand dataset for our proposed model.
Figure 9. KATH classifier of optimized data over IPN hand dataset for our proposed model.
Applsci 12 06481 g009
Figure 10. A few example images of the IPN Hand dataset.
Figure 10. A few example images of the IPN Hand dataset.
Applsci 12 06481 g010
Figure 11. A few example images of the IPN Hand dataset.
Figure 11. A few example images of the IPN Hand dataset.
Applsci 12 06481 g011
Figure 12. Comparison results of Precision, recall, and F-1 Score using different classifiers over IPN Hand dataset.
Figure 12. Comparison results of Precision, recall, and F-1 Score using different classifiers over IPN Hand dataset.
Applsci 12 06481 g012
Figure 13. Comparison results of precision, recall, and F-1 Score using different classifiers over the Jester dataset.
Figure 13. Comparison results of precision, recall, and F-1 Score using different classifiers over the Jester dataset.
Applsci 12 06481 g013
Table 1. A comprehensive review of relevant research.
Table 1. A comprehensive review of relevant research.
Hand Gestures Recognition for Controlling Smart Home Appliances
MethodsMain Contributions
H. Khanh et al. (2019) [24]The system was developed for controlling smart home appliances using two deep learning models fused with mobile sensors to recognize hand gestures. The mobile sensors were instrumented on smartwatches, smartphones, and smart appliances. The deep learning models helped in the learning and representation of the mobile sensors’ data.
Ransalu et al. (2012) [25]The HGR model was developed to automate the home appliances using hand gestures. First, the hand was detected using the Viola-Jones object detection algorithm. To segment the hand from the image, the YCbCr skin color segmentation technique was used. The filtered hand was refined using dilation and eroding. At last, a multilayer perceptron was used to classify the four-hand gestures, i.e., ready, swing, on/speedup & off.
P. N. et al. (2017) [26]The model consists of a few steps. The hand gestures were captured using the web camera and were then preprocessed to detect the hands. Corner point detection was used to de-noise the images using a MATLAB simulation tool. Based on the gestures, different threshold values were used for controlling the home appliances. The threshold values were generated using the Fast Fourier transform algorithm and the appliances were controlled by the micro-controllers.
V. Utpal et al. (2011) [27]For controlling home appliances and electronic gadgets using hand gestures, the authors detected hands using the YCbCr skin color segmentation model and traced edges. For gesture recognition, the number of fingers was counted, and its orientation was analyzed. The reference background was stored from each frame which was compared with the next frame for reliable hand gesture recognition.
Santhana et al. (2020) [28]They developed a hand gesture recognition system using Leap motion sensors. The system was customized to recognize multiple motion-based hand gestures for controlling smart home appliances. The system was trained using a customized dataset containing various hand gestures to control daily household devices using a deep neural network.
Qi et al. (2013) [29]They developed a hand recognition system for controlling television. The system was categorized into three sub-categories. (1) For static hand gesture recognition, hand features were extracted using a histogram-oriented gradient (HOG), and for recognition, Adaboost training was used. (2) For dynamic hand gesture recognition, first the hand trajectory was recognized and passed through the HMM model for recognition. (3) For finger click recognition, a specific depth threshold was fixed to detect the fingers. The distance between the palm and the fingertip was calculated. The accumulated variance was calculated for each fingertip to recognize the finger click gesture.
Yueh et al. (2018) [30]The authors developed a system to control a TV using hand gestures. First, the hand was detected through skin segmentation and the hand contour was extracted. After that, the system was trained using CNN to recognize hand gestures that were categorized into five branches; (1) menu, (2) direction, (3) go back, (4) mute/unmute, and (5) nothing. After that, CNN also helped in tracking the hand joints to detect commands: (1) increase/ decrease the speed, (2) clicking, and (3) cursor movement.
Table 2. Hand detection accuracies over IPN Hand dataset.
Table 2. Hand detection accuracies over IPN Hand dataset.
Hand GesturesNumber of SamplesPlain BackgroundAccuracy (%)Cluttered BackgroundAccuracy (%)
POF30301002583.3
PTF30301002686.6
COF30301002686.6
CTF302893.32686.6
TU302996.62996.6
TD302996.630100
TL30279030100
TR302893.330100
OT302996.630100
DCOF30301002996.6
DCTF30301002996.6
ZI30301002893.3
ZO302996.630100
Mean Accuracy Rate 97.1% 94.3%
POF = pointing with one finger, PTF = pointing with two fingers, COF = click with one finger, CTF = click with two fingers, TU = throw up, TD = throw down, TL = throw left, TR = throw right, OT = open twice, DCOF = double click with one finger, DCTF = double click with two fingers, ZI = zoom in, ZO = zoom out.
Table 3. Hand detection accuracies over Jester Hand dataset.
Table 3. Hand detection accuracies over Jester Hand dataset.
Hand GesturesNumber of SamplesPlain BackgroundAccuracy (%)Cluttered BackgroundAccuracy (%)
SD30301002583.3
SL302996.62790
SR302893.32480
SU302893.32686.6
TD302996.62480
TU302893.330100
ZIF302996.62583.3
ZOF30279030100
S302996.62583.3
RF30301002996.6
RB30301002376.6
PI302893.32893.3
SH302996.630100
Mean Accuracy Rate 95.6% 88.6%
SD = swiping down, SL = swiping left, SR = swiping right, SU = swiping up, TD = thumb down, TU = thumb up, ZIF = zooming in with full hand, ZOF = zooming out with full hand, S = stop, RF=rolling hand forward, RB = rolling hand backward, PI = pulling hand in, SH = shaking hand.
Table 4. Confusion Matrix results over IPN Hand dataset.
Table 4. Confusion Matrix results over IPN Hand dataset.
ClassPOFPTFCOFCTFTUTDTLTROTDCOFDCTFZIZO
POF8101000000000
PTF0900001000000
COF00100000000000
CTF0109000000000
TU1000900000000
TD0010080001000
TL0200007000010
TR1000000800010
OT0000000090100
DCOF00000000010000
DCTF0000000000901
ZI00000000000100
ZO0000000000019
Hand Gestures classification mean accuracy = 88.46%
POF = pointing with one finger, PTF = pointing with two fingers, COF = click with one finger, CTF = click with two fingers, TU = throw up, TD = throw down, TL = throw left, TR = throw right, OT = open twice, DCOF = double click with one finger, DCTF = double click with two fingers, ZI = zoom in, ZO= zoom out.
Table 5. Confusion Matrix results over Jester dataset.
Table 5. Confusion Matrix results over Jester dataset.
GesturesSDSLSRSUTDTUZIFZOFSRFRBPISH
SD9000100000000
SL0900000010000
SR0090100000000
SU0018100000000
TD00001000000000
TU0000180010000
ZIF0000009001000
ZOF0000100900000
S0100000080010
RF0000100009000
RB00000000001000
PI0000000000190
SH0000001000117
Hand Gestures classification mean accuracy = 87.69%
SD = swiping down, SL = swiping left, SR = swiping right, SU = swiping up, TD = thumb down, TU = thumb up, ZIF = zooming in with full hand, ZOF = zooming out with full hand, S = stop, RF = rolling hand forward, RB = rolling hand backward, PI = pulling hand in, SH = shaking hand.
Table 6. Hand gestures recognition results from the proposed model with other state-of-the-art techniques.
Table 6. Hand gestures recognition results from the proposed model with other state-of-the-art techniques.
AuthorsIPN Hand Dataset (%)AuthorsJester Dataset (%)
Yamaguchi et al. (2022) [70]60.00 Zhou et al. (2018) [71]82.02
Gammulle et al. (2021) [72]80.03Shi et al. (2019) [73]82.34
Garcia et al. (2020) [68]82.36Kopuklu et al. (2018) [74]84.70
TSN [75]68.01MFFs [76]84.70
Proposed method88.46Proposed method87.69
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ansar, H.; Ksibi, A.; Jalal, A.; Shorfuzzaman, M.; Alsufyani, A.; Alsuhibany, S.A.; Park, J. Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier. Appl. Sci. 2022, 12, 6481. https://doi.org/10.3390/app12136481

AMA Style

Ansar H, Ksibi A, Jalal A, Shorfuzzaman M, Alsufyani A, Alsuhibany SA, Park J. Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier. Applied Sciences. 2022; 12(13):6481. https://doi.org/10.3390/app12136481

Chicago/Turabian Style

Ansar, Hira, Amel Ksibi, Ahmad Jalal, Mohammad Shorfuzzaman, Abdulmajeed Alsufyani, Suliman A. Alsuhibany, and Jeongmin Park. 2022. "Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier" Applied Sciences 12, no. 13: 6481. https://doi.org/10.3390/app12136481

APA Style

Ansar, H., Ksibi, A., Jalal, A., Shorfuzzaman, M., Alsufyani, A., Alsuhibany, S. A., & Park, J. (2022). Dynamic Hand Gesture Recognition for Smart Lifecare Routines via K-Ary Tree Hashing Classifier. Applied Sciences, 12(13), 6481. https://doi.org/10.3390/app12136481

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop