Next Article in Journal
Automated Vehicle Counting from Pre-Recorded Video Using You Only Look Once (YOLO) Object Detection Model
Next Article in Special Issue
Breast Cancer Detection with an Ensemble of Deep Learning Networks Using a Consensus-Adaptive Weighting Method
Previous Article in Journal
A Joint De-Rain and De-Mist Network Based on the Atmospheric Scattering Model
Previous Article in Special Issue
The Effectiveness of Pan-Sharpening Algorithms on Different Land Cover Types in GeoEye-1 Satellite Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework

Department of Computer Science, Kansas State University, Manhattan, KS 66506, USA
*
Author to whom correspondence should be addressed.
J. Imaging 2023, 9(7), 130; https://doi.org/10.3390/jimaging9070130
Submission received: 1 May 2023 / Revised: 13 June 2023 / Accepted: 19 June 2023 / Published: 26 June 2023
(This article belongs to the Special Issue Image Processing and Computer Vision: Algorithms and Applications)

Abstract

:
Vision-based human activity recognition (HAR) has emerged as one of the essential research areas in video analytics. Over the last decade, numerous advanced deep learning algorithms have been introduced to recognize complex human actions from video streams. These deep learning algorithms have shown impressive performance for the video analytics task. However, these newly introduced methods either exclusively focus on model performance or the effectiveness of these models in terms of computational efficiency, resulting in a biased trade-off between robustness and computational efficiency in their proposed methods to deal with challenging HAR problem. To enhance both the accuracy and computational efficiency, this paper presents a computationally efficient yet generic spatial–temporal cascaded framework that exploits the deep discriminative spatial and temporal features for HAR. For efficient representation of human actions, we propose an efficient dual attentional convolutional neural network (DA-CNN) architecture that leverages a unified channel–spatial attention mechanism to extract human-centric salient features in video frames. The dual channel–spatial attention layers together with the convolutional layers learn to be more selective in the spatial receptive fields having objects within the feature maps. The extracted discriminative salient features are then forwarded to a stacked bi-directional gated recurrent unit (Bi-GRU) for long-term temporal modeling and recognition of human actions using both forward and backward pass gradient learning. Extensive experiments are conducted on three publicly available human action datasets, where the obtained results verify the effectiveness of our proposed framework (DA-CNN+Bi-GRU) over the state-of-the-art methods in terms of model accuracy and inference runtime across each dataset. Experimental results show that the DA-CNN+Bi-GRU framework attains an improvement in execution time up to 167× in terms of frames per second as compared to most of the contemporary action-recognition methods.

1. Introduction

The recent advancements in artificial intelligence (AI), in particular, deep learning-driven vision algorithms, and microelectronics have made possible automated surveillance on Internet of things (IoT) and edge devices [1]. Generally, these surveillance systems are comprised of multiple interconnected cameras deployed in public places, such as offices, roads, shopping malls, hospitals, and airports, to enhance public safety and security [2]. The primary objective behind the deployment of surveillance systems in the aforementioned places is to instantly detect abnormalities by recognizing the anomalous human behavior or activity in a video stream which could result in injury or unlawful conduct. Human activity recognition is a process to analyze the hidden sequential pattern and predict the status of activity based on the perceptual context in input video streams. Generally, in videos, human activity is a combination of different movements of human body parts (i.e., hands, legs, or a combination of both). For instance, running involves rapid movement of hands and legs; similarly, throwing object involves the backward and forward force of arms and hands. Human activity recognition has numerous potential applications, such as in smart surveillance systems [3], video summarization [4], content-based video retrieval [5], sports and healthcare [6], and human–computer interactions [7]. In video, each frame contributes spatial information in sequential order which forms a sequential pattern containing human activity that cannot be recognized in a single video frame. Throwing a ball or a dart (which involves forward and backward force of an arm and hand) have the same action pose in the starting frame and thus the discrimination between these two distinct activities is challenging with respect to recognizing the action in a single frame. Investigating the same movements of an arm and hand in succeeding frames together with the information from previous frames will enable effective recognition of human activities in video stream data.
The earlier developed methods in initial research for vision-based activity recognition are exclusively focused on activities performed by a single person/actor in simple and controlled environments. In contrast, the current research focuses on more challenging and realistic human activities recorded with a cluttered complex background, variation in viewpoint, occlusion in background, inter- and intra-class variations, and pose variations. The existing vision-based human activity-recognition methods can be categorized into two classes namely: (i) traditional handcrafted feature-based, and (ii) deep learning-based human activity-recognition methods. The traditional handcrafted feature-based methods [8,9,10,11,12] use manually designed handcrafted or hand-engineered features (which requires extensive human efforts with prior knowledge of scene understanding) followed by statistical machine learning models to recognize the activity. For instance, several traditional image features have been utilized to analyze videos, such as histogram of 3D oriented gradients (HOG3D), histogram optical flow (HOF) [13], motion boundary histogram (MBH) [14], and extended speeded up robust feature (SURF) feature descriptors. The hand-engineered features must be designed specifically for each particular environment based on scene perceptual complexity. These types of manually designed handcrafted feature-based methods are ineffective while tackling long-term temporal dependencies and complex scenes. Recently, deep learning-based methods have made incredible breakthroughs in various domains of image processing and computer vision, and have been actively used for human activity recognition [15,16,17,18,19]. These deep learning-based methods have obtained state-of-the-art performance by extracting deep progressive discriminative features using different convolutional neural network (CNN) kernels and exploiting a gradient learning strategy. Unlike traditional handcrafted features of an image, deep CNNs learn progressively strong features (containing low-level, mid-level, and high-level features) that help to keep track of all types of visual semantics in image data.
Deep learning-based methods have enhanced the activity-recognition solutions in two perspectives. First, CNNs have the ability to extract more generic and semantically rich features than those of traditional handcrafted feature descriptors. Due to this generic feature extraction enabled by CNNs, CNNs have proliferated in a variety of complex computer vision tasks including 3D image reconstruction [20], image and video captioning [21], and text-to-image generation [22] that cannot be accomplished using traditional handcrafted feature-based methods due to their limitations in terms of features and learning strategies. Secondly, deep learning offers efficient architectures called recurrent neural networks (RNNs) which have the ability to learn representations of human activity from a batch of frames (sequence of frames or temporal representation of human activity) rather than a single frame. Earlier traditional methods consider frame-level classification of human activity in videos, rather than understanding the activity in a sequence of frames that greatly limits their performance for complex and multi-person activities. To cope with this challenge, deep learning-based methods have adopted RNNs for better understanding and recognition of complex human activities in videos. Normally, in deep learning-based methods, RNNs are placed right after CNNs, where the CNN architecture is responsible for extracting deep discriminative features from videos and the RNN is responsible for learning the hidden sequential patterns in the extracted CNN features. The performance of these deep learning methods is good compared to traditional methods; however, these methods are computationally very expensive due to their hybrid and complex CNN and RNN architectures.
The above-mentioned deep learning-based activity-recognition methods have attained exceptional performance. Most of the existing AI-assisted activity-recognition methods have adopted large yet effective pre-trained CNN architectures trained on a large-scale image dataset having tens of millions of trained parameters. Fusing such a computationally expensive feature descriptor backbone architecture with long short-term memory (LSTM) networks or multi-layer LSTMs (LSTMs having several layers with the same settings) greatly increases the computational complexity of the overall method, thereby compromising on the better tradeoff between model accuracy and complexity. Considering the demand for computationally efficient yet effective approaches that provide a balanced tradeoff between model accuracy and complexity for deployment on resource-constrained IoT and/or edge devices, in this paper, we propose a deep learning-based computationally efficient yet effective method for activity-recognition problems that can be deployed even on resource-constrained edge devices in the IoT-enabled surveillance environment. Our main contributions in this work are as follows:
  • We propose a computationally efficient cascaded spatial–temporal learning approach for human activity recognition. The proposed system utilizes deep discriminative RGB features guided by a channel–spatial attention mechanism and long-term modeling of action-centric features for reliable recognition of human activities in video streams.
  • We propose a light-weight CNN architecture having a total of eight convolutional layers where the maximum number of kernels used per layer is 64 with spatial dimensions 3 × 3 . With these constrained settings, we have developed a compact yet efficient CNN architecture for deep discriminative feature extraction as opposed to complex deep CNNs utilized by other contemporary works in their activity-recognition models using transfer learning.
  • We design a stacked dual channel–spatial attention mechanism with residual skip connection for spatial saliency extraction from video frames. The developed dual attentional module is placed after each two-consecutive convolutional layers of the developed CNN model which helps the network to extract saliency-aware deep discriminative features for localizing the action-specific regions in video frames.
  • We propose a bi-directional GRU network with three bi-directional layers (having forward and backward pass) that efficiently capture the long-term temporal patterns of human actions in both forward and backward directions, which greatly enhances the reusability of features, improves the feature propagation, and alleviates the issue of gradients vanishing.
  • We demonstrate the effectiveness and suitability of the proposed encapsulated dual attention CNN and bi-directional GRU framework (DA-CNN+Bi-GRU) for resource-constrained IoT and edge devices by comparing the model accuracy and execution/inference time of the DA-CNN+Bi-GRU framework with various baseline methods as well as contemporary human action-recognition methods.
The remainder of this paper is organized as follows. Section 2 provides a brief overview of the related works covering different types of methods introduced for human activity recognition. The proposed DA-CNN+Bi-GRU framework and its technical components are discussed in detail in Section 3. Section 4 presents an extensive experimental evaluation of DA-CNN+Bi-GRU as compared to other methods based on different metrics. Finally, Section 5 concludes the paper with possible future research directions.

2. Related Works on Human Activity Recognition

In recent years, human action and activity recognition have been widely studied and have received an exceptional amount of attention from computer vision researchers due to the recent success of deep learning for image-classification and object-detection tasks. Comprehensive reviews of both traditional and deep learning-based methods have been presented in numerous surveys [23,24]. The reported literature on human action and activity recognition can be summarized in terms of handcrafted feature-based methods, deep learning feature-based methods, long-term temporal modeling-based methods, and attention model-based methods. This section presents a brief discussion on these representative methods and a brief summary of previous related works.

2.1. Handcrafted Feature-Based Methods

Numerous traditional handcrafted feature-based methods have been proposed to localize spatial and temporal variations in videos using manually hand-engineered feature descriptors. Generally, these handcrafted feature-based methods can be structured as a feature-extraction and -encoding pipeline having three phases including key feature point detection (spatial and temporal feature points), quantization of detected features, and feature encoding. The first phase involves the extraction of spatial–temporal features from video frames, followed by feature quantization in the second phase which quantizes local motion-centric features. Lastly, the quantized spatial–temporal features are then encoded into feature vectors (known as action feature vectors) having fixed dimensions. For instance, inspired by the feature-extraction mechanism of the scale-invariant feature transform (SIFT) descriptor, Scovanner et al. [25] adopted the SIFT algorithm feature-extraction strategy and extended their feature space from 2D to 3D for encoding hidden action patterns. As single feature representation is not able to capture human actions, numerous multi-feature representative descriptors have been proposed in the literature. Laptev et al. [26] have proposed a multiscale spatial–temporal feature-based approach by utilizing space–time extension and the Harris operator. They first extract multi-scale spatial–temporal features from video frames and then characterize the appearance and motion of local features using a volumetric histogram of oriented gradients (HOG). The retrieved multi-scale spatial–temporal features are then fed to a non-linear support vector machine (SVM) for action recognition. In [27], Ryoo and Matthies inspected the behavior of local and global motion features to recognize first person activities in video data. Their proposed methods exclusively focus on temporal structures depicted in first person action/activity videos. These traditional handcrafted feature-based methods have shown progressive improvement over the years by presenting more efficient approaches; however, these methods are time-consuming (lacking end-to-end recognition strategy), labor-intensive (requiring extensive human efforts to extract generic and more discriminative features), and difficult to adopt in diverse scenarios.

2.2. Deep Learning Feature-Based Methods

Deep learning feature-based methods are the current mainstream methods to solve the problem of complex human action and activity recognition in videos. With the recent success in the computer vision domain for high-level vision tasks including image enhancement [28], image segmentation [29], and video captioning [30], CNNs have been actively investigated, addressing human action and activity-recognition problems. Numerous CNN-assisted methods have been presented [31,32,33,34,35,36,37] with deep CNN architectures with 2D convolution kernels applied across convolutional layers of the CNN. These convolutional layers extract deep discriminative spatial features with translation invariance from action video frames, offering reasonable action-recognition performance without using temporal modeling. For instance, Karpathy et al. [31] presented a single-stage CNN architecture for action recognition where they trained their proposed model on a large-scale sports video dataset benchmark, namely the Sports-1M dataset. Although their method achieves better results than traditional handcrafted feature-based methods, the presented architecture is unable to cope with temporal modeling. To overcome this issue, several two-stream CNN architectures were introduced [32,33,37] to obtain both spatial and temporal modeling of human action where one architecture performs spatial modeling of spatial contextual features and the second architecture performs temporal modeling using extracted optical flow features. The addition of a second network improves the performance by introducing temporal modeling to the CNN-based action-recognition approach; however, it equally increases the computational complexity of the overall two-stream CNN approach. To achieve spatial modeling and temporal cues within a single CNN architecture without compromising on model complexity, 3D CNNs [38,39,40] were introduced for human action-recognition tasks. For instance, Tran et al. [38] exploit the powerful characteristics of 3D CNN to recognize human action in sports videos, where they trained their proposed architecture on a large-scale benchmark dataset and have shown promising results. However, these 3D CNN-based approaches work well with short-term temporal modeling and lack the ability to cope with long temporal modeling. Oikonomou et al. [41] examined the effectiveness of advanced data-driven classifiers for action recognition by focusing on a specific set of joint coordinates (x, y, and z-axes). The underlying assumption is that observing this particular set of joints is sufficient for accurately perceiving each action in real-life scenarios. Their objective is to explore the capabilities of joint analysis in improving pose-based action-recognition systems. Consequently, they established correlations between specific joints and corresponding actions, identifying the most influential joints in the process. Shah et al. [42] addressed the task of action recognition using joint-based information. In contrast to alternative modalities, they leveraged the arrangement of joints and their motion to create concise models that capture essential human motion details for activity recognition. Their approach introduces a novel model for joint-based action recognition. The model initially extracts motion features from individual joints independently using a shared motion encoder, and subsequently performs collective reasoning. Holte et al. [43] examined the latest methods in multi-view techniques for the estimation of 3D human poses and recognition of human activities. They explored various fields where these technologies are applied and the specific needs associated with them. These areas include advanced human–computer interaction (HCI), assisted living, interactive games based on gestures, intelligent driver assistance systems, movies, 3D TV and animation, physical therapy, autonomous mental development, smart environments, sport motion analysis, video surveillance, and video annotation. They then conducted a thorough analysis of recent approaches proposed to meet these specific requirements and categorized them accordingly. Nandagopal et al. [44] proposed a novel method for activity recognition called KPE-DCNN. This technique involves several stages: the input video is first transformed into a series of frames, followed by key point extraction using a customized OpenPose model. The extracted key points are then used to classify human activities by training an optimized DCNN model. The goal of KPE-DCNN is to accurately extract key points and effectively recognize different activities based on these points. Zhou et al. [45] proposed a cascaded architecture to tackle the complex task of multi-stage, coarse-to-fine human–object interaction (HOI) understanding. In their approach, each stage of the architecture consists of an instance localization network that progressively refines HOI proposals. These refined proposals are then passed on to an interaction-recognition network. Notably, both networks maintain connections with their respective predecessors from the previous stage, facilitating cross-stage information exchange. The interaction-recognition network comprises two key components: a relation ranking module for selecting high-quality HOI proposals and a triple-stream classifier for relation prediction. These modules synergistically utilize carefully designed human-centric relation features to achieve effective interaction understanding. To study instance-aware posing of human body parts, Zhou et al. [46] proposed a novel bottom-up approach that simultaneously addresses the tasks of category-level human semantic segmentation and multi-person pose estimation in a joint and end-to-end fashion. This approach yields a compact, efficient, and robust framework that leverages structural information across various levels of human granularity, thereby mitigating the challenges associated with person partitioning. The key innovation of this work is the learning of a dense-to-sparse projection field, which facilitates the explicit association of dense human semantics with sparse keypoints. This projection field is progressively refined throughout the network’s feature pyramid, resulting in an improved performance and representation of the relationships between human semantics and pose keypoints. The resulting framework benefits from its ability to incorporate and enhance structural information, providing a powerful tool for the given tasks.

2.3. Temporal Modeling-Based Methods

The temporal modeling approach has been actively used to overcome the issue of long-term temporal modeling, where researchers have introduced a special kind of neural network called RNN, which has the ability to deal with long-term sequences. Different variants of RNNs have been introduced for action-recognition problems including LSTM [47], bi-directional LSTM [48], and GRU [49], which are comparatively more efficient than RNNs in terms of memorizing contents for long periods of time. For instance, Yue et al. [34] presented a two-stream CNN architecture to extract both spatial (edge, color and shape) features and temporal (optical flow) features stacked with the LSTM model for temporal modeling of human activity. Similarly, Amin et al. [47] presented a two-stream CNN architecture followed by a multi-layer LSTM to recognize human activity in videos. They first extracted spatial salient and optical flow features and then fed the extracted features to multi-layer LSTM to localize human action in video sequences. Ibrahim et al. [50] proposed a two-stream temporal modeling-based activity-recognition framework to recognize a team or group of activities. Their proposed method consist of two LSTM networks; the first LSTM learns the representation of a single person action, whereas the second LSTM is responsible for understanding collective activity by aggregating individual actions in a sequence of frames. Biswas et al. [51] presented a special variant of RNN named structural RNN for group activity recognition. Their proposed method consists of series of interconnected RNNs structured to analyze human actions and their mutual interactions in video sequences. To accurately learn representations of human activity in feature-encoded video frames, Shugao et al. [52] reformulated ranking loss to efficiently detect human activities. They first extracted deep discriminative CNN features from video frames using VGG19, which are then fed into LSTM to analyze hidden sequential patterns and recognize human activities. Muhammad et al. [6] presented a spatio-termporal approach for recognizing salient events in soccer videos, where they used a pretrained ResNet50 architecture for deep feature extraction and a multilayer LSTM for event recognition from the hidden sequential patterns. These hybrid CNN+LSTM approaches exhibited significant performance for vision-based human action- and activity-recognition tasks; these methods are computationally complex due to intensive computation caused by CNN feature extraction and human action modeling by LSTM.

2.4. Attention Mechanism-Based Methods

In the recent past, attention-based method have demonstrated great potential for a variety of high-level vision tasks including image segmentation [53], video captioning [54], and visual question answering (VQA) [55]. More recently, attention mechanisms combined with CNN and RNN networks have been widely used for human action-recognition tasks and have achieved noticeable improvements in action-recognition performance. For instance, Baradel et al. [56] introduced a novel spatio-temporal attention mechanism for human action recognition. Their approach automatically directs attention to the most significant human hands and detects the most discriminative moments within an action. Unlike conventional soft-attention mechanisms, they employed an RNN to handle attention in a fully differentiable manner. Notably, they diverged from the typical practice of using the hidden RNN state as an input to the attention model. Instead, they generated attention distributions using external information, specifically human articulated poses. Islam et al. [57] proposed Multi-GAT, a hierarchical multi-modal HAR approach that incorporates graphical attention. Their method focuses on learning complementary features from multiple modalities in a hierarchical manner. To disentangle and extract salient modality-specific features facilitating feature interactions, they devised a multi-modal mixture-of-experts model. Furthermore, they introduced a novel message-passing-based graphical attention approach, which captures cross-modal relations to extract complementary multi-modal features. Long et al. [58] introduced keyless attention as a sophisticated and efficient approach to better address the sequential characteristics of data. Additionally, through a comprehensive comparison of various multi-modal fusion techniques, they discovered that multimodal keyless attention fusion achieves the highest success in capturing interactions between different modalities. Song et al. [59] proposed a spatial and temporal attention model to investigate discriminative features in human action recognition and detection using skeleton data. Their approach utilizes RNNs with LSTM units to build the network architecture. Their proposed model learns to selectively focus on discriminative joints within each input frame and assigns varying levels of attention to the outputs of different frames. To ensure effective training for action recognition, they introduced a regularized cross-entropy loss and devised a joint training strategy. Additionally, leveraging temporal attention, they developed a technique to generate temporal proposals for action detection. Cho et al. [60] introduced three variations of the self-attention network (SAN) called SAN-V1, SAN-V2, and SAN-V3. These variants effectively extract high-level semantics by capturing long-range correlations. The authors also incorporated a temporal segment network (TSN) into their SAN variants, leading to notable enhancements in overall performance. Although these attention-driven methods have been widely used for human action-recognition task and have obtained noticeable improvements over handcrafted feature-based methods and other non-attention deep learning methods, they perform well only on clean red, green, and blue (RGB) video data and mostly fail while dealing with noisy color (RGB) video data.

3. Proposed Human Activity-Recognition Framework

This section presents in detail the insights of our proposed DA-CNN+Bi-GRU human action-recognition framework and its core components. For better understanding, the proposed approach is divided into three distinct modules, where each module is separately discussed. The first core component of our method is the newly introduced lightweight CNN architecture having a small number of trainable parameters. The second core component is a dual attention (channel and spatial attention) module, which is used to embed a dual attention mechanism to the CNN module to enable our CNN model to extract salient features from video frames. The last key component of our framework is a bi-directional GRU network for learning long-term encoded patterns of human actions. The conceptual workflow of the DA-CNN+Bi-GRU framework is depicted in Figure 1.

3.1. Overview of Proposed CNN Architecture

Recognizing human actions in video data is indeed a challenging problem, where video data represent complex human actions over a series of frames in the form of different hidden visual contents that include temporal flow of objects in frames, varying textures, object-specific edges, and colors. For better representation and modeling of human actions, these visual contents need to be analyzed effectively via an approach that allows an activity-recognition system to recognize complex human actions or activity in video sequences. To effectively extract the defining visual features of these hidden action contents, CNN-based approaches are widely used to recognize human actions in videos. Although the presented CNN-based approaches have shown remarkable performance, their computational complexity and execution/inference times are very high due to large network architectures. To avoid such high computational complexity and long runtime, we propose a lightweight CNN architecture coupled with channel and spatial attention. The proposed CNN architecture contains a total of eight convolutional layers, where each two consecutive convolutional layers are followed by a max pooling layer and a dual attention block (containing both channel and spatial attention). The first two convolutional layers each apply 16 kernels on input video frames with the kernel size 3 × 3 , whereas the third and fourth convolutional layers each apply 32 kernels on the output of the first dual attention block with the kernel size 3 × 3 .
Similarly, the fifth and sixth convolutional layers each apply 32 kernels on the output of the second dual attention block with the kernel size 3 × 3 . The last pair of the convolutional layers each apply 64 kernels on the output of the third dual attention block with the kernel size 3 × 3 and then forward the estimated feature maps to the last dual attention block. The output of the last dual attention block is processed by a global average pooling layer, the output of which is then flattened by a flatten layer. The output of the flatten layer is fused with a bi-directional GRU network for later long short-term sequence learning. The architectural details of our proposed CNN architecture are listed in Table 1. It is worth noticing that we used at most 64 convolutional kernels per layer and a fixed 3 × 3 kernel size that greatly help to reduce the computational complexity as low as possible with a negligible effect on model performance.

3.2. Dual Attention Module

To exclusively focus on the most salient regions of video frames, we propose an attention-driven CNN architecture to efficiently localize the salient regions and enhance feature representation. The proposed attention mechanism is formed by modifying the convolutional block attention module (CBAM) [61] by replacing the 7 × 7 convolution layer with a 3 × 3 convolution layer, followed by the fusion of the spatial attention module with the intermediate output of the channel attention module through element-wise product operation. A detailed graphical overview of the proposed dual attention block is presented in Figure 2. The fusion of both channel and spatial attentions not only helps to reduce the overall parameters overhead, but also enables the proposed CNN architecture to extract salient features. Therefore, the formation of network layers is constructed in such a way that we place a stacked dual attention module after each two consecutive convolutional layers of our proposed network. The channel attention module estimates the weighted contribution of RGB channels by applying intermediate channel attention A C on the output feature maps F M of the previous convolutional layer to obtain the channel attention A t t C . The computed output from A t t C is then forwarded to the spatial attention module, which localizes promising object-specific regions by applying spatial attention on the computed channel attention feature maps A t t C . Finally, the architecture obtains the refined feature maps F R M by fusing the spatial attention feature maps A t t S with the input feature maps F M via a residual skip connection using element-wise addition operation. Mathematically, A t t C , A t t S , and F R M can be formulated as follows:
A t t C H × W × C = A C ( F M H × W × C ) F M H × W × C ,
A t t S H × W × C = A S ( A t t C H × W × C ) A t t C H × W × C ,
F R M H × W × C = A t t S H × W × C F M H × W × C
Here, H, W, and C denote the height, width, and number of channels of the feature maps, respectively. A C and A S are the intermediate channel attention and the intermediate spatial attention, respectively. F R M is the final refined feature maps obtained by fusing spatial attention and input feature maps F M .

3.2.1. Channel Attention

In pattern-recognition problems, particularly in image/object recognition, each color channel contributes differently based on the appearance of color in an image. During the training process, a CNN model generates feature maps from input image data by extracting deep discriminative features through multiple convolutional layers. Within these feature maps, certain channels have a higher contribution than others in the recognition process, emphasizing their importance in capturing relevant information. Unlike the earlier attention-based approaches that used either global max pooling layer or global average pooling layer, DA-CNN+Bi-GRU uses both global max pooling and global average pooling to extract more effective features. Global max pooling emphasizes highly activated values by selecting the maximum value from the receptive field, whereas global average pooling estimates the equally weighted feature maps for each channel.
The computed feature maps are then forwarded to a shared multilayer perceptron (MLP) containing two fully connected layers, namely f c 1 and f c 2 , having 128 and 512 nodes, respectively. The shared MLP learns the non-linearity between the two fully connected layers using the rectified linear unit (ReLU) activation function, and outputs two individual feature vectors, namely V C m a x 1 × 1 × C and V C a v g 1 × 1 × C , for global max pooling and global average pooling, respectively. The computed feature vectors are then combined via an element-wise addition operation, and then forwarded to a sigmoid activation function σ , which normalizes the feature values to obtain intermediate channel attention features A C 1 × 1 × C . The obtained intermediate channel attention features A C 1 × 1 × C are then fused with the input feature maps F M H × W × C using a residual skip connection by performing an element-wise multiplication operation, which results in the ultimate channel attention feature maps A t t C H × W × C as depicted in Figure 2. Mathematically, the channel attention and its components can be expressed as follows:
V C m a x 1 × 1 × C = f c 2 ( R e L U ( f c 1 ( m a x p o o l ( F M H × W × C ) ) ) ) ,
V C a v g 1 × 1 × C = f c 2 ( R e L U ( f c 1 ( a v g p o o l ( F M H × W × C ) ) ) ) ,
A C 1 × 1 × C = σ ( V C m a x 1 × 1 × C V C a v g 1 × 1 × C ) ,
A t t C H × W × C = A C 1 × 1 × C F M H × W × C ,
Here, V C m a x 1 × 1 × C and V C a v g 1 × 1 × C are the feature vectors obtained from global max pooling and global average pooling operations, respectively. In the above equations, F M H × W × C represents the input feature maps, σ denotes the sigmoid activation function, whereas A t t C H × W × C is the final channel attention output.

3.2.2. Spatial Attention

The spatial attention mechanism focuses on object saliency in the given feature maps by paying more attention to important features across each color channel and localizing salient regions. To highlight the salient object-specific regions in the feature maps, our design exploits inter-spatial features and their relationship among channels, which greatly help to trace the target object in the feature maps. DA-CNN+Bi-GRU computes the relation of inter-spatial features among channels by applying max pooling and average pooling to the input channel attention feature maps to obtain max pooled channel attention A t t C m a x H × W × 1 and average pooled channel attention A t t C a v g H × W × 1 , respectively.
The max pooled channel attention A t t C m a x H × W × 1 and average pooled channel attention A t t C a v g H × W × 1 are concatenated and then forwarded to a single convolutional layer C o n v 3 × 3 , which applies a 3 × 3 convolution kernel on pooled feature maps to form single-channel convoluted feature maps. These convoluted feature maps are then processed by a sigmoid activation function, which normalizes the learned features and produces intermediate spatial attention features A S H × W × 1 . Finally, the obtained intermediate spatial attention features A S H × W × 1 are fused with the input channel attention feature maps A t t C H × W × C using a residual skip connection by performing an element-wise multiplication operation, which results in final spatial attention feature maps A t t S H × W × C , as depicted in Figure 2. Mathematically, spatial attention A t t S W × H × C and its component can be expressed as follows:
A t t C m a x H × W × 1 = m a x p o o l ( A t t C H × W × C ) ,
A t t C a v g H × W × 1 = a v g p o o l ( A t t C H × W × C ) ,
A S H × W × 1 = σ ( C o n v 3 × 3 ( A t t C m a x H × W × 1 A t t C a v g H × W × 1 ) ) ,
A t t S H × W × C = A S H × W × 1 A t t C H × W × C ,
where A t t C m a x H × W × 1 and A t t C a v g H × W × 1 are the global max and average pooled features, respectively. σ is the sigmoid activation function and ⊎ represents the concatenation operation. A t t S H × W × C is the final obtained spatial attention. The representative saliency maps of different human actions generated by our proposed method are depicted in Figure 3.

3.3. Learning Human Action Patterns via Bi-Directional GRU

Videos can be conceptualized as a stack of frames that encapsulate the sequential flow of diverse visual contents within a specific time duration. To understand the visual contents, mainstream computer vision approaches first extract deep discriminative features from the video frames using CNNs and then combine the extracted features in sequential order to maintain the semantic flow of the video. Second, the feature-encoded videos are then processed via RNNs to learn the representation of visual contents from hidden sequential patterns. Specifically, for human activity-recognition problems, two special variants of RNNs are actively used by researchers that include LSTMs and GRUs. The LSTM unit is comprised of different gates including input, output, forget gates, and other memory components, whereas the GRU contains an update gate, a reset gate, and an activation function. The LSTM is comparatively more complex than the GRU in terms of the number and formation of gates which leads to relatively higher computational complexity requiring more computational resources. Therefore, in this paper, we propose to use GRU with bi-directional flow of learning strategy, which effectively learns from the encoded hidden sequential pattern.
The bi-directional GRU consists of two layers, namely forward and backward layer, where both layers process the same sequence in different sequential order. The forward layer reads the input sequence from left to right, that is, from X t 1 to X n where n is the length of sequence. On the other hand, the backward layer reads the input sequence in reverse order from right to left, that is, from X t + n to X t 1 as shown in Figure 4. Both forward and backward GRU layers consist of GRU cells, where each cell consists of two gates, namely a reset r and an update gate μ , with two activation functions that include sigmoid and tanh. The reset gate decides whether the GRU needs to forget or retain the portion of information based on its values (between 0 and 1). When the output value of the reset gate is near 0, the reset gate forgets the information from the previous portion of the sequence, whereas if the reset gate value is near 1, the reset gate retains the previous portion of the sequence. The update gate decides the amount of information from the previous hidden state to be retained to the current hidden state based on its values (between 0 and 1). When the value of the update gate is near 0, the updated gate simply forgets the portion of information from the previous hidden state and retains the portion of information from the previous hidden state to the current hidden state when the value is close to 1. Mathematically, the operation of these gates can be expressed as follows:
r t = σ ( w r · x t + u r · h t 1 ) ,
μ t = σ ( w μ · x t + u μ · h t 1 ) ,
h ˜ t = tanh ( w · x t + r t · u · h t 1 ) ,
h t = ( 1 μ t ) · h t 1 + μ t · h ˜ t ,
y t = σ ( w o · h t ) ,
where r t and μ t represent the reset and update gates, respectively, having values between 0 and 1. In the above equations, w and u are the weight variables, x t is the input to the GRU layer, w o is the weight variable between input and output layer, y t represents the output layer node at time step t. h ˜ t is the candidate hidden state of the current node, h t is the current hidden state, and h t 1 is the hidden state of the previous node.

4. Experimental Results and Discussion

In this section, we present a detailed experimental evaluation of our proposed DA-CNN+Bi-GRU human activity-recognition framework. We evaluate the effectiveness of our proposed framework by analyzing the performance with and without the key components (channel attention, spatial attention, bi-directional GRU) of our framework. First, we describe the implementation details and performance evaluation metrics that we used in this research. Next, we briefly discuss the datasets we used for benchmarking experiments. We then compare the DA-CNN+Bi-GRU framework with state-of-the-art human action-recognition methods across each tested dataset. Finally, we present the human action-recognition visualization and then conduct runtime analysis of our proposed approach for real-time human activity recognition.

4.1. Implementation Details

The DA-CNN+Bi-GRU framework is implemented using a well-known deep learning framework called TensorFlow version 2.0 in Python language 3 on a computing system with an Intel Xeon (R) processor with processor frequency 3.50 GHz and 32 GB of dedicated main memory. The computing system is also equipped with an NVIDIA GeForce GTX 1080 graphics processing unit (GPU) having a graphics random-access memory of 8 GB. For training and validation, we divided the datasets into a ratio of 70% and 30%, where for training we used 70% of the data and the remaining 30% of the data were used for validation. The training process is run for 300 epochs and the weights are initialized with a random weight initializer, whereas the batch size is set to 16. To adjust weight values during training, we used the Adam optimizer with static learning rate of 0.0001. The DA-CNN+Bi-GRU network utilizes categorical cross-entropy loss, which controls the weight adjustment based on network prediction during training. For sequence learning, we used a sequence length of 16 frames without overlapping for both forward and backward pass of bi-directional GRU, where we used three bi-directional GRU layers with 32 GRUs per layer. Moreover, we used two different performance evaluation metrics to assess the overall performance of our proposed method. The first metric is the accuracy metric, which is used to evaluate the activity-recognition performance of our framework and other contemporary methods. The second metric is frames per second (FPS) or alternatively seconds per frame (SFP), which measures the runtime of our proposed framework and other contemporary methods.

4.2. Datasets

To verify the effectiveness of the DA-CNN+Bi-GRU framework, we conducted extensive experiments on three challenging human action datasets that include YouTube action, UCF50, HMDB51, UCF101, and Kinetics-600 datasets. Each dataset consists of multiple action videos having varying duration, different viewpoints, and FPS. These datasets are discussed in detail in the following subsections.

4.2.1. YouTube Action Dataset

The YouTube action dataset [62] is a commonly used action-recognition dataset containing diverse sports and other action video clips collected from YouTube. The collected videos clips are very challenging due to variations in viewpoint, camera motion, cluttered background, and varying poses and appearances of objects in the scene. The dataset contains 1640 video clips categorized into 11 action categories, where the duration of videos ranges between 2 and 5 s having a frame rate of 29 FPS and a resolution of 320 × 240 . The collected action clips in all action categories are grouped into 25 distinct groups containing four or more video clips, where each video clip in the same group shares common visual features, such as background, viewpoint, and the person or actor.

4.2.2. UCF50 Dataset

The UCF50 dataset [63] is one of the challenging large-scale human activity-recognition datasets, containing videos of diverse human actions captured with varying viewpoints, camera motions, object poses and appearances, and background clutter. The dataset contains a total of 6676 video clips categorized into 50 different classes, where the duration of video clips ranges between 2 and 3 s with a frame rate of 25 FPS and a resolution of 320 × 240 . The video clips in all 50 categories are further grouped into 25 groups, where each group comprises at least four video clips, where a video clip in a single group shares common features of actions, such as the same person performing an action, the same viewpoint, and the same background.

4.2.3. HMDB51 Dataset

HMDB51 [64] is one of the challenging datasets commonly used for human action recognition in videos. The videos in this dataset are collected from difference sources including movies, public databases, YouTube, and Google videos. The dataset comprises a total of 6849 action video clips categorized into 51 classes, where each class contains at least 101 video clips having a duration of 2 to 3 s with a frame rate of 30 FPS and a resolution of 320 × 240 . The collected action video clips can be generally categorized into five different types of actions that include facial actions, facial actions with object manipulation, general body movements, body movements and interaction with objects, and body movements while interacting with humans.

4.2.4. UCF101 Dataset

According to the literature based on human action/activity recognition, UCF101 [65] is a very challenging dataset comprising videos which resemble real-world activities. The dataset consists of 13,320 videos collected from YouTube, which are categorised into 101 action classes, where in each class there are 100 to 200 video clips of human actions performed by different subjects. The duration of each clip is between 2 and 3 s with a frame rate of 25 FPS and a frame resolution of 320 × 240 . The collected video clips are retrieved based on five major human activities that include human interaction with objects, human interaction with humans, human body motion only, playing of musical instruments, and humans performing sports activities.

4.2.5. Kinetics-600 Dataset

Kinetics-600 [66] is an extremely large human action video dataset, comprising 480 K video clips (∼10 s) of different human actions that are categorized into 600 actions classes. The video clips in the dataset are collected from YouTube videos which are then labeled based on the action in video clips, where each video clip has a variable resolution, field of view, and frame rate. The dataset has three distinct sets, namely train, validation, and test. The train set contains 500∼950 videos per class, whereas in the validation set, each class has 45∼50 videos. The test set and validation set have the same number of videos; however, the test set is not labeled.

4.3. Assessment of Our Framework with Baseline Methods

This research is built up on the exploration of various possible solutions for vision-based human action recognition, where we developed several spatial–temporal methods, assessed their performances, and developed our final proposed method. To obtain the optimal approach, we explored different spatial–temporal solutions and successively developed four different baseline methods that include CNN+LSTM, CNN+Bi-LSTM, CNN+GRU, and CNN+Bi-GRU, and we analyzed their performances in terms of model precision. To obtain a fair comparison, we trained each baseline method on five different datasets (i.e., YouTube action, UCF50, HMDB51, UCF101, and Kinetics-600 datasets). These datasets are then used for training the DA-CNN+Bi-GRU framework. The network settings of these baseline methods are listed in detail in Table 2, where it can be perceived that CNN+LSTM and CNN+GRU methods use a total of 11 spatial–temporal layers including 8 convolutional and 3 temporal layers. Similarly, CNN+Bi-LSTM and CNN+Bi-GRU methods use a total of 14 layers, including 8 convolutional and 6 temporal layers (with 3 forward and 3 backward pass layers). Finally, the proposed framework (DA-CNN+Bi-GRU) has a total of 18 layers comprising 12 convolutional layers (8 convolutional and 4 attentional) and 6 temporal layers (3 forward and 3 backward pass layers).
The training performance (in terms of accuracy) of each baseline method along with our proposed method is depicted in Figure 5. It can be seen from Figure 5 that DA-CNN+Bi-GRU performs better than other baseline methods in terms of accuracy. For instance, in Figure 5a for the YouTube action dataset, our method achieves the best accuracy score throughout 300 epochs. In Figure 5b for the HMDB51 dataset, our method (DA-CNN+Bi-GRU) starts as the second-best method in early training epochs where CNN+Bi-GRU dominates; however, after 20 epochs, DA-CNN+Bi-GRU attains the best accuracy as compared to the other baseline methods and remains the best till the end of the training. Similarly, in Figure 5c for the UCF50 dataset, our method (DA-CNN+Bi-GRU) does not perform the best in the first 35 epochs, where CNN+Bi-GRU dominates; however, after 35 epochs, the DA-CNN+Bi-GRU starts improving and finally trains with the best accuracy at the 300th epoch. In Figure 5d for the UCF101 dataset, the proposed DA-CNN+Bi-GRU starts with the best training accuracy in the very early epochs and shows the best performance throughout the training phase (for 300 epochs) and finishes with the best training accuracy. Finally, in Figure 5e for the Kinetics-600 dataset, the proposed method obtains the best training accuracy throughout the training phase followed by the CNN+Bi-GRU method with the second best performance in terms of training accuracy.
We also demonstrate the performance of our proposed DA-CNN+Bi-GRU method on YouTube action, HMDB51, UCF50, UCF101, and Kinetics-600 datasets using confusion matrix and category-wise accuracy metrics. The obtained results for the confusion matrix and category-wise accuracy metrics are depicted in Figure 6 and Figure 7, respectively. The obtained performances of these baseline methods along with our proposed method across five benchmark datasets are presented in Table 3. From Table 3, it can be noticed that DA-CNN+Bi-GRU dominates all the baseline methods across each dataset. For instance, the proposed framework attains the best accuracy score of 98.0% over the YouTube action dataset as compared to all the baseline methods, whereas CNN (spatial attention only) + Bi-GRU obtains the second-best accuracy score of 95.6%. Similarly, on the UCF50 dataset, the proposed framework obtains the highest accuracy score of 98.5%, whereas the runner-up is CNN (spatial attention only) + Bi-GRU with an accuracy of 95.7%. For the HMDB51 dataset, it can be seen that our proposed method dominates all the baseline methods by achieving the best accuracy score of 79.3%, whereas CNN (spatial attention only) + Bi-GRU is the runner-up method, attaining the second-best accuracy score of 74.5%. For the HMDB51 dataset, the proposed framework attains the highest accuracy score of 79.3%, whereas CNN (spatial attention only) + Bi-GRU is the runner-up method, obtaining the second-best accuracy of 74.5%. For the UCF101 dataset, the proposed DA-CNN+Bi-GRU outperforms all the baseline methods by obtaining an accuracy of 97.6%, whereas the runner-up is the baseline method CNN (spatial attention only) + Bi-GRU which obtains an accuracy of 95.8%. Finally, for the Kinetics-600 dataset, the proposed DA-CNN+Bi-GRU achieves the best accuracy of 86.7% amongst all the baseline approaches, whereas the CNN (spatial attention only) + Bi-GRU method attains the second best accuracy of 85.6%. The best and the runner-up results are indicated in bold and italics, respectively, in Table 3.

4.4. Comparison with State-of-the-Art Methods

To show the effectiveness of the DA-CNN+Bi-GRU framework for the human activity-recognition task, we conducted an extensive comparative analysis of our method with the state-of-the-art methods in terms of overall accuracy. The quantitative comparisons of our method with the state-of-the-art methods for YouTube action, UCF50, HMDB51, UCF101, and Kinetics-600 datasets are listed in Table 4, Table 5, Table 6, Table 7 and Table 8, respectively. The best results in these tables are represented in bold, whereas the runner-up results are highlighted in italics. Considering the presented results, it can be noticed that DA-CNN+Bi-GRU outperforms state-of-the-art methods on UCF50, HMDB51, and UCF101 datasets, whereas it attains runner-up performance on the YouTube action and Kinetics-600 datasets. For the YouTube action dataset, the STDN [67] has the best performance with an accuracy of 98.2%, whereas DA-CNN+Bi-GRU attains the runner-up performance by obtaining an accuracy of 98.0%, which is within 0.2% accuracy of the best-performing STDN [67]. Thus, for most practical purposes, the DA-CNN+Bi-GRU framework attains performance comparable to STDAN [67]. Other methods compared include multi-task hierarchical clustering [68], BT-LSTM [69], deep autoencoder [70], two-stream attention LSTM [71], weighted entropy-variance-based feature selection [72], dilated CNN+BiLSTM+RB [73], DS-GRU [52], and local-global features + QSVM [74], which obtain 89.7%, 85.3%, 96.2%, 96.9%, 94.5%, 89.0%, 97.1%, and 82.6% accuracies, respectively.
For the UCF50 dataset, DA-CNN+Bi-GRU dominates the state-of-the-art methods by obtaining the best accuracy of 98.5%, whereas (LD-BF) + (LD-DF) [77] obtains the second-best accuracy of 96.7%. Local-global features + QSVM [74] achieves the lowest accuracy of 69.4%, whereas the rest of the methods including multi-task hierarchical clustering [68], deep autoencoder [70], ensemble model with swarm-based optimization [75], DS-GRU [52], and ViT+LSTM [76] obtain 93.2%, 96.4%, 92.2%, 95.2%, and 96.1% accuracies, respectively. For the HMDB51 dataset comprising challenging action videos, our proposed method achieves the best results by obtaining an accuracy of 79.3%, whereas the runner-up method is evidently deep learning [78] which attains an accuracy of 77.0%. The multi-task hierarchical clustering method [68] achieves an accuracy of 51.4%, which is the lowest among all the comparative methods on the HMDB51 dataset. The rest of the comparative methods including STPP+LSTM [79], optical flow + multi-layer LSTM [47], TSN [80], IP-LSTM [81], deep autoencoder [70], TS-LSTM + temporal-inception [82], HATNet [83], correlational CNN+LSTM [84], STDAN [67], DB-LSTM+SSPF [48], DS-GRU [52], TCLC [85], ViT+LSTM [76], semi-supervised temporal gradient learning [86], AdaptFormer [87], SVT (Linear) [88], SVT (Fine-tune) [88], SVFormer-S [89], and SVFormer-B [89] obtain accuracies of 70.5%, 72.2%, 70.7%, 58.6%, 70.3%, 69.0%, 74.8%, 66.2%, 56.5%, 75.1%, 72.3%, 71.5%, 73.7%, 75.9%, 55.6%, 57.8%, 67.2%, 59.7%, and 68.2%, respectively.
For the UCF101 dataset, the proposed DA-CNN+Bi-GRU outperforms all the comparative methods by achieving the best accuracy of 97.6% followed by the runner-up method RTS [90], which attains an accuracy of 96.4%. The multi-task hierarchical clustering method [68] attains the lowest accuracy of 76.3% amongst all the comparative methods, followed by SVFormer-S [89], which achieves the second-lowest accuracy of 79.1% amongst the other considered comparative methods on the UCF101 dataset. The rest of the comparative methods including saliency-aware 3DCNN with LSTM [91], spatiotemporal multiplier networks [92], long-term temporal convolutions [39], OFF [93], TVNet [94], attention cluster [95], CNN with Bi-LSTM [96], Videolstm [97], two-stream convnets [98], mixed 3D-2D convolutional tube [99], TS-LSTM+temporal-inception [82], TSN+TSM [100], STM [101], correlational CNN+LSTM [84], SVT (Linear) [88], SVT (Fine-tune) [88], ConvNet Transformer [102], and SVFormer-B [89] achieve accuracies of 84.0%, 87.0%, 82.4%, 96.0%, 95.4%, 94.6%, 92.8%, 89.2%, 84.9%, 88.9%, 91.1%, 94.3%, 96.2%, 92.8%, 90.8%, 93.7%, 86.1%, and 86.7%, respectively. Finally, for the Kinetics-600 dataset, the MTV-H [103] achieves the best accuracy of 89.6%, followed by the proposed DA-CNN+Bi-GRU method which attains an accuracy of 86.7%, whereas the MTV-B [103] (variant of MTV-H) achieves an accuracy of 84.0%. The GCF-Net [104] and global and local-aware attention [105] methods attain the lowest accuracies of 70.0% and 70.0%, respectively. The rest of the comparative methods including SlowFast [106], Stnet [107], LGD-3D [108], D3D+S3D-G [109], MoviNet [110], MM-ViT [111], Swin-B [112], and Swin-L [112] achieve accuracies of 81.8%, 76.3%, 82.7%, 79.1%, 83.5%, 83.8%, and 85.9%, respectively. Considering the overall comparative analysis, DA-CNN+Bi-GRU obtains performance comparable to the best-performing method on the YouTube action and Kinetics-600 datasets, and greatly dominates the state-of-the-art comparative methods on UCF50, HMDB51, and UCF101 datasets, thus demonstrating the superiority of our proposed method over the existing action-recognition methods.
Table 6. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for HMDB51 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 6. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for HMDB51 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodYearAccuracy (%)
Multi-task hierarchical clustering [68]201751.4
STPP+LSTM [79]201770.5
Optical flow + multi-layer LSTM [47]201872.2
TSN [80]201870.7
IP-LSTM [81]201958.6
Deep autoencoder [70]201970.3
TS-LSTM + temporal-inception [82]201969.0
HATNet [83]201974.8
Correlational CNN + LSTM [84]202066.2
STDAN [67]202056.5
DB-LSTM+SSPF [48]202175.1
DS-GRU [52]202172.3
TCLC [85]202171.5
Evidential deep learning [78]202177.0
ViT+LSTM [76] 2021 73.7
Semi-supervised temporal gradient learning [86]202275.9
AdaptFormer [87] 2022 55.6
SVT (Linear) [88] 2022 57.8
SVT (Fine-tune) [88] 202267.2
SVFormer-S [89] 202359.7
SVFormer-B [89] 202368.2
DA-CNN+Bi-GRU (Proposed)202379.3
Table 7. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for UCF101 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 7. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for UCF101 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodYearAccuracy (%)
Multi-task hierarchical clustering [68]201776.3
Saliency-aware 3DCNN with LSTM [91]201784.0
Spatiotemporal multiplier networks [92]201787.0
Long-term temporal convolutions [39]201782.4
RTS [90]201896.4
OFF [93]201896.0
TVNet [94]201895.4
Attention cluster [95]201894.6
CNN with Bi-LSTM [96]201892.8
Videolstm [97]201889.2
Two stream convnets [98]201884.9
Mixed 3D-2D convolutional tube [99]201888.9
TS-LSTM + temporal-inception [82]201991.1
TSN+TSM [100]201994.3
STM [101]201996.2
Correlational CNN + LSTM [84]202092.8
SVT (Linear) [88]202290.8
SVT (Fine-tune) [88]202293.7
ConvNet Transformer [102]202386.1
SVFormer-S [89]202379.1
SVFormer-B [89]202386.7
DA-CNN+Bi-GRU (Proposed)202397.6
Table 8. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for Kinetics-600 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 8. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for Kinetics-600 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodYearAccuracy (%)
SlowFast [106]201981.8
Stnet [107]201976.3
LGD-3D [108]201982.7
GCF-Net [104]202070.0
D3D+S3D-G [109]202079.1
MoviNet [110]202183.5
Global and local-aware attention [105]202170.0
MM-ViT [111]202283.5
Swin-B [112]202283.8
Swin-L [112]202285.9
MTV-B [103]202284.0
MTV-H [103]202289.6
DA-CNN+Bi-GRU (Proposed)202386.7

4.5. Action-Recognition Visualization

To validate the recognition efficiency of DA-CNN+Bi-GRU, we tested DA-CNN+Bi-GRU on 15% of test videos taken from each dataset (including YouTube action, UCF50, HMDB51, UCF101, and Kinetics-600 datasets). The prepared test sets are validated for the action-recognition task using our proposed framework and the visual results from the test experiments are depicted in Figure 8. In Figure 8, the representative frames of the predicted action clips are presented along with their ground truths, model predicted actions, and confidence scores over the probability prediction bar graphs for better understanding of readers. It can be perceived from the presented visual results that the DA-CNN+Bi-GRU framework predicts most of the actions including brush hair, volleyball spiking, basketball, climb, fall floor, bench press, horse race, billiards, diving, baseball pitch, and hula hoop with 0.99% probability or 99% confidence. However, for some action classes, such as clap, fencing, golf swing, and high jump, the DA-CNN+Bi-GRU framework also generates non-zero probabilities for wrong action classes; however, these probabilities for wrong action classes are still very low and thus do not affect the prediction of actual action class. Hence, the obtained qualitative visual results verify the effectiveness of our proposed framework for practical use in different vision-based human action-recognition and -monitoring environments.

4.6. Runtime Analysis

To analyze the effectiveness and feasibility of the DA-CNN+Bi-GRU framework for practical applications in real-time environments, we estimated the runtime of our method for action-recognition tasks in terms of SPF and FPS with and without using GPU resources. The obtained runtime results are then compared with the state-of-the-art methods. Table 9 presents and compares the runtime of our proposed framework with the running times of the contemporary action-recognition methods. Results in Table 9 demonstrate that the DA-CNN+Bi-GRU framework outperforms the state-of-the-art methods when executing on both GPU and central processing unit (CPU) platforms. Results indicate that our proposed framework attains 0.0036 SPF and 300 FPS while running on GPU, whereas it attains 0.0049 SPF and 250 FPS while running on CPU. Results further show that the second-best execution time results on GPU are achieved by [93], which are 0.0048 SPF and 206 FPS. In Table 9, the best runtime results are indicated with bold text and the runner-up results are indicated with italics. Experimental results indicate that for the SPF metric, the DA-CNN+Bi-GRU framework can provide an improvement of up to 26.1× when running on GPU and an improvement of 87.76× when running on CPU as compared to other contemporary activity-recognition methods. Experimental results further reveal that for the FPS metric, the DA-CNN+Bi-GRU framework can provide an improvement of up to 28.3× when running on GPU and an improvement of up to 166.6× when running on CPU as compared to other contemporary activity-recognition methods.
We also scaled the runtime inference results of the state-of-the-art human action-recognition methods in Table 9 to the hardware specifications used in our framework (i.e., 3.5 GHz CPU and 1607 MHz GPU) to provide a fair comparison of the inference speed. The scaled runtime inference results of the state-of-the-art human action-recognition methods are presented in Table 10. Although scaling does not provide 100% accuracy for processor/GPU runtime because of different instruction set architectures and memory subsystems utilized by different processor/GPU architectures, scaling provides plausible estimates and facilitates relative comparisons [113,114]. From the scaled results in Table 10, it can be seen that the STTP+LSTM [79] method has the best SPF and FPS values of 0.0023 and 423.58, respectively, for the GPU inference. The OFF [93] method has the runner up SPF and FPS of 0.0029 and 331.04, respectively, for the GPU inference, followed by our proposed method which has SPF and FPS of 0.0036 and 300, respectively. Our proposed method thus obtains the third-best SPF and FPS values for the GPU inference. On the other hand, for inference on the CPU, our proposed method delivers the best SPF and FPS values of 0.0049 and 250, respectively, followed by Optical flow + multi-layer LSTM [47], which attains the runner up SPF and FPS values of 0.17 and 3.71, respectively. We note that our method, however, provides a better accuracy on human activity-recognition tasks than the STTP+LSTM [79] method and the OFF [93] method. For the scaled experimental results, it can be observed that for the FPS metric, the DA-CNN+Bi-GRU framework can provide an improvement of 2.82×, on average, when running on GPU and an improvement of up to 94.34×, on average, when running on CPU as compared to other contemporary activity-recognition methods. It is also worth mentioning here that the storage requirement of the DA-CNN+Bi-GRU framework is just 5.4 MB, and thus our framework can be run on resource-constrained IoT and edge devices with very limited memory including today’s smart cameras, Arduino, and Raspberry pi. These runtime and storage requirement results demonstrate that the proposed framework is a suitable candidate for deployment on resource-constrained IoT and edge devices as the proposed framework exhibits better accuracy, lower execution time, and low storage requirements as compared to contemporary activity-recognition methods.

5. Conclusions and Future Research Directions

In this work, we proposed a cascaded spatial–temporal discriminative feature-learning framework for human activity recognition in video streams. The proposed method encapsulates the attentional (channel and spatial attention) CNN architecture and bi-directional GRU network as a unified framework for single instance training and efficient spatial temporal modeling of human actions. The attentional CNN architecture comprises channel and spatial attentions, which help retrieve the prominent discriminative features from the object-specific regions, and thus generate high quality saliency-aware feature maps. The bi-directional GRU learns the temporal modeling of long-term human action sequences using two-way gradient learning (i.e., forward and backward pass), which allows the DA-CNN+Bi-GRU framework to utilize the learned knowledge not only from the previous frames but also from the upcoming/next frames. Such bi-directional modeling of human actions greatly helps our method to improve the learning ability while training and the prediction precision while inferencing. To evaluate the efficiency of DA-CNN+Bi-GRU, we conducted extensive experiments on five publicly available human action benchmark datasets. The obtained experimental results are compared with the state-of-the-art methods on five benchmark human action-recognition datasets, including YouTube action, UCF50, HMDB51, UCF101, and Kinetics-600 datasets. Experimental results verify the effectiveness of our method in terms of both model robustness and computational efficiency. Further, we analyzed the runtime performance of our proposed framework in terms of seconds per frame (SPF) and frames per second (FPS) for both CPU and GPU execution environments. The obtained runtime assessment results reveal that our proposed framework can attain an improvement of up to 88× for the SPF metric and up to 167× for the FPS metric as compared to other contemporary action-recognition methods. Additionally, our proposed framework requires a storage of only 5.3 MB, which makes it feasible for deployment on devices with limited memory. Thus, the overall efficiency of our framework in terms of recognition performance (accuracy), low execution time, and low storage requirements, makes DA-CNN+Bi-GRU a strong candidate for real-time IoT and edge applications.
Currently, the DA-CNN+Bi-GRU method only uses the spatial attention (channel and spatial attention) mechanism, which is indeed very effective. However, in the future, we plan to use the temporal attention mechanism together with spatial attention because such hybrid attention has a great potential to improve the human activity-recognition performance.

Author Contributions

Conceptualization, H.U. and A.M.; methodology, H.U. and A.M.; software, H.U.; validation, H.U.; formal analysis, H.U. and A.M.; investigation, A.M.; resources, A.M.; data curation, H.U.; writing—original draft preparation, H.U. and A.M.; writing—review and editing, H.U. and A.M.; visualization, H.U. and A.M.; supervision, A.M.; project administration, A.M.; funding acquisition, A.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Air Force Office of Scientific Research (AFOSR) Contract Number FA9550-22-1-0040. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the Air Force, the Air Force Research Laboratory (AFRL), and/or AFOSR.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analysed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

The authors would like to acknowledge Erik Blasch from the Air Force Research Laboratory (AFRL) for his guidance, support, and feedback on this research.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

CNN Convolutional neural network
DA-CNN Dual attention convolutional neural network
CBAM Convolutional block attention module
IoT Internet of things
RNN Recurrent neural network
GRU Gated recurrent unit
Bi-GRU Bi-directional gated recurrent unit
SPF Seconds per frame
FPS Frames per second

References

  1. Munir, A.; Blasch, E.; Kwon, J.; Kong, J.; Aved, A. Artificial Intelligence and Data Fusion at the Edge. IEEE Aerosp. Electron. Syst. Mag. 2021, 36, 62–78. [Google Scholar] [CrossRef]
  2. Munir, A.; Kwon, J.; Lee, J.H.; Kong, J.; Blasch, E.; Aved, A.; Muhammad, K. FogSurv: A Fog-Assisted Architecture for Urban Surveillance Using Artificial Intelligence and Data Fusion. IEEE Access 2021, 9, 111938–111959. [Google Scholar] [CrossRef]
  3. Huang, C.; Wu, Z.; Wen, J.; Xu, Y.; Jiang, Q.; Wang, Y. Abnormal Event Detection Using Deep Contrastive Learning for Intelligent Video Surveillance System. IEEE Trans. Ind. Inform. 2021, 18, 5171–5179. [Google Scholar] [CrossRef]
  4. Sahu, A.; Chowdhury, A.S. Together Recognizing, Localizing and Summarizing Actions in Egocentric Videos. IEEE Trans. Image Process. 2021, 30, 4330–4340. [Google Scholar] [CrossRef] [PubMed]
  5. Qi, M.; Qin, J.; Yang, Y.; Wang, Y.; Luo, J. Semantics-Sware Spatial–Temporal Binaries for Cross-Modal Video Retrieval. IEEE Trans. Image Process. 2021, 30, 2989–3004. [Google Scholar] [CrossRef]
  6. Muhammad, K.; Ullah, H.; Obaidat, M.S.; Ullah, A.; Munir, A.; Sajjad, M.; de Albuquerque, V.H.C. AI-Driven Salient Soccer Events Recognition Framework for Next Generation IoT-Enabled Environments. IEEE Internet Things J. 2021, 2202–2214. [Google Scholar] [CrossRef]
  7. Ng, W.; Zhang, M.; Wang, T. Multi-Localized Sensitive Autoencoder-Attention-Lstm for Skeleton-Based Action Recognition. IEEE Trans. Multimed. 2021, 24, 1678–1690. [Google Scholar] [CrossRef]
  8. Asghari, P.; Soleimani, E.; Nazerfard, E. Online Human Activity Recognition Employing Hierarchical Hidden Markov Models. J. Ambient Intell. Humaniz. Comput. 2020, 11, 1141–1152. [Google Scholar] [CrossRef] [Green Version]
  9. Ehatisham-Ul-Haq, M.; Javed, A.; Azam, M.A.; Malik, H.M.; Irtaza, A.; Lee, I.H.; Mahmood, M.T. Robust Human Activity Recognition Using Multimodal Feature-Level Fusion. IEEE Access 2019, 7, 60736–60751. [Google Scholar] [CrossRef]
  10. Naveed, H.; Khan, G.; Khan, A.U.; Siddiqi, A.; Khan, M.U.G. Human Activity Recognition Using Mixture of Heterogeneous Features and Sequential Minimal Optimization. Int. J. Mach. Learn. Cybern. 2019, 10, 2329–2340. [Google Scholar] [CrossRef]
  11. Franco, A.; Magnani, A.; Maio, D. A Multimodal Approach for Human Activity Recognition Based on Skeleton and RGB Data. Pattern Recognit. Lett. 2020, 131, 293–299. [Google Scholar] [CrossRef]
  12. Elmadany, N.E.D.; He, Y.; Guan, L. Information Fusion for Human Action Recognition via Biset/Multiset Globality Locality Preserving Canonical Correlation Analysis. IEEE Trans. Image Process. 2018, 27, 5275–5287. [Google Scholar] [CrossRef]
  13. Dileep, D.; Sreeni, K. Anomalous Event Detection in Crowd Scenes using Histogram of Optical Flow and Entropy. In Proceedings of the 2021 Fourth International Conference on Microelectronics, Signals & Systems (ICMSS), Kollam, India, 18–19 November 2021; pp. 1–6. [Google Scholar]
  14. Yenduri, S.; Perveen, N.; Chalavadi, V. Fine-Grained Action Recognition Using Dynamic Kernels. Pattern Recognit. 2022, 122, 108282. [Google Scholar] [CrossRef]
  15. Luvizon, D.C.; Picard, D.; Tabia, H. Multi-Task Deep Learning for Real-Time 3D Human Pose Estimation and Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2752–2764. [Google Scholar] [CrossRef] [Green Version]
  16. Li, J.; Liu, X.; Zhang, W.; Zhang, M.; Song, J.; Sebe, N. Spatio-Temporal Attention Networks for Action Recognition and Detection. IEEE Trans. Multimed. 2020, 22, 2990–3001. [Google Scholar] [CrossRef]
  17. Ghose, S.; Prevost, J.J. Autofoley: Artificial Synthesis of Synchronized Sound Tracks for Silent Videos with Deep Learning. IEEE Trans. Multimed. 2020, 23, 1895–1907. [Google Scholar] [CrossRef]
  18. Lu, L.; Lu, Y.; Yu, R.; Di, H.; Zhang, L.; Wang, S. GAIM: Graph Attention Interaction Model for Collective Activity Recognition. IEEE Trans. Multimed. 2019, 22, 524–539. [Google Scholar] [CrossRef]
  19. Liu, K.; Gao, L.; Khan, N.M.; Qi, L.; Guan, L. A Multi-Stream Graph Convolutional Networks-Hidden Conditional Random Field Model for Skeleton-Based Action Recognition. IEEE Trans. Multimed. 2020, 23, 64–76. [Google Scholar] [CrossRef]
  20. Hu, P.; Ho, E.S.l.; Munteanu, A. 3DBodyNet: Fast Reconstruction of 3D Animatable Human Body Shape From a Single Commodity Depth Camera. IEEE Trans. Multimed. 2021, 24, 2139–2149. [Google Scholar] [CrossRef]
  21. Yan, C.; Hao, Y.; Li, L.; Yin, J.; Liu, A.; Mao, Z.; Chen, Z.; Gao, X. Task-Adaptive Attention for Image Captioning. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 43–51. [Google Scholar] [CrossRef]
  22. Xia, W.; Yang, Y.; Xue, J.H.; Wu, B. Tedigan: Text-Guided Diverse Face Image Generation and Manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 November 2021. [Google Scholar]
  23. Pareek, P.; Thakkar, A. A Survey on Video-Based Human Action Recognition: Recent Updates, Datasets, Challenges, and Applications. Artif. Intell. Rev. 2021, 54, 2259–2322. [Google Scholar] [CrossRef]
  24. Kong, Y.; Fu, Y. Human Action Recognition and Prediction: A Survey. Int. J. Comput. Vis. 2022, 130, 1366–1401. [Google Scholar] [CrossRef]
  25. Scovanner, P.; Ali, S.; Shah, M. A 3-Dimensional Sift Descriptor and Its Application to Action Recognition. In Proceedings of the 15th ACM International Conference on Multimedia, Augsburg, Germany, 25–29 September 2007; pp. 357–360. [Google Scholar]
  26. Laptev, I.; Marszalek, M.; Schmid, C.; Rozenfeld, B. Learning Realistic Human Actions from Movies. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  27. Ryoo, M.S.; Matthies, L. First-Person Activity Recognition: Feature, Temporal Structure, and Prediction. Int. J. Comput. Vis. 2016, 119, 307–328. [Google Scholar] [CrossRef]
  28. Ullah, H.; Muhammad, K.; Irfan, M.; Anwar, S.; Sajjad, M.; Imran, A.S.; de Albuquerque, V.H.C. Light-DehazeNet: A Novel Lightweight CNN Architecture for Single Image Dehazing. IEEE Trans. Image Process. 2021, 30, 8968–8982. [Google Scholar] [CrossRef] [PubMed]
  29. Chen, T.; Yao, Y.; Zhang, L.; Wang, Q.; Xie, G.; Shen, F. Saliency Guided Inter-and Intra-Class Relation Constraints for Weakly Supervised Semantic Segmentation. IEEE Trans. Multimed. 2022, 25, 1727–1737. [Google Scholar] [CrossRef]
  30. Aafaq, N.; Mian, A.S.; Akhtar, N.; Liu, W.; Shah, M. Dense Video Captioning with Early Linguistic Information Fusion. IEEE Trans. Multimed. 2022, 25, 2309–2322. [Google Scholar] [CrossRef]
  31. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei-Fei, L. Large-Scale Video Classification with Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 27 June–2 July 2014; pp. 1725–1732. [Google Scholar]
  32. Simonyan, K.; Zisserman, A. Two-Stream Convolutional Networks for Action Recognition in Videos. Adv. Neural Inf. Process. Syst. 2014, 27. [Google Scholar]
  33. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y. Towards Good Practices for Very Deep Two-Stream Convnets. arXiv 2015, arXiv:1507.02159. [Google Scholar]
  34. Yue-Hei Ng, J.; Hausknecht, M.; Vijayanarasimhan, S.; Vinyals, O.; Monga, R.; Toderici, G. Beyond Short Snippets: Deep Networks for Video Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 4694–4702. [Google Scholar]
  35. Wu, Z.; Wang, X.; Jiang, Y.G.; Ye, H.; Xue, X. Modeling Spatial–Temporal Clues in a Hybrid Deep Learning Framework for Video Classification. In Proceedings of the 23rd ACM International Conference on Multimedia, Brisbane, Australia, 26 October 2015; pp. 461–470. [Google Scholar]
  36. Wang, X.; Farhadi, A.; Gupta, A. Actions Transformations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2658–2667. [Google Scholar]
  37. Feichtenhofer, C.; Pinz, A.; Zisserman, A. Convolutional Two-Stream Network Fusion for Video Action Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1933–1941. [Google Scholar]
  38. Tran, D.; Bourdev, L.; Fergus, R.; Torresani, L.; Paluri, M. Learning Spatiotemporal Features with 3D Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 4489–4497. [Google Scholar]
  39. Varol, G.; Laptev, I.; Schmid, C. Long-Term Temporal Convolutions for Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 1510–1517. [Google Scholar] [CrossRef] [Green Version]
  40. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D Convolutional Neural Networks for Human Action Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
  41. Oikonomou, K.M.; Kansizoglou, I.; Manaveli, P.; Grekidis, A.; Menychtas, D.; Aggelousis, N.; Sirakoulis, G.C.; Gasteratos, A. Joint-Aware Action Recognition for Ambient Assisted Living. In Proceedings of the 2022 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 21–23 June 2022; pp. 1–6. [Google Scholar]
  42. Shah, A.; Mishra, S.; Bansal, A.; Chen, J.C.; Chellappa, R.; Shrivastava, A. Pose and Joint-Aware Action Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, WI, USA, 3–8 January 2022; pp. 3850–3860. [Google Scholar]
  43. Holte, M.B.; Tran, C.; Trivedi, M.M.; Moeslund, T.B. Human pose estimation and activity recognition from multi-view videos: Comparative explorations of recent developments. IEEE J. Sel. Top. Signal Process. 2012, 6, 538–552. [Google Scholar] [CrossRef]
  44. Nandagopal, S.; Karthy, G.; Oliver, A.S.; Subha, M. Optimal Deep Convolutional Neural Network with Pose Estimation for Human Activity Recognition. Comput. Syst. Sci. Eng. 2022, 44, 1719–1733. [Google Scholar] [CrossRef]
  45. Zhou, T.; Wang, W.; Qi, S.; Ling, H.; Shen, J. Cascaded human-object interaction recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4263–4272. [Google Scholar]
  46. Zhou, T.; Yang, Y.; Wang, W. Differentiable Multi-Granularity Human Parsing. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 8296–8310. [Google Scholar] [CrossRef]
  47. Ullah, A.; Muhammad, K.; Del Ser, J.; Baik, S.W.; de Albuquerque, V.H.C. Activity Recognition Using Temporal Optical Flow Convolutional Features and Multilayer LSTM. IEEE Trans. Ind. Electron. 2018, 66, 9692–9702. [Google Scholar] [CrossRef]
  48. He, J.Y.; Wu, X.; Cheng, Z.Q.; Yuan, Z.; Jiang, Y.G. DB-LSTM: Densely Connected Bi-Directional LSTM for Human Action Recognition. Neurocomputing 2021, 444, 319–331. [Google Scholar] [CrossRef]
  49. Sun, X.; Xu, H.; Dong, Z.; Shi, L.; Liu, Q.; Li, J.; Li, T.; Fan, S.; Wang, Y. CapsGaNet: Deep Neural Network Based on Capsule and GRU for Human Activity Recognition. IEEE Syst. J. 2022, 16, 5845–5855. [Google Scholar] [CrossRef]
  50. Ibrahim, M.S.; Muralidharan, S.; Deng, Z.; Vahdat, A.; Mori, G. A Hierarchical Deep Temporal Model for Group Activity Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1971–1980. [Google Scholar]
  51. Biswas, S.; Gall, J. Structural Recurrent Neural Network (SRNN) for Group Activity Analysis. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1625–1632. [Google Scholar]
  52. Ullah, A.; Muhammad, K.; Ding, W.; Palade, V.; Haq, I.U.; Baik, S.W. Efficient Activity Recognition Using Lightweight CNN and DS-GRU Network for Surveillance Applications. Appl. Soft Comput. 2021, 103, 107102. [Google Scholar] [CrossRef]
  53. Li, X.; Zhao, Z.; Wang, Q. Abssnet: Attention-Based Spatial Segmentation Network for Traffic Scene Understanding. IEEE Trans. Cybern. 2021, 52, 9352–9362. [Google Scholar] [CrossRef]
  54. Deng, J.; Li, L.; Zhang, B.; Wang, S.; Zha, Z.; Huang, Q. Syntax-Guided Hierarchical Attention Network for Video Captioning. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 880–892. [Google Scholar] [CrossRef]
  55. Yang, Z.; He, X.; Gao, J.; Deng, L.; Smola, A. Stacked Attention Networks for Image Question Answering. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 21–29. [Google Scholar]
  56. Baradel, F.; Wolf, C.; Mille, J. Human Action Recognition: Pose-Based Attention Draws Focus to Hands. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 604–613. [Google Scholar]
  57. Islam, M.M.; Iqbal, T. Multi-gat: A Graphical Attention-Based Hierarchical Multimodal Representation Learning Approach for Human Activity Recognition. IEEE Robot. Autom. Lett. 2021, 6, 1729–1736. [Google Scholar] [CrossRef]
  58. Long, X.; Gan, C.; Melo, G.; Liu, X.; Li, Y.; Li, F.; Wen, S. Multimodal Keyless Attention Fusion for Video Classification. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  59. Song, S.; Lan, C.; Xing, J.; Zeng, W.; Liu, J. Spatio-Temporal Attention-Based LSTM Networks for 3D Action Recognition and Detection. IEEE Trans. Image Process. 2018, 27, 3459–3471. [Google Scholar] [CrossRef]
  60. Cho, S.; Maqbool, M.; Liu, F.; Foroosh, H. Self-Attention Network for Skeleton-Based Human Action Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 635–644. [Google Scholar]
  61. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  62. Liu, J.; Luo, J.; Shah, M. Recognizing Realistic Actions from Videos “in the Wild”. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1996–2003. [Google Scholar]
  63. Reddy, K.K.; Shah, M. Recognizing 50 Human Action Categories of Web Videos. Mach. Vis. Appl. 2013, 24, 971–981. [Google Scholar] [CrossRef] [Green Version]
  64. Kuehne, H.; Jhuang, H.; Garrote, E.; Poggio, T.; Serre, T. HMDB: A Large Video Database for Human Motion Recognition. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2556–2563. [Google Scholar]
  65. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A Dataset of 101 Human Actions Classes from Videos in the Wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  66. Carreira, J.; Noland, E.; Banki-Horvath, A.; Hillier, C.; Zisserman, A. A Short Note About Kinetics-600. arXiv 2018, arXiv:1808.01340. [Google Scholar]
  67. Zhang, Z.; Lv, Z.; Gan, C.; Zhu, Q. Human Action Recognition Using Convolutional LSTM and Fully Connected LSTM With Different Attentions. Neurocomputing 2020, 410, 304–316. [Google Scholar] [CrossRef]
  68. Liu, A.A.; Su, Y.T.; Nie, W.Z.; Kankanhalli, M. Hierarchical Clustering Multi-Task Learning for Joint Human Action Grouping and Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 102–114. [Google Scholar] [CrossRef]
  69. Ye, J.; Wang, L.; Li, G.; Chen, D.; Zhe, S.; Chu, X.; Xu, Z. Learning Compact Recurrent Neural Networks With Block-Term Tensor Decomposition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9378–9387. [Google Scholar]
  70. Ullah, A.; Muhammad, K.; Haq, I.U.; Baik, S.W. Action Recognition Using Optimized Deep Autoencoder and CNN for surveillance Data Streams of Non-Stationary Environments. Future Gener. Comput. Syst. 2019, 96, 386–397. [Google Scholar] [CrossRef]
  71. Dai, C.; Liu, X.; Lai, J. Human Action Recognition Using Two-Stream Attention Based LSTM Networks. Appl. Soft Comput. 2020, 86, 105820. [Google Scholar] [CrossRef]
  72. Afza, F.; Khan, M.A.; Sharif, M.; Kadry, S.; Manogaran, G.; Saba, T.; Ashraf, I.; Damaševičius, R. A Framework of Human Action Recognition Using Length Control Features Fusion and Weighted Entropy-Variances Based Feature Selection. Image Vis. Comput. 2021, 106, 104090. [Google Scholar] [CrossRef]
  73. Muhammad, K.; Ullah, A.; Imran, A.S.; Sajjad, M.; Kiran, M.S.; Sannino, G.; de Albuquerque, V.H.C. Human Action Recognition Using Attention Based LSTM Network with Dilated CNN Features. Future Gener. Comput. Syst. 2021, 125, 820–830. [Google Scholar] [CrossRef]
  74. Al-Obaidi, S.; Al-Khafaji, H.; Abhayaratne, C. Making Sense of Neuromorphic Event Data for Human Action Recognition. IEEE Access 2021, 9, 82686–82700. [Google Scholar] [CrossRef]
  75. Zhang, L.; Lim, C.P.; Yu, Y. Intelligent Human Action Recognition Using an Ensemble Model of Evolving Deep Networks with Swarm-Based Optimization. Knowl.-Based Syst. 2021, 220, 106918. [Google Scholar] [CrossRef]
  76. Hussain, A.; Hussain, T.; Ullah, W.; Baik, S.W. Vision transformer and deep sequence learning for human activity recognition in surveillance videos. Comput. Intell. Neurosci. 2022, 2022, 3454167. [Google Scholar] [CrossRef]
  77. Du, Z.; Mukaidani, H. Linear Dynamical Systems Approach for Human Action Recognition with Dual-Stream Deep Features. Appl. Intell. 2022, 52, 452–470. [Google Scholar] [CrossRef]
  78. Bao, W.; Yu, Q.; Kong, Y. Evidential Deep Learning for Open Set Action Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 13349–13358. [Google Scholar]
  79. Wang, X.; Gao, L.; Wang, P.; Sun, X.; Liu, X. Two-Stream 3-D Convnet Fusion for Action Recognition in Videos with Arbitrary Size and Length. IEEE Trans. Multimed. 2017, 20, 634–644. [Google Scholar] [CrossRef]
  80. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal Segment Networks for Action Recognition in Videos. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2740–2755. [Google Scholar] [CrossRef] [Green Version]
  81. Yu, S.; Xie, L.; Liu, L.; Xia, D. Learning Long-Term Temporal Features with Deep Neural Networks for Human Action Recognition. IEEE Access 2019, 8, 1840–1850. [Google Scholar] [CrossRef]
  82. Ma, C.Y.; Chen, M.H.; Kira, Z.; AlRegib, G. TS-LSTM and Temporal-Inception: Exploiting Spatiotemporal Dynamics for Activity Recognition. Signal Process. Image Commun. 2019, 71, 76–87. [Google Scholar] [CrossRef] [Green Version]
  83. Diba, A.; Fayyaz, M.; Sharma, V.; Paluri, M.; Gall, J.; Stiefelhagen, R.; Van Gool, L. Holistic Large Scale Video Understanding. arXiv 2019, arXiv:1904.11451. [Google Scholar]
  84. Majd, M.; Safabakhsh, R. Correlational Convolutional LSTM for Human Action Recognition. Neurocomputing 2020, 396, 224–229. [Google Scholar] [CrossRef]
  85. Zhu, L.; Fan, H.; Luo, Y.; Xu, M.; Yang, Y. Temporal Cross-Layer Correlation Mining for Action Recognition. IEEE Trans. Multimed. 2021, 24, 668–676. [Google Scholar] [CrossRef]
  86. Xiao, J.; Jing, L.; Zhang, L.; He, J.; She, Q.; Zhou, Z.; Yuille, A.; Li, Y. Learning from Temporal Gradient for Semi-Supervised Action Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3252–3262. [Google Scholar]
  87. Chen, S.; Ge, C.; Tong, Z.; Wang, J.; Song, Y.; Wang, J.; Luo, P. Adaptformer: Adapting vision transformers for scalable visual recognition. arXiv 2022, arXiv:2205.13535. [Google Scholar]
  88. Ranasinghe, K.; Naseer, M.; Khan, S.; Khan, F.S.; Ryoo, M.S. Self-supervised video transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2874–2884. [Google Scholar]
  89. Xing, Z.; Dai, Q.; Hu, H.; Chen, J.; Wu, Z.; Jiang, Y.G. Svformer: Semi-supervised video transformer for action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 18816–18826. [Google Scholar]
  90. Zhu, Y.; Newsam, S. Random Temporal Skipping for Multirate Video Analysis. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 542–557. [Google Scholar]
  91. Wang, X.; Gao, L.; Song, J.; Shen, H. Beyond Frame-Level CNN: Saliency-Aware 3-D CNN with LSTM for Video Action Recognition. IEEE Signal Process. Lett. 2016, 24, 510–514. [Google Scholar] [CrossRef]
  92. Feichtenhofer, C.; Pinz, A.; Wildes, R.P. Spatiotemporal Multiplier Networks for Video Action Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4768–4777. [Google Scholar]
  93. Sun, S.; Kuang, Z.; Sheng, L.; Ouyang, W.; Zhang, W. Optical Flow Guided Feature: A Fast and Robust Motion Representation for Video Action Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1390–1399. [Google Scholar]
  94. Fan, L.; Huang, W.; Gan, C.; Ermon, S.; Gong, B.; Huang, J. End-to-End Learning of Motion Representation for Video Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6016–6025. [Google Scholar]
  95. Long, X.; Gan, C.; De Melo, G.; Wu, J.; Liu, X.; Wen, S. Attention Clusters: Purely Attention Based Local Feature Integration for Video Classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7834–7843. [Google Scholar]
  96. Ullah, A.; Ahmad, J.; Muhammad, K.; Sajjad, M.; Baik, S.W. Action Recognition in Video Sequences Using Deep Bi-Directional LSTM with CNN Features. IEEE Access 2017, 6, 1155–1166. [Google Scholar] [CrossRef]
  97. Li, Z.; Gavrilyuk, K.; Gavves, E.; Jain, M.; Snoek, C.G. Videolstm Convolves, Attends and Flows for Action Recognition. Comput. Vis. Image Underst. 2018, 166, 41–50. [Google Scholar] [CrossRef] [Green Version]
  98. Han, Y.; Zhang, P.; Zhuo, T.; Huang, W.; Zhang, Y. Going Deeper with Two-Stream ConvNets for Action Recognition in Video Surveillance. Pattern Recognit. Lett. 2018, 107, 83–90. [Google Scholar] [CrossRef]
  99. Zhou, Y.; Sun, X.; Zha, Z.J.; Zeng, W. Mict: Mixed 3D/2D Convolutional Tube for Human Action Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 449–458. [Google Scholar]
  100. Song, X.; Lan, C.; Zeng, W.; Xing, J.; Sun, X.; Yang, J. Temporal–Spatial Mapping for Action Recognition. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 748–759. [Google Scholar] [CrossRef]
  101. Jiang, B.; Wang, M.; Gan, W.; Wu, W.; Yan, J. STM: Spatiotemporal and Motion Encoding for Action Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2000–2009. [Google Scholar]
  102. Phong, N.H.; Ribeiro, B. Video Action Recognition Collaborative Learning with Dynamics via PSO-ConvNet Transformer. arXiv 2023, arXiv:2302.09187. [Google Scholar]
  103. Yan, S.; Xiong, X.; Arnab, A.; Lu, Z.; Zhang, M.; Sun, C.; Schmid, C. Multiview Transformers for Video Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3333–3343. [Google Scholar]
  104. Hsiao, J.; Chen, J.; Ho, C. Gcf-Net: Gated Clip Fusion Network for Video Action Recognition. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 699–713. [Google Scholar]
  105. Zheng, Z.; An, G.; Wu, D.; Ruan, Q. Global and Local Knowledge-Aware Attention Network for Action Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 334–347. [Google Scholar] [CrossRef]
  106. Feichtenhofer, C.; Fan, H.; Malik, J.; He, K. Slowfast Networks for Video Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6202–6211. [Google Scholar]
  107. He, D.; Zhou, Z.; Gan, C.; Li, F.; Liu, X.; Li, Y.; Wang, L.; Wen, S. STNET: Local and Global Spatial–Temporal Modeling for Action Recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 8401–8408. [Google Scholar]
  108. Qiu, Z.; Yao, T.; Ngo, C.W.; Tian, X.; Mei, T. Learning Spatio-Temporal Representation with Local and Global Diffusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 12056–12065. [Google Scholar]
  109. Stroud, J.; Ross, D.; Sun, C.; Deng, J.; Sukthankar, R. D3d: Distilled 3D Networks for Video Action Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 625–634. [Google Scholar]
  110. Kondratyuk, D.; Yuan, L.; Li, Y.; Zhang, L.; Tan, M.; Brown, M.; Gong, B. Movinets: Mobile Video Networks for Efficient Video Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtually, 19–25 June 2021; pp. 16020–16030. [Google Scholar]
  111. Chen, J.; Ho, C.M. MM-ViT: Multi-Modal Video Transformer for Compressed Video Action Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, WI, USA, 3–8 January 2022; pp. 1910–1921. [Google Scholar]
  112. Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video Swin Transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 3202–3211. [Google Scholar]
  113. Munir, A.; Gordon-Ross, A.; Lysecky, S.; Lysecky, R. A Lightweight Dynamic Optimization Methodology and Application Metrics Estimation Model for Wireless Sensor Networks. Elsevier Sustain. Comput. Inform. Syst. 2013, 3, 94–108. [Google Scholar] [CrossRef]
  114. Alghamdi, Y.; Munir, A.; Ahmad, J. A Lightweight Image Encryption Algorithm Based on Chaotic Map and Random Substitution. Entropy 2022, 24, 1344. [Google Scholar] [CrossRef]
Figure 1. A graphical overview of our proposed activity-recognition framework. The proposed framework consists of three main modules: CNN architecture, dual channel and spatial attention module, and bi-directional GRU network. The CNN module utilizes a dual-attention mechanism to effectively extract salient CNN features from video frames, whereas the bi-directional GRU network is used to learn the activity representation for hidden sequential patterns.
Figure 1. A graphical overview of our proposed activity-recognition framework. The proposed framework consists of three main modules: CNN architecture, dual channel and spatial attention module, and bi-directional GRU network. The CNN module utilizes a dual-attention mechanism to effectively extract salient CNN features from video frames, whereas the bi-directional GRU network is used to learn the activity representation for hidden sequential patterns.
Jimaging 09 00130 g001
Figure 2. The building blocks of dual attention blocks containing channel and spatial attention mechanisms in detail.
Figure 2. The building blocks of dual attention blocks containing channel and spatial attention mechanisms in detail.
Jimaging 09 00130 g002
Figure 3. Visual representation of the salient object-specific regions computed with our dual attention mechanism.
Figure 3. Visual representation of the salient object-specific regions computed with our dual attention mechanism.
Jimaging 09 00130 g003
Figure 4. The building block of bi-directional single GRU layer.
Figure 4. The building block of bi-directional single GRU layer.
Jimaging 09 00130 g004
Figure 5. Validation history of our proposed DA-CNN+Bi-GRU framework along with other tested baseline methods for 300 epochs over three benchmark action datasets: (a) Validation history for YouTube action dataset, (b) Validation history for UCF50 dataset, (c) Validation history for HMDB51 dataset, (d) Validation history for UCF101 dataset, and (e) Validation history for Kinetics-600 dataset.
Figure 5. Validation history of our proposed DA-CNN+Bi-GRU framework along with other tested baseline methods for 300 epochs over three benchmark action datasets: (a) Validation history for YouTube action dataset, (b) Validation history for UCF50 dataset, (c) Validation history for HMDB51 dataset, (d) Validation history for UCF101 dataset, and (e) Validation history for Kinetics-600 dataset.
Jimaging 09 00130 g005
Figure 6. Confusion matrices computed for the proposed DA-CNN+Bi-GRU for the test sets of five tested datasets: (a) YouTube Action dataset, (b) HMDB51 dataset, (c) UCF50 dataset, (d) UCF101 dataset, and (e) Kinetics-600 dataset.
Figure 6. Confusion matrices computed for the proposed DA-CNN+Bi-GRU for the test sets of five tested datasets: (a) YouTube Action dataset, (b) HMDB51 dataset, (c) UCF50 dataset, (d) UCF101 dataset, and (e) Kinetics-600 dataset.
Jimaging 09 00130 g006
Figure 7. Category-wise accuracy of the proposed DA-CNN+Bi-GRU on the test sets of five tested datasets: (a) YouTube Action dataset, (b) HMDB51 dataset, (c) UCF50 dataset, (d) UCF101 dataset, and (e) Kinetics-600 dataset.
Figure 7. Category-wise accuracy of the proposed DA-CNN+Bi-GRU on the test sets of five tested datasets: (a) YouTube Action dataset, (b) HMDB51 dataset, (c) UCF50 dataset, (d) UCF101 dataset, and (e) Kinetics-600 dataset.
Jimaging 09 00130 g007
Figure 8. The visual recognition results of our proposed DA-CNN+Bi-GRU framework with predicted classes and their confidence scores for the test videos taken from the YouTube action, UCF50, and HMDB51 datasets.
Figure 8. The visual recognition results of our proposed DA-CNN+Bi-GRU framework with predicted classes and their confidence scores for the test videos taken from the YouTube action, UCF50, and HMDB51 datasets.
Jimaging 09 00130 g008
Table 1. Architectural details of our proposed CNN architecture.
Table 1. Architectural details of our proposed CNN architecture.
LayerInput ChannelsNumber of KernelsKernel SizeStridePaddingOutput Channels
Conv 1316 3 × 3 1116
Conv 21616 3 × 3 1116
Max pooling
Channel Attention
Spatial Attention
Conv 33232 3 × 3 1132
Conv 43232 3 × 3 1132
Max pooling
Channel Attention
Spatial Attention
Conv 532323×31132
Conv 632323×31132
Max pooling
Channel Attention
Spatial Attention
Conv 73264 3 × 3 1164
Conv 83264 3 × 3 1164
Max pooling
Channel Attention
Spatial Attention
Global Average Pooling
Flatten
Table 2. Network settings of experimented baseline methods and our proposed framework.
Table 2. Network settings of experimented baseline methods and our proposed framework.
MethodSpatial Block LayersTemporal Block Layers
CNN+LSTM8 convolutional3 LSTM
CNN+Bi-LSTM8 convolutional6 LSTM (3 forward and 3 backward)
CNN+GRU8 convolutional3 GRU
CNN+Bi-GRU8 convolutional6 GRU (3 forward and 3 backward)
DA-CNN+Bi-GRU12 convolutional (8 convolutional and 4 attentional)6 GRU (3 forward and 3 backward)
Table 3. Quantitative comparative analysis of our proposed framework with other baseline methods. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 3. Quantitative comparative analysis of our proposed framework with other baseline methods. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodDatasetAccuracy (%)
CNN+LSTMYouTube action64.7
CNN+Bi-LSTMYouTube action84.2
CNN+GRUYouTube action88.5
CNN+Bi-GRUYouTube action92.1
CNN (channel attention only)+Bi-GRU YouTube action 94.2
CNN (spatial attention only)+Bi-GRU YouTube action 95.6
DA-CNN+Bi-GRU (Proposed)YouTube action98.0
CNN+LSTMUCF5076.3
CNN+Bi-LSTMUCF5083.3
CNN+GRUUCF5087.6
CNN+Bi-GRUUCF5093.6
CNN (channel attention only)+Bi-GRU UCF50 95.1
CNN (spatial attention only)+Bi-GRU UCF50 95.7
DA-CNN+Bi-GRU (Proposed)UCF5098.5
CNN+LSTMHMDB5156.7
CNN+Bi-LSTMHMDB5163.2
CNN+GRUHMDB5168.0
CNN+Bi-GRUHMDB5172.4
CNN (channel attention only)+Bi-GRU HMDB51 73.9
CNN (spatial attention only)+Bi-GRU HMDB51 74.5
DA-CNN+Bi-GRU (Proposed)HMDB5179.3
CNN+LSTMUCF10183.9
CNN+Bi-LSTMUCF10186.8
CNN+GRUUCF10190.7
CNN+Bi-GRUUCF10194.2
CNN (channel attention only)+Bi-GRU UCF101 95.1
CNN (spatial attention only)+Bi-GRU UCF101 95.8
DA-CNN+Bi-GRU (Proposed)UCF10197.6
CNN+LSTMKinetics-60073.2
CNN+Bi-LSTMKinetics-60077.9
CNN+GRUKinetics-60081.5
CNN+Bi-GRUKinetics-60084.3
CNN (channel attention only)+Bi-GRU Kinetics-600 84.9
CNN (spatial attention only)+Bi-GRU Kinetics-600 85.6
DA-CNN+Bi-GRU (Proposed)Kinetics-60086.7
Table 4. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods on YouTube action dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 4. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods on YouTube action dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodYearAccuracy (%)
Multi-task hierarchical clustering [68]201789.7
BT-LSTM [69]201885.3
Deep autoencoder [70]201996.2
STDAN [67]202098.2
Two-stream attention LSTM [71]202096.9
Weighted entropy-variance-based feature selection [72]202194.5
Dilated CNN+BiLSTM+RB [73]202189.0
DS-GRU [52]202197.1
Local-global features + QSVM [74]202182.6
DA-CNN+Bi-GRU (Proposed)202398.0
Table 5. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for UCF50 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 5. Quantitative comparative analysis of our proposed method with the state-of-the-art action-recognition methods for UCF50 dataset. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodYearAccuracy (%)
Multi-task hierarchical clustering [68]201793.2
Deep autoencoder [70]201996.4
Ensemble model with swarm-based optimization [75]202192.2
DS-GRU [52]202195.2
Local-global features + QSVM [74]202169.4
ViT+LSTM [76] 2021 96.1
(LD-BF) + (LD-DF) [77]202297.5
DA-CNN+Bi-GRU (Proposed)202398.5
Table 9. Runtime analysis of our proposed framework with the state-of-the-art human action-recognition methods (without scaling). The best and the runner-up results are highlighted in bold and italic, respectively.
Table 9. Runtime analysis of our proposed framework with the state-of-the-art human action-recognition methods (without scaling). The best and the runner-up results are highlighted in bold and italic, respectively.
MethodSeconds per Frame (SPF)YearFrames per Second (FPS)
GPUCPUGPUCPU
STPP+LSTM [79]0.0053-2017186.6-
CNN with Bi-LSTM [96]0.0570-201720-
OFF [93]0.0048-2018206-
Videolstm [97]0.0940-201810.6-
Optical flow + multi-layer LSTM [47]0.03560.182018303.5
Deep autoencoder [70]0.04300.432019241.5
TSN+TSM [100]0.0167-201960-
IP-LSTM [81]0.0431-201923.2-
STDN [67]0.0075-2020132-
DS-GRU [52]0.0400-202125-
MoviNet [110]0.0833-202112-
(LD-BF) + (LD-DF) [77]0.0670-202214-
DA-CNN+Bi-GRU (Proposed)0.00360.00492023300250
Table 10. Runtime analysis of our proposed framework with the state-of-the-art human action-recognition methods scaled to our framework’s hardware specifications. The best and the runner-up results are highlighted in bold and italic, respectively.
Table 10. Runtime analysis of our proposed framework with the state-of-the-art human action-recognition methods scaled to our framework’s hardware specifications. The best and the runner-up results are highlighted in bold and italic, respectively.
MethodSeconds per Frame (SPF)YearFrames per Second (FPS)
GPUCPUGPUCPU
STPP+LSTM [79]0.0023-2017423.58-
CNN with Bi-LSTM [96]0.0354-201732.14-
OFF [93]0.0029-2018331.04-
Videolstm [97]0.0584-201817.03-
Optical flow + multi-layer LSTM [47]0.02210.17201848.213.71
Deep autoencoder [70]0.02670.40201938.561.59
TSN+TSM [100]0.0167-201960-
IP-LSTM [81]0.0268-201937.28-
STDN [67]0.0046-2020212.12-
DS-GRU [52]0.0248-202140.17-
MoviNet [110]0.0645-202115.48-
(LD-BF) + (LD-DF) [77]0.0416-202222.49-
DA-CNN+Bi-GRU (Proposed)0.00360.00492023300250
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ullah, H.; Munir, A. Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework. J. Imaging 2023, 9, 130. https://doi.org/10.3390/jimaging9070130

AMA Style

Ullah H, Munir A. Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework. Journal of Imaging. 2023; 9(7):130. https://doi.org/10.3390/jimaging9070130

Chicago/Turabian Style

Ullah, Hayat, and Arslan Munir. 2023. "Human Activity Recognition Using Cascaded Dual Attention CNN and Bi-Directional GRU Framework" Journal of Imaging 9, no. 7: 130. https://doi.org/10.3390/jimaging9070130

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop