Next Article in Journal
Synergistic Multi-Granularity Rough Attention UNet for Polyp Segmentation
Previous Article in Journal
The Effect of Simulated Dose Reduction on the Performance of Artificial Intelligence in Chest Radiography
Previous Article in Special Issue
Supervised and Self-Supervised Learning for Assembly Line Action Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Machine Learning for Human Activity Recognition: State-of-the-Art Techniques and Emerging Trends

by
Md Amran Hossen
and
Pg Emeroylariffion Abas
*
Faculty of Integrated Technologies, Universiti Brunei Darussalam, Bandar Seri Begawan BE 1410, Brunei
*
Author to whom correspondence should be addressed.
J. Imaging 2025, 11(3), 91; https://doi.org/10.3390/jimaging11030091
Submission received: 18 February 2025 / Revised: 14 March 2025 / Accepted: 15 March 2025 / Published: 20 March 2025

Abstract

:
Human activity recognition (HAR) has emerged as a transformative field with widespread applications, leveraging diverse sensor modalities to accurately identify and classify human activities. This paper provides a comprehensive review of HAR techniques, focusing on the integration of sensor-based, vision-based, and hybrid methodologies. It explores the strengths and limitations of commonly used modalities, such as RGB images/videos, depth sensors, motion capture systems, wearable devices, and emerging technologies like radar and Wi-Fi channel state information. The review also discusses traditional machine learning approaches, including supervised and unsupervised learning, alongside cutting-edge advancements in deep learning, such as convolutional and recurrent neural networks, attention mechanisms, and reinforcement learning frameworks. Despite significant progress, HAR still faces critical challenges, including handling environmental variability, ensuring model interpretability, and achieving high recognition accuracy in complex, real-world scenarios. Future research directions emphasise the need for improved multimodal sensor fusion, adaptive and personalised models, and the integration of edge computing for real-time analysis. Additionally, addressing ethical considerations, such as privacy and algorithmic fairness, remains a priority as HAR systems become more pervasive. This study highlights the evolving landscape of HAR and outlines strategies for future advancements that can enhance the reliability and applicability of HAR technologies in diverse domains.

1. Introduction

Human activity recognition (HAR) has gained significant attention due to its potential in interpreting human actions across various domains. Consider a high-risk industrial setting, such as an offshore platform, where HAR is used to identify hazardous activities. The effectiveness of HAR in such environments is influenced by external factors, including fluctuating lighting conditions due to changing weather. Vision-based HAR systems may struggle in low-light conditions, potentially missing critical safety incidents. In contrast, depth sensors, which are not affected by lighting conditions, can effectively recognise hazardous movements, allowing real-time safety monitoring and intervention. This example underscores the practical significance of HAR, demonstrating its capability beyond theoretical research.
Beyond workplace safety, HAR plays a crucial role in healthcare, smart environments, sports analytics, surveillance and security systems, autonomous driving systems, human–computer interaction (HCI), and human–robot interaction (HRI) [1,2,3,4,5,6,7]. In healthcare, HAR enables remote patient monitoring [8,9,10], while in smart environments, it has the potential to optimise energy consumption [11]. In sport analytics, HAR enhances sports performance analysis [12], and in security surveillance, it assists in detecting suspicious behaviour [13]. These diverse applications highlight the growing impact of HAR across multiple fields.
Recent advancements in machine learning (ML) and deep learning (DL) technologies [14,15,16,17,18,19,20,21,22,23,24,25] have significantly enhanced the scalability and accuracy of HAR systems [26]. Early stages of HAR research relied primarily on grayscale or RGB videos [27,28], but the field has since expanded over time to include a variety of other sensing modalities, including motion capture devices (Mocap), infrared sequences, event streams, depth sensors, point clouds, 2D/3D skeleton, audio streams, Wi-Fi signals, radio waves, Light Detection and Ranging (LiDAR), and a combination of environmental and proximity sensors [29,30,31,32,33,34,35,36,37,38,39]. The integration of wearable devices such as accelerometers [40,41,42], gyroscopes, and mobile phones has further enriched the HAR field. Multimodal approaches [43] offer a more comprehensive insights into human activities, particularly in scenarios where single modality sensing is inadequate.
Over the past decade, numerous reviews have focused on various aspects of HAR, such as human motion analysis, motion analysis in video data, posture prediction, HAR from 3D data, depth data, vision-based data, sensor-based HAR, wearable sensor-based HAR, human pose estimation, skeleton-based HAR, and RFID-based HAR [44,45,46,47,48,49,50,51,52,53]. Moreover, several reviews have reviewed the diverse datasets and modalities used in HAR [54,55,56,57]. Despite this extensive body of work, a comprehensive examination encompassing a wide array of data modalities including silhouettes, point clouds, range sensors, RGB, depth, skeleton, infrared sequences, event streams, motion capture devices (Mocap), acoustic sensors, wearable devices, radar, Wi-Fi, and sensor fusion remains limited. This review aims to fill this gap by providing an integrated perspective on the latest advancements in HAR across these diverse modalities.
This review presents a thorough review of the latest breakthroughs in machine learning technologies which are tailored for improving HAR. It examines the significance of various methodologies proposed in the literature together with their applicability. The primary objective is to offer a comprehensive summary that highlights recent advancements as well as their implications. By combining insights from various studies, this review provides a cohesive view of current HAR technologies. This review also identifies unresolved issues and limitations within existing approaches that underscore critical needs and urge for further research to develop solutions that bridge these gaps. By identifying potential areas for exploration, this review serves as a roadmap for advancing the field of human activity recognition and fostering a deeper understanding of both its current capabilities and future potentials.
This review paper is carefully structured into six sections to provide a systematic exploration of human activity recognition using machine learning and deep learning techniques. After this introduction, fundamental concepts and historical developments within HAR is presented in Section 2 which sets up the stage for deeper discussions. Various data modalities that are used for HAR research are reviewed in Section 3 which also details the specific sensors and technologies that are instrumental in advancing the field. Section 4 explores the publicly available datasets that are pivotal for training and evaluating HAR systems, along with a discussion on their characteristics and applications. Different machine learning techniques which are frequently applied to HAR are analysed in Section 5, which also highlights contributions and comparative evaluation. Subsequent sections identify the challenges and research gaps in the field of HAR and also propose potential research directions that could address these. This review concludes by presenting a summary of the findings and their implications for future technology development in human activity recognition, emphasizing the continued need for innovation and interdisciplinary collaboration.

2. Review Methodology

This section details the systematic review process undertaken for this study, describing the eligibility criteria for the study selection, search strategy, and screening process. This ensures transparency and reproducibility. The methodology follows the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) 2020 [58,59] guidelines for systematic reviews.

2.1. Eligibility Criteria

2.1.1. Inclusion Criteria

An article was considered for this review if any of the following conditions are satisfied:
  • The article either investigates or proposes methods for HAR.
  • It uses sensor data relevant to HAR, such as RGB, depth, infrared, motion capture devices, skeleton joint positions, wearable devices, acoustic sensors, radar, WiFi, LiDAR, proximity sensors, or combinations of these.
  • It applies machine learning or deep learning techniques for activity classification.
  • The article is published in English in a peer-reviewed journal or conference proceeding.

2.1.2. Exclusion Criteria

An article was excluded from this review if it met any of the following conditions:
  • The article focuses on sensor development without direct relevance to HAR.
  • The article does not employ machine learning or deep learning for activity recognition.
  • The article primarily focuses on human detection, pose estimation, or motion identification without activity classification.
  • Reviews and non-peer-reviewed, non-English articles.

2.2. Search Criteria

To identify relevant studies, a comprehensive search was conducted across the following databases up to 30 August 2024:
  • Scopus
  • Google Scholar
  • IEEE Xplore Digital Library
  • ACM Digital Library
The following keyword combinations were used to query each database during the search process:
  • Activity-related keywords: [“human activity recognition” OR “activity monitoring” OR “gesture recognition” OR “Activity Discovery”].
  • Machine learning-related keywords: [machine learning OR deep learning OR convolutional neural network OR recurrent neural network OR support vector machine OR random forest OR Reinforcement Learning OR Graph Convolutional Networks OR Generative Adversarial Networks OR Autoencoders OR Artificial neural network OR ANN].
  • Sensor-related keywords: [RGB OR Depth OR Infrared OR Motion capture devices OR Mocap OR Skelton joint positions OR Wearable devices OR accelerometer OR gyroscope OR Acoustic sensors OR Radar OR Wifi OR Lidar OR Proximity sensors OR Fusion HAR].

2.3. Screening Strategy

The PRISMA 2020 guidelines were adopted for the article selection process, as illustrated in Figure 1 (PRISMA flowchart). The selection process consisted of four stages:
  • Identification:
    The initial search yielded 7938 records from the seven databases: ACM (479), IEEE Xplore (4760), Elsevier (563), Springer (1570), Wiley (18), Taylor and Francis (32), and Nature (286).
  • Screening:
    A total of 5759 records were screened based on titles and abstracts to remove studies that were not relevant to the scope of this review. In total, 1949 records were excluded for not utilising machine learning or deep learning techniques for HAR. Additionally, 3890 articles were found to be inaccessible due to paywall restrictions, despite utilizing institutional subscriptions and exploring open-access alternatives. These constraints limited access to potentially relevant studies, but efforts were made to ensure that the available open-access literature provided a comprehensive and representative dataset.
  • Eligibility:
    Following the initial screening, 1869 full-text articles were assessed for eligibility based on the predefined inclusion and exclusion criteria. During this stage, 650 articles were excluded for reasons such as lack of implementation details, a narrow focus on partial-body movements, or the absence of performance metrics essential for evaluating HAR systems. Additionally, 860 articles were excluded due to significant overlap with existing studies, ensuring that only unique and methodologically sound research was retained for review.
  • Inclusion:
    After completing the eligibility assessment, 359 studies met the inclusion criteria and were included for detailed analysis. These studies represent a broad range of sensor modalities, machine learning methodologies, and emerging trends in HAR research. The selected studies provide valuable insights into the field, addressing key challenges and identifying potential research directions.

3. Human Activity Recognition (HAR)

3.1. Definition and Categories of Human Activity Recognition

3.1.1. Overview of Human Activity Recognition

Human Activity Recognition (HAR) is the automated process that can identify and interpret human actions using a sensing modality and a learning algorithm. HAR involves several stages including data capturing, data processing, and analysing data from various types of sources. Sensors which are utilised to classify and understand human actions often include devices such as wearable sensors, cameras, and depth sensors [55]. The primary objective of any HAR system is to interpret a diverse range of human activities from gestures to complex sequences like dancing to improve our understanding of human activities in different contexts [60].
Leveraging advanced sensors and computational techniques, HAR enables the development of sophisticated systems capable of perceiving, predicting, and responding to human activities. These capabilities are pivotal for driving progress in several domains, including healthcare [61], smart environments [62], robotics [63], surveillance, and human–computer interaction [64]. The interdisciplinary nature of HAR integrates elements from computer vision, signal processing, machine learning, and sensor technologies, enabling the accurate recognition and interpretation of human activities.
HAR significantly impacts a wide range of sectors, including healthcare, security and surveillance, smart homes, autonomous driving, entertainment, sports analytics, industrial automation, and education. The diverse applications of HAR in these fields are illustrated in Figure 2, showcasing the transformative role of this technology in enhancing our understanding and interaction with human activities.

3.1.2. Categories of Activities

Human activities encompass a broad spectrum of actions and behaviours that can be classified into several categories based on their nature and complexity, as illustrated in Figure 3. These categories, widely recognised in the literature [65,66], provide a structured framework for understanding the various types of activities performed by humans.
Gestures represent the most fundamental form of human activity, involving intentional body movements or signals made with the hands, arms, or entire body to convey information, commands, or expressions. These can range from simple gestures, such as waving or pointing, to more intricate movements used in sign language or other communication interfaces [67,68].
Building upon these basic gestures are atomic actions, which are simple, fundamental movements that are performed independently. Examples include walking, running, sitting, standing, jumping, or crouching. These atomic actions often serve as the foundational elements for recognising more complex activities and behaviours [69].
Interactions elevate the complexity of human activities by involving engagements between humans or between humans and objects. Characterised by a dynamic exchange of information, actions, or responses [70,71], human-to-human interactions include communication, cooperation, or collaboration, such as a conversation between two people. Human–object interactions [65,72], on the other hand, involve individuals engaging with physical objects or systems, such as manipulating tools or interacting with technology.
When multiple individuals coordinate their actions, they participate in group activities. These activities often require synchronisation, cooperation, and communication among participants to achieve shared goals or outcomes [73,74]. Examples of group activities include team sports like football or volleyball [75], performing in an orchestra, participating in collaborative projects, or engaging in collective performances [76,77].
Behaviours encompass a broad range of actions, expressions, and responses exhibited by individuals or groups. This category includes social behaviours, emotional expressions, habitual actions, decision-making patterns, and psychological responses observed in various contexts.
Events are more complex forms of human activities that typically involve multiple individuals and occur within specific environments. Examples of such events include music festivals, concerts, and weddings, where various activities take place simultaneously and involve numerous participants.
Based on the nature of motion or postures, activities can be further categorised into static actions, transitional actions, and dynamic actions. Static actions involve minimal movement, such as sitting and reading a book, whereas transitional actions involve a change in posture, such as sitting down or standing up. Dynamic actions are more complex and involve significant motion, such as walking and running. Activities can also be grouped into broader categories, including daily living activities, remote health monitoring tasks, sports actions, and activities related to security and safety, such as suspicious or dangerous activities.

3.2. Stages of HAR

The HAR process typically starts with data acquisition, where raw sensor data are collected from various sources, such as accelerometers, gyroscopes, or cameras, to capture human movements and actions. This stage involves selecting the appropriate sensors based on the type of activities and the environment in which they occur, determining suitable sampling rates, and ensuring data quality to facilitate accurate activity recognition. Figure 4 illustrates the various stages involved in the HAR process.
Following data acquisition, the next stage is data pre-processing, which aims to improve data quality and extract the relevant features necessary for activity classification. Pre-processing techniques include data cleaning to remove irrelevant or erroneous data, data filtering to eliminate noise, normalisation to standardise the data, and background removal in the case of video data. Additional techniques such as feature extraction, data segmentation, feature selection, and dimensionality reduction are applied to prepare the data for model training and to enhance the accuracy of subsequent activity recognition stages.
Once the data have been pre-processed, the focus shifts to model selection and training. In this stage, the most suitable machine learning or deep learning algorithms are chosen based on the nature of the data and the specific activity recognition tasks. This involves configuring model architectures, optimising hyperparameters, and training the model using labelled training data to develop robust models capable of accurately classifying activities from sensor data in real time. In cases where labelled data are unavailable, unsupervised methods are applied to estimate the number of activities in the dataset. Techniques such as clustering are used to group similar data points, after which class labels are assigned by examining cluster centres or random cluster members [76].
Following model training, fine-tuning techniques such as hyperparameter optimisation and transfer learning are employed to further refine model performance. Hyperparameter tuning involves adjusting parameters like learning rate and batch size to improve model accuracy, while transfer learning leverages pre-trained models to enhance performance on new but related tasks, reducing the need for large amounts of labelled data.
The final stage is activity recognition, where the trained model is deployed to classify real-time sensor data and infer human activities. Once activities are accurately identified, the information can be used in various applications to perform autonomous actions, such as alerting emergency response teams or notifying a family member in case of a fall. This ability to interpret and act upon human activities in real time highlights the transformative potential of HAR across numerous domains.

3.3. Activity Discovery and Activity Recognition

Various pervasive applications are made possible by human activity recognition, which converts noisy, low-level sensor input into high-level human activities [77]. The understanding of human activities includes both activity recognition and activity pattern discovery [78]. Activity recognition is centred on the precise recognition of human activities using a predefined activity model. Thus, a human activity recognition researcher first develops a high-level conceptual model, which is subsequently put into practice by creating an appropriate pervasive system [78]. Two assumptions are made in an activity recognition process: firstly, the set of activities is fixed, and the user activity model is fixed. However, both fall short when a system is deployed in an open and flexible settings, for an extended period of time, and target a broader spectrum of user. First of all, it is impractical to presume that only the pre-defined set of activities will be performed by users over an extended period of time. The pre-trained model may misclassify or fail to recognise a novel activity performed by a user if it was not observed or learned during the training period. Secondly, as user activities change over time, it can be challenging, or even infeasible, to collect all possible user activities during the training period. Both of the aforementioned scenarios may cause activity detection algorithms to perform worse, which could provide users with unwanted services.
Conversely, activity pattern discovery focuses more on immediately identifying unknown patterns from sensor data without the need for pre-established models or presumptions. In order to find activity patterns, a researcher that specialises in activity pattern detection first creates a ubiquitous system and then examines the sensor data. Despite their differences, the two approaches share the goal of advancing human activity technology [78]. They also work well together since the identified activity pattern can be used to define subsequent activities that can be identified and monitored.
The steps involving human activity recognition and activity pattern discovery are depicted in Figure 5 and Figure 6, respectively. Both the activity recognition and activity discovery process start with segmented activities; feature extraction/selection is an intermediate step which is followed by the activity representation step where features from consecutive frames are combined. Activity recognition requires the selection of an appropriate learning algorithm while activity discovery uses a clustering algorithm to estimate the number of activities in the dataset which are then annotated for classification. Methods often used by human activity discovery researchers are tabulated in Table 1.

3.4. Types of Sensors Used for Human Activity Recognition

Human activity recognition (HAR) methods can be broadly categorised into several categories based on the types of sensors used: non-visual sensors, visual-based sensors, context-aware systems, and hybrid approaches [69]. Each category leverages different modalities to capture and analyse human activities, offering unique advantages and addressing specific challenges associated with activity recognition. Understanding the strengths and limitations of these sensor types is crucial for developing effective HAR systems tailored to diverse applications and environments. Figure 7 provides an overview of the various HAR categories and their corresponding sensor modalities.
Different modalities of HAR, each with their respective advantages and limitations, are summarised in Table 2. By combining multiple modalities, hybrid methods aim to mitigate the drawbacks of single sensor approaches and improve the robustness and accuracy of activity recognition systems. Hybrid methods are typically characterised into three categories such as early fusion, late fusion, and hybrid fusion. In early fusion raw data or low-level features are combined at an early stage for HAR. While in the late fusion, each sensor modality is processed independently, and classification results are combined final classification. Hybrid fusion combines elements from both early and late fusion for improved classification.

4. Data Collection and Pre-Processing

In human activity recognition (HAR), the accuracy and reliability of the recognition systems heavily depend on the quality and diversity of the data used for model training and testing. Data collection plays a critical role in capturing relevant information from different sources, while data pre-processing ensures that this information is clean, structured, and ready for analysis. This section delves into both the publicly available datasets for HAR, categorised by different sensing modalities, and the various pre-processing techniques used to optimise the data for machine learning models.

4.1. Publicly Available Datasets for Human Activity Recognition

Publicly available datasets are essential resources for advancing and benchmarking research in human activity recognition (HAR). These datasets span a broad spectrum of data modalities, including vision-based, sensor-based, skeleton-based, and multimodal datasets.
Vision-based datasets typically consist of RGB images or video streams that capture human movements. RGB datasets allow researchers to analyse visual patterns and behaviour with the surrounding context. On the other hand, depth-based datasets include 3D spatial information from sensors such as Kinect or LiDAR. For instance, 3D datasets offer detailed data about object and human positioning in space. Sensor-based datasets are collected from sensors such as the accelerometers, gyroscopes, or inbuilt smartphone sensors. Sensor-based data provide motion and orientation data that are critical for activity recognition. Skeleton-based datasets focus on tracking key joint positions of the human body which simplify the representation of human motion while still enabling accurate activity analysis. Multimodal datasets integrate various data streams such as RGB, depth, infrared, and skeleton coordinates. Multimodal datasets provide a comprehensive and enriched dataset for recognising complex human activities and it is for scenarios where one sensing modality may seem inadequate. Key publicly available datasets for HAR from each category are listed in Table 3. The datasets that are mostly used from each sensor type with most subjects’ action classes are listed here; for a comprehensive list of publicly available HAR datasets, please refer to Supplementary Material S1, Tables S1–S7.

4.2. Data Pre-Processing

Data pre-processing is a critical stage in human activity recognition (HAR) that significantly enhances the quality, consistency, and usability of sensor data [124]. As raw data collected from various sensors can be noisy, incomplete, or inconsistent, effective pre-processing is essential for preparing the data for subsequent analysis and model development. Pre-processing techniques include a series of steps aimed at cleaning, transforming, and organizing data, ensuring it is suitable for activity recognition tasks [83].

4.2.1. Importance of Data Pre-Processing in Human Activity Recognition

Regardless of the sensing medium, data pre-processing plays a pivotal role in ensuring the quality and reliability of human activity recognition (HAR) systems [124]. It acts as a critical foundation for enhancing the accuracy and performance of models. It is essential for cleaning, organizing, and transforming raw sensor data into a usable format [124].
Similarly, data pre-processing is essential in RGB-based HAR, where raw video frames captured by RGB cameras often contain a wide range of unnecessary information, such as background clutter, lighting variations, and irrelevant movements [116]. Pre-processing techniques, such as background subtraction, frame cropping, and normalisation, help isolate human subjects, remove extraneous visual noise, and standardise colour channels for uniform input [125].

4.2.2. Data Pre-Processing Techniques

Data pre-processing involves a variety of techniques that are designed to transform raw and unprocessed data into a structured and refined format. Processed data ultimately enhance the accuracy and performance of human activity recognition (HAR) systems [126].

Data Preparation

Data preparation is a foundational step in pre-processing that involves two key tasks: activity segmentation and noise removal.
Activity segmentation is the process of dividing continuous data streams into smaller, manageable sections that correspond to specific actions or activities. This segmentation allows researchers to isolate and analyse individual activities more effectively, leading to more accurate labelling and classification of human actions [127]. Figure 8 illustrates an example where continuous human activities (top row) and hand joint movements over time (bottom row) are depicted.
Noise removal is equally important for HAR, as it involves eliminating unwanted disturbances from the raw data. If left unprocessed, these artifacts can obscure meaningful patterns and introduce inaccuracies during analysis. Noise removal techniques often include filtering and smoothing [128].

Feature Engineering

Feature extraction involves identifying relevant patterns from raw data that are indicative of specific human activities [129]. In human activity recognition (HAR), extracted features may include statistical measures such as mean, variance, or skewness [129], frequency domain characteristics like spectral entropy [62], or temporal patterns such as joint angles [129], joint positions [130], and joint orientations [131].
Once the relevant features have been identified, feature selection techniques are applied to retain only the most useful ones, while discarding irrelevant, redundant, or noisy features [132]. This step is crucial for reducing the dimensionality of the data and simplifying the subsequent analysis. Common methods for feature selection include techniques such as Principal Component Analysis (PCA) [83], Recursive Feature Elimination (RFE) [133], Gaussian filters [134], and various filtering approaches [135] that rank features based on their importance to the task.

Data Transformation and Standardisation

Data transformation and standardisation ensure consistency and comparability across different features or data types [136]. These techniques help stabilise training processes, reduce biases caused by varying data scales, and improve the performance of human activity recognition (HAR). Scaling and normalisation involve adjusting the range of input features to a standard scale, such as normalising pixel values between 0 and 1, or standardising features to have a mean of 0 and a standard deviation of 1 [137].

Data Enhancement and Handling Class Imbalances

Data enhancement and addressing class imbalances are critical for improving the performance of deep learning-based models in human activity recognition (HAR) [138]. Deep learning models require large amounts of diverse labelled data to generalise well [139]. Data augmentation is a key technique used to artificially expand the dataset by generating additional training samples through transformations [140] such as changing illumination, adding noise, applying rotations, translations, flips, crops, scaling, and zooms to the original data [141]. These transformations help diversify the training data, allowing models to learn from different variations of the same activity and thus reducing overfitting.
For sequential data, such as time-series data from sensors or video sequences, techniques like padding and trimming ensure that all sequences have uniform feature dimensions [142], which is necessary for efficient batch processing in deep learning models. Padding involves appending zeros to shorter sequences, while trimming longer sequences to a fixed length.
Another crucial component of data improvement is the encoding of categorical variables which involves converting categorical variables into numerical representations using methods like one-hot encoding. Encoding allows deep learning models to process non-numerical data effectively [143]. This is particularly useful in HAR when dealing with labels or other categorical variables that need to be converted into a format compatible with machine learning algorithms.
A common challenge in HAR datasets is class imbalance where certain activities occur more frequently than others. Class imbalance often leads to biased models. To address this, HAR researchers often use techniques like oversampling minority classes or under-sampling majority classes [143]. Oversampling is helpful to generate additional data points for underrepresented activities. Under-sampling reduces the data points from overrepresented activities. These approaches help achieve a balanced dataset which is crucial for ensuring that the model does not favour the majority class, and the model can effectively learn to recognise minority class activities.

5. Machine Learning Techniques in Human Activity Recognition

Before the widescale adoption of machine learning techniques, traditional approaches to human activity recognition (HAR) relied heavily on rule-based systems [144]. These methods were designed to detect and classify activities based on predefined criteria and certain threshold values. For example, to identify accelerometer-based activities, thresholding techniques were employed to recognise specific motions like walking and running by detecting changes in acceleration patterns that crossed predetermined limits [145]. Similarly, in time-series analysis, methods like peak detection or pattern matching were also used to identify repetitive activities such as cycling or repetitive arm movements.
The sections below provide a detailed overview of the various machine learning techniques employed in HAR, their methodologies, applications, and performance benefits, emphasising the role of both shallow learning and deep learning models in advancing the field.

5.1. Shallow Learning Models

5.1.1. Supervised Learning Techniques

Supervised machine learning techniques are fundamental to human activity recognition (HAR) as they rely on labelled datasets to train models for classifying human activities from various sensor modalities. These techniques are highly effective in extracting and learning patterns from data sources such as RGB images, wearable devices, mobile phone sensors, depth cameras, and skeletal tracking systems. By utilizing features extracted from these modalities, supervised learning algorithms can discern and classify distinct human actions with high accuracy.
Popular supervised learning algorithms in HAR include Decision Trees (DT), Random Forests (RF), Support Vector Machines (SVM), K-Nearest Neighbours (KNN), and gradient boosting methods like XGBoost and NGBoost. Each algorithm has its strengths and trade-offs depending on the sensor data and application scenarios.
Decision Trees (DT) [146,147,148,149,150] are widely used for classifying activities by partitioning the feature space based on threshold values. They are effective in handling continuous and categorical sensor data, such as accelerometers and gyroscope readings, as well as frame-wise features extracted from RGB images. However, a key limitation of Decision Trees is their tendency to overfit, especially when the tree becomes too deep, leading to poor generalisation on complex data.
Random Forest (RF), an ensemble method that combines multiple decision trees, help to mitigate overfitting and improve the overall accuracy of HAR systems [151,152,153,154,155,156]. RF models excel in capturing more complex patterns from high-dimensional sensor data and are commonly used in applications requiring robust classification. However, RF models are computationally expensive, especially when the number of trees is large, and they can suffer from interpretability challenges, particularly in the presence of highly imbalanced datasets.
Support Vector Machines (SVMs) are well-suited for HAR involving complex boundary separations in high-dimensional feature spaces [157,158,159,160,161,162]. They effectively utilize kernel methods to manage non-linear relationships between features from different sensors, making them ideal for differentiating activities like walking and running. However, SVMs are sensitive to kernel and hyperparameter configurations, and tuning them for optimal performance can be computationally intensive, particularly for large datasets.
K-Nearest Neighbour (KNN) is a simple yet powerful instance-based learning algorithm. It classifies activities based on the majority vote of the nearest data points in the feature space, which makes it intuitive and effective for HAR applications. However, KNN becomes computationally expensive for large datasets and struggles with performance when distinguishing between complex or overlapping activity patterns [163,164,165,166].
Hidden Markov Models (HMMs) are utilized to model sequential dependencies in HAR data, making them particularly useful for recognising activities like walking, running, or sitting, where temporal patterns are crucial. HMMs efficiently handle state transitions in time-series data, but their assumptions of fixed states and transition probabilities limit their flexibility in capturing continuous or highly variable human activities [167,168,169].
Naive Bayes classifiers are used for efficient classification in HAR by assuming conditional independence among features. They perform well in scenarios where this assumption holds true, but in real-world applications, sensor features are often correlated, which can reduce the model’s accuracy. Naive Bayes classifiers are commonly applied to sensor data from accelerometers, gyroscopes, magnetometers, and skeletal tracking systems [170,171,172,173].
Gradient boosting methods, such as XGBoost and NGBoost [174,175,176], have gained traction in HAR research for their ability to handle complex, multi-sensor data with high accuracy. These algorithms iteratively refine predictions to minimize errors, making them particularly effective for imbalanced datasets. However, gradient boosting methods are sensitive to hyperparameter tuning and require substantial computational resources, especially for large datasets.
Stochastic Gradient Descent (SGD) is commonly used in HAR for optimizing linear models, neural networks, and SVMs [177,178,179]. It updates model parameters incrementally based on small batches of sensor data, making it efficient for large-scale HAR applications. However, careful tuning of the learning rate and other hyperparameters is necessary to avoid convergence problems.
Supervised learning models are typically trained using labelled data, where a portion of the dataset is used for training, while the remaining data are reserved for testing and validation [41]. The models improve their ability to generalise across diverse activities using an iterative optimisation process. Hyperparameter tuning and cross-validation are frequently applied during the process. The main advantage of supervised learning in HAR is its ability to learn complex mappings between input features and activity labels. These methods lead to robust and real-time recognition performance across various environments. Despite their effectiveness, supervised learning methods require a substantial amount of labelled data for training [180]. The performance can degrade where labelled data are scarce; also, factors like high variability in activity contexts leads to poor performance [181].
Nonetheless, they continue to be widely used in HAR systems powering applications ranging from healthcare monitoring, rehabilitation, and human–computer interaction to smart environments. The versatility of supervised learning algorithms allows them to integrate information from multiple sensor modalities, enhancing the robustness and accuracy of activity recognition systems. Table S8 in Supplementary Material S1 provides an overview of various machine learning-based HAR techniques across different sensor modalities, showcasing the versatility of supervised learning methods in this domain.

5.1.2. Unsupervised Learning and Clustering Techniques

Unsupervised machine learning methods offer a valuable approach to human activity recognition (HAR) by uncovering hidden patterns and structures within sensor data without the need for labels [182,183]. These methods are especially useful when labelled data are scarce or unavailable, allowing the HAR system to autonomously learn activity patterns directly from raw data. In HAR systems utilising modalities such as RGB video, mobile phones, wearable sensors, depth sensors, Mocap, and skeleton data, unsupervised learning algorithms like clustering or dimensionality reduction techniques—such as Principal Component Analysis (PCA) and t-distributed Stochastic Neighbour Embedding (t-SNE) [83]—are commonly employed to identify underlying patterns within the data.
Clustering techniques, including K-means [184,185,186,187], Spectral Clustering [188,189], Agglomerative Clustering [188], DBSCAN [190], and Gaussian Mixture Models (GMM) [191,192], enable the grouping of similar activities into clusters by detecting natural groupings based on the similarity of features. These methods are essential for revealing inherent similarities and differences among various activities and sensor modalities, helping to cluster activities that share similar characteristics. For example, clustering algorithms can automatically group activities like walking, running, and sitting by analysing motion data collected from accelerometers and gyroscopes in wearable sensors.
Unsupervised learning techniques are particularly beneficial for anomaly detection and activity discovery in HAR. These methods allow the system to detect unknown or abnormal activities without requiring labelled examples. For instance, in healthcare settings, unsupervised learning can identify anomalous patterns in patient behaviour, such as unexpected falls or irregular walking patterns, which may signal health issues. Similarly, deep clustering [188], particle swarm optimisation, and game theory [83] methods have been explored to discover new activity patterns in HAR datasets, offering the potential for automatic identification of previously unseen or unlabelled activities.
By leveraging complementary information from various sensor modalities, unsupervised learning enables HAR systems to identify, group, and characterise activities based on their inherent similarities, contributing to applications such as anomaly detection, behaviour profiling, and activity summarisation. The flexibility of unsupervised learning techniques makes them invaluable for cases where labelled data are limited or unavailable, and they help the HAR system to generalise better across different activity contexts and sensor environments. Table S9 in Supplementary Material S1 summarises various unsupervised learning techniques and their applications in HAR using different modalities.

5.1.3. Semi-Supervised Learning Techniques

Semi-supervised learning methods strike a balance between supervised and unsupervised approaches, offering a pragmatic solution for human activity recognition (HAR) in situations where labelled data are limited or expensive to obtain [193]. These techniques leverage both labelled and unlabelled data, which are particularly advantageous in real-world applications where acquiring extensive labelled datasets is often challenging, but abundant unlabelled data are readily available. This hybrid learning paradigm is well-suited for HAR systems that utilise multiple data modalities, such as RGB video, mobile phones, wearable sensors, depth sensors, and skeletal data in smart home environments [194].
One innovative application of semi-supervised learning in HAR is the Federated Learning (FL) framework proposed by Yang et al. [193]. Their approach, FedHAR, addresses data scarcity by combining federated, semi-supervised, and transfer learning techniques. In this framework, a global model is trained using a small amount of labelled data from each user and subsequently personalised for individual users through transfer learning. FedHAR was validated on the MobiAct and WISDM datasets, demonstrating its effectiveness in adapting to personalised activity recognition scenarios.
A semi-supervised adversarial learning approach using Long Short-Term Memory (LSTM) networks for human activity recognition (HAR) is proposed by the authors in reference [194]. This method used both labelled and unlabelled data to train the model, enhancing its ability to adapt to new activities and changes in activity routines. The evaluation was conducted on datasets collected from smart home environments with heterogeneous sensors, specifically the Kasteren and CASAS datasets [195]. The proposed method achieved an accuracy of over 95% which surpassed existing state-of-the-art methods.
A semi-supervised active transfer learning (SATL) method for human activity recognition (HAR) to reduce the burden of manual data labelling is proposed by Oh et al. [196]. The method used previously learned data to reduce labelling tasks on new data. A DNN-based basic model was trained on an initial training set and used to construct a correct classifier model. Both models are then used to analyse unlabelled data and select the most informative samples for labelling by a human annotator. The method was evaluated on the HCI-HAR and mHealth dataset, achieving over 95% accuracy while significantly reducing the amount of data requiring manual labelling. A context-aware semi-supervised method was proposed by [197] which used the HCI-HAR [198], WISDM, PAMAP2 [199], mHealth [200] datasets for model validation.
Techniques like self-training, co-training, and graph-based semi-supervised learning iteratively refine model predictions by incorporating information from both labelled and unlabelled samples [201]. In self-training, a model is initially trained on a small amount of labelled data, and the confident predictions on unlabelled data are iteratively added to the training set [201]. Co-training involves training multiple models on different views or subsets of the data, where each model helps label the unlabelled data for the other. Graph-based methods use graph representations to propagate label information from labelled to unlabelled data based on their similarity. A semi-supervised method for HAR using multimodal data from Instagram, particularly image and text data, is proposed by Kim et al. [201]. The goal was to improve recognition accuracy while using a limited amount of labelled data, leveraging the abundance of unlabelled data on social media. The proposed method achieved a recognition accuracy of 71.58%, outperforming existing HAR methods on multimodal datasets.
Semi-supervised ensemble learning for HAR using the CASAS Kyoto dataset was explored by Patricia et al. [202]. The researchers applied a distance-based clustering analysis method to identify clusters of activity characteristics, which then served as the input for a supervised classification process. Evaluation using quality metrics indicates favourable results, with the average accuracy exceeding 95% when using the Agglomerative Clustering algorithm with the Bagging classifier. This paper [203] focused on leveraging unlabelled data to improve the performance of Convolutional Neural Networks (CNNs) in HAR. Two semi-supervised architectures, CNN-Encoder-Decoder and CNN-Ladder, were used for this purpose. The study evaluated the proposed method on three datasets: ActiTracker, PAMAP2, and mHealth. The paper reported up to an 18% improvement in F1-score compared to previous baseline methods.
A semi-supervised framework that used generative adversarial networks (GANs) with temporal convolutions for human activity recognition (HAR) is suggested by Hazar et al. [204]. The framework aimed to overcome the restrictions faced by conventional approaches, particularly the reliance on large, labelled datasets. The framework was evaluated using the PAMAP2, Opportunity-locomotion, and LISSI HAR datasets under real-world semi-supervised scenarios, examining factors such as inter-subject training, labelled data amount, class number, and IMU positioning effects on performance. Results demonstrated high classification performance and generalisation ability, with 25% improvement compared to baseline methods when limited annotated data were provided.
By combining the strength of labelled data with the abundance of unlabelled data, semi-supervised methods help the model generalise more effectively across diverse activity contexts and sensor modalities [197]. This approach enhances the robustness and scalability of HAR systems, enabling them to adapt to new activities, sensor configurations, and unseen environments with minimal supervision. Semi-supervised learning also allows for more accurate and reliable activity recognition in real-world applications like healthcare monitoring, smart environments, and human–computer interaction, where labelled data are often limited [194]. Various techniques in semi-supervised HAR are tabulated in Table S10 in Supplementary Material S1 showcasing their applications across multiple modalities.

5.2. Deep Learning Models

5.2.1. CNN, RNN, GCN, and LSTM-Based Methods

Deep learning-based methods have become the cornerstone of human activity recognition (HAR) across various sensing modalities, such as RGB [205], mobile phones [206], wearable sensors [207], depth sensors [208], and skeleton data [209]. This shift from traditional machine learning to deep learning approaches started gaining significant momentum in the mid-2010s when researchers realised the potential of deep neural networks to automatically learn hierarchical representations from raw sensor data without manual feature engineering [210]. Key deep learning architectures include Convolutional Neural Networks (CNNs) [15], Recurrent Neural Networks (RNNs) [19], Graph Convolutional Networks (GCNs) [21], Long Short-Term Memory (LSTM) [211], and Transformers architectures. Each architecture serves a distinct purpose: CNNs are effective for spatial data analysis, such as images or depth maps, while RNNs, LSTMs, and GCNs excel at capturing temporal dynamics or sequential dependencies in time-series data, such as sensor readings from wearables, mobile phones, or skeleton-based data. To capture long-range dependencies and the global context in sequential data, transformers are used for HAR due to their ability which make them effective for modalities like skeleton sequences, images, and time-series sensor data. The ability of these networks to process raw input directly, bypassing manual feature extraction, allows for more efficient recognition of complex human activities across diverse environments and modalities.
Deep learning models in HAR have greatly enhanced the field’s ability to interpret multimodal sensor data, enabling the extraction of more meaningful and robust features [62]. For example, CNNs were applied to process spatial features from images by authors in reference [212]. The authors in [72] applied RNNs to capture temporal relationships in time-series data. These models can be integrated into multimodal frameworks that fuse information from different sensors. Visual data from RGB cameras combined with accelerometer or gyroscope readings [213] can be used by these hybrid methods. This multimodal approach enhances accuracy and robustness, allowing HAR systems to perform more effectively in real-world applications, including healthcare monitoring [61], smart environments [214], and human–computer interaction [215]. A key advantage of deep learning over traditional methods is its ability to automatically learn complex, hierarchical features from large datasets, which are often necessary to capture the subtleties and intricacies of human activities [216]. This is particularly valuable in HAR scenarios where large-scale, diverse, and complex data are involved. Deep learning methods can generalise well across varied contexts and adapt to new activities or environments, making them more versatile and reliable.
In recent years, several advancements in deep learning have been introduced to further enhance HAR capabilities. The authors in reference [205] introduced a method for 3D human pose estimation from a single image using a Convolutional Neural Network (CNN) and transfer learning. To improve the accuracy of 3D pose estimation, the authors leveraged a large-scale 2D pose dataset. The methodology involves training a CNN to predict the 2D pose and then lifting the 2D predictions to 3D using a separate network. This study [94] introduced a novel dataset called DHP19 for 3D human pose estimation using dynamic vision sensors (DVSs). The dataset consists of synchronised recordings from four DVS cameras and a Vicon motion capture system. The paper presented a reference study on 3D human pose estimation using the DHP19 dataset and reported a 3D mean per joint position error or MPJPE of 87.54mm on the test set, which significantly reduced joint estimating errors with a reduced frame rate.
Convolutional Neural Networks (CNNs) have been used for human action recognition from 3D joints. For instance, a scene flow to action map description for RGB+D-based action recognition with CNNs was proposed by Wang et al. [217]. Here, 3D skeleton data were transformed to ten feature arrays and used as the input to CNNs for action recognition by Ke et al. [218]. In order to transform the visual appearance of human body parts obtained from various perspectives to a view-invariant space for depth-based human object interaction analysis, Rahmani et al. [219] created a deep CNN model.
A novel method called Spatial Temporal Graph Convolutional Networks (ST-GCNs) for skeleton-based action recognition was presented in this study [220]. The model works by constructing a set of spatial-temporal graph convolutions on the skeleton sequences to capture motion information. The authors evaluated the ST-GCN method on the Kinetics and NTU-RGB+D datasets. On Kinetics, ST-GCN achieved a top 1% accuracy of 30.7% and a top 5% accuracy of 52.8%, while on NTU-RGB+D, the model outperformed previous state-of-the-art skeleton-based models with cross-subject accuracy of 81.5% and cross-view accuracy of 88.3%. A new video-level framework called Temporal Segment Network (TSN) for action recognition in videos is proposed by [221] which aimed to model long-range temporal structures with a segment-based sampling and aggregation module. The model was evaluated on five challenging video-based action recognition benchmarks: HMDB51, UCF101, THUMOS14, ActivityNet v1.2, and Kinetics400. TSN achieved a state-of-the-art performance on these datasets, with accuracies of 71.0%, 94.9%, 80.1%, 89.6%, and 75.7%, respectively.
Actional-Structural Graph Convolutional Networks (AS-GCN) for skeleton-based action recognition was introduced in this work [222]. AS-GCN incorporated the actional links between joints in consecutive frames, in addition to the structural links between joints in a single frame. The model was evaluated on the NTU-RGB+D dataset and achieved an accuracy of 86.8% and 94.2% on the cross-subject and cross-view benchmark.
Richly Activated Graph Convolutional Network (RA-GCN) for robust skeleton-based action recognition was suggested by authors in this study [223]. RA-GCN applied a multi-stream framework with an activation module to handle noisy or incomplete skeleton data. The model was evaluated on the NTU RGB+D 60 and 120 datasets and outperforms state-of-the-art methods, particularly in scenarios with occlusion and jittering with an accuracy score of 87.3% and 81.1% on cross-subject settings.
Another graph-based method called Two-Stream Adaptive Graph Convolutional Networks (2S-AGCNs) for skeleton-based action recognition was proposed in this study [224]. The 2S-AGCN method used both joint and bone information by constructing two adaptive graphs, one for each type of information. The model was evaluated on the Kinect300 and NTU RGB+D dataset. The method achieves state-of-the-art performance on both the cross-subject and cross-view benchmarks with an accuracy score of 88.5% 95.1%. On the Kinetics dataset, the top 1% accuracy was 36.1% and top 5% accuracy was 58.7%.
Skeleton-based action recognition using Directed Graph Neural Networks (DGNs) was introduced in this study [225]. DGNs explicitly modelled the dependencies between joints by considering the direction of information flow. The model was evaluated on the NTU-RGB+D and Skeleton-Kinetics datasets and achieved competitive performance compared to other state-of-the-art methods. Feedback Graph Convolutional Network (FGCN) is another method for skeleton-based action recognition [226]. FGCN proposed a feedback mechanism into graph convolutional networks, enabling low-level layers to access semantic information from higher-level layers. The model is evaluated on multiple datasets including the NTU-RGB+D60, NTU-RGB+D120, Northwestern-UCLA datasets and achieves state-of-the-art performance. Shift Graph Convolutional Network (Shift-GCN) for skeleton-based action recognition is proposed by the authors [227]. Shift-GCN was designed to capture the temporal evolution of the skeleton sequence by using temporal shift modules. The model was assessed on NTU RGB+D, NTU RGB+D -120, and Northwestern-UCLA datasets and achieved accuracy of 96.5%, 85.9%, and 94.6% respectively.
Decoupled Spatial-Temporal Attention Network (DSTA-Net) was introduced by Shi et al. [228] for skeleton-based action and gesture recognition. In the DSTA-Net, spatial and temporal features are decoupled using attention modules without the need for hand-crafted features. Additionally, the method also proposed using four data streams such as spatial-temporal, spatial, slow-temporal, and fast-temporal streams. DSTA-Net was evaluated on two hand gesture recognition datasets, SHREC [229] and DHG [230], and two human action recognition datasets, NTU-60 and NTU-120. On the hand gesture recognition datasets, the accuracies were 91.5%/96.4% while the accuracies on the action recognition datasets were 86.6% and 89.0%, respectively. Liu et al. [231] proposed a new method for skeleton-based action recognition that disentangles and unifies graph convolutions which consists of a disentangled multi-scale aggregation scheme to remove redundant dependencies between features in different neighbourhoods, and a unified spatial-temporal graph convolution operator (G3D) to facilitate direct information propagation between joints in both spatial and temporal domains. The method was assessed on the NTU RGB+D 120, NTU RGB+D 60, and Kinetics Skeleton 400 datasets.
Kwon et al. [232] explored a new perspective on self-similarity for learning generalised motion representation by proposing a neural block called SELFY, based on spatial-temporal self-similarity (STSS). The authors validated the effectiveness of SELFY on four RGB benchmark datasets: Diving-48, Something-Something-V1 and V2, and FineGym datasets with an accuracy of 41.6%, 84.4%, 91.1%, 87.7%, respectively. Davoodikakhki et al. [233] introduced human action classification using hierarchical classification, network pruning, and skeleton-based pre-processing to improve model robustness and performance. The method was tested on the NTU RGB 60, NTU RGB 120, N-UCLA, and UTD Action Dataset. Video-Pose Network (VPN++) for action recognition is proposed by Das et al. [234] that combines RGB video and 3D pose data through knowledge distillation. VPN++ uses two levels of distillation: feature-level and attention-level, to improve overall HAR performance. The authors evaluated the method on multiple datasets, including Smarthome dataset [235], NTU60, NTU120, and N-UCLA datasets with an average accuracy of 71.0%, 90.7%, 86.2%, and 93.5%, respectively.
Video Swin Transformer, a new type of vision transformer architecture for video recognition that uses shifted windows to compute self-attention across video frames, allowing efficient and effective modelling of long-range temporal information, is introduced by Liu et al. [236]. Experiments on several video recognition benchmarks, including the Kinetics-400, Kinetics-600, and Something-Something V2 with an average accuracy score of 84.9%, 85.9%, and 69.6% demonstrate improved performance compared to previous convolutional and transformer-based approaches. A framework for video recognition using webly-supervised learning from multiple data sources named OmniSource was introduced by Duan et al. [237] which uses a task-specific data collection strategy and a unified framework to leverage various web data formats and perform activity recognition.
Attention mechanisms [25], for example, have allowed models to focus on the most relevant parts of the input data, leading to more accurate and efficient activity recognition. Graph-based networks, such as GCNs [21], have shown promising results in capturing spatial and relational dependencies in skeleton-based activity recognition. Additionally, transfer learning has emerged as a key technique [238], enabling models trained on large, pre-existing datasets to be fine-tuned for specific HAR tasks with limited data. As deep learning continues to evolve, novel architectures, such as attention-based transformers and reinforcement learning frameworks, are being explored to improve HAR’s performance further. These methods have shown potential in overcoming challenges such as data imbalance, noise, and complex activity patterns, making HAR systems more scalable and adaptable to real-world environments. Omnisource achieved 96.4%, 71.7%, and 83.6% accuracy on the UCF101, HMDB51, and Kinetics400 dataset. PoseConv3D is a novel framework for HAR that relies on a 3D heatmap volume instead of a graph sequence as the base representation of human skeletons is traduced by Duan et al. [34]. The authors tested the proposed method on five skeleton benchmark datasets achieving an average accuracy of 97.1%, 90.3%, 47.7%, 94.3%, 76.7%, and 55.6% on the NTU60, NTU120, Kinetics, FineGYM, HMDB51, and UCF101datasets, respectively. Table S11 in Supplementary Material S1 provides a summary of the various deep learning-based methods published in recent years, highlighting their performance across different modalities and datasets.

5.2.2. Attention-Based Networks

Attention-based networks, particularly transformers, have emerged more recently as highly effective methods for human activity recognition (HAR) across a variety of sensing modalities, including RGB, mobile phones, wearable sensors, depth, and skeleton data. Originally developed for natural language processing, transformers have been adapted for sequential data modelling in HAR due to their ability to capture long-range dependencies and focus on relevant portions of the input sequence [239]. Unlike traditional recurrent architectures, transformers use attention mechanisms that allow the network to weigh different parts of the input based on their relevance, making them highly effective for tasks involving complex and multivariate sensor data [240]. The self-attention mechanism in transformers dynamically adjusts the importance assigned to each part of the input, enabling the model to focus on salient patterns and discriminative information crucial for recognising activities [239]. This ability is particularly useful in HAR systems, where human movements can exhibit complex temporal dependencies. In comparison to other architectures, attention-based models excel in their capacity to handle long-term dependencies without suffering from issues like vanishing gradients, which is a common problem in Recurrent Neural Networks (RNNs) and Long Short-Term Memory (LSTM) networks [241].
A Two-Stream Inflated 3D ConvNet (I3D) based on 2D ConvNet inflation is introduced by Carreira et al. [242] which used RGB and flow for action recognition on complex datasets such as the UCF-101, HMDB-51, and Kinects dataset with accuracy scores of 97.9%, 80.2%, and 78.7%. Girdhar et al. [243] proposed a Video Action Transformer Network, a novel architecture that localises and recognises actions in video clips by attending to the entire video sequence. The authors evaluated the performance of the model using frame-level mean average precision (frame-AP) at an intersection of union (IOU) threshold of 0.5.
Li et al. [244] introduced GroupFormer, a clustered spatial-temporal transformer network for group activity recognition. In the first step, the model extract features for individual people in a video and then applies a clustering algorithm to group them. Subsequently, the grouped features are fed into a transformer network that learns to capture the interactions between members in each group and predict the group activity. The model was pre-trained on Kinetics [242] using the I3D model and evaluated on the Volleyball dataset [72] and the Collective Activity dataset [245], and achieved an average accuracy of 95.7%, and 96.3%, respectively.
The Video Transformer Network (VTN), a transformer-based system for video recognition, is presented by Neimark et al. [246]. The VTN uses a 2D Convolutional Neural Network as a backbone to extract spatial features. Then, a transformer network is used to model the temporal relationships between frames, enabling the classification of actions by attending to the entire video sequence. The model is evaluated on the Kinetics-400 and Moments in Time datasets for video action recognition. The results show competitive accuracy while exhibiting significant speed advantages during training and inference compared to state-of-the-art methods.
VidTr, a video transformer network that recognises video actions without the use of convolutions, is presented in this study [247]. The network takes raw video frames as the input and uses a transformer architecture with separable spatial and temporal attention to learn action representations. Six widely used datasets are used to evaluate the model. The datasets are the Kinetics-400, Kinetics-700, Something-Something V2, Charades [103], UCF-101 [99], and HMDB-51 [248]. VidTr achieved a state-of-the-art performance on these datasets, demonstrating its effectiveness in capturing long-term temporal dependencies for video action recognition.
The method [240] proposed by authors investigates the use of self-attention mechanisms for video comprehension. It presents TimeSformer, a transformer-based architecture that applies different spatio-temporal attention mechanisms to capture interactions between video patches. The authors investigated various attention schemes, including divided space–time attention, joint space–time attention, and sparse local attention. A number of video action recognition benchmarks, including Kinetics-400, Kinetics-700, Something-Something V2, and HowTo100M are used to assess the performance of the proposed model. The TimeSformer achieved competitive results on these datasets, demonstrating the effectiveness of self-attention for video comprehension.
A video transformer model with space–time mixing attention that aimed for efficient video recognition is presented in the study [249]. The model restricts time attention to a local window and utilises efficient space–time mixing to attend to spatial and temporal locations without significant overhead. Lightweight global temporal-only attention mechanisms are integrated for further accuracy improvements. Experiments on Something-Something V2, Kinetics, and Epic Kitchens [122] datasets show high recognition accuracy and improved efficiency compared to other video transformer models.
Motionformer, a video transformer architecture that introduces a novel trajectory attention mechanism for video action recognition is proposed by the authors in reference [250]. The trajectory attention effectively captures the motion patterns in videos by aggregating data along implicitly determined motion trajectories. Motionformer was evaluated on three video-based action recognition datasets: Kinetics-400, Something-Something V2, and Epic Kitchens-100. The model achieved comparable results, demonstrating the effectiveness of the trajectory attention mechanism for video understanding.
This study [251] introduced Multiscale Vision Transformers (MViT), a transformer architecture for video and image recognition. The MViT model incorporated multiscale feature hierarchies, starting with high spatial resolution and low channel capacity in early layers, and progressively increasing the channel capacity while reducing spatial resolution in deeper layers. Both low-level and high-level visual information are effectively captured by this architecture. The model is evaluated on the Kinetics-400, Kinetics-600, Kinetics-700, Something-Something V2, and ImageNet datasets. MViT achieves state-of-the-art performance on these datasets, showcasing its effectiveness in video and image recognition tasks.
A pure transformer-based model for video classification titled ViViT is proposed by Arnab et al. [252]. By adapting the success of transformer models in image classification, the proposed model used them to handle the spatio-temporal nature of video data. From input videos, ViViT extracts spatio-temporal tokens and encodes them using a series of transformer layers. The proposed method used an efficient model variant that factorises the spatial and temporal dimensions of the input to address the challenge of long token sequences in videos. ViViT is evaluated on multiple video classification benchmarks, including Kinetics-400 and 600, Epic Kitchens, Something-Something V2, and Moments in Time. The model achieved state-of-the-art results on datasets evaluated which outperformed previous methods based on deep 3D convolutional networks.
A novel visual representation learning approach, namely the TokenLearner, using a few adaptively learned tokens for image and video recognition is presented in this study [253]. Instead of relying on pre-defined splitting strategies or densely sampled patches, TokenLearner learns to extract important tokens from the input data. This efficient tokenisation representation captures long-range interactions in videos and effectively models spatial contents in images. TokenLearner was evaluated on the ImageNet, Kinetics-400, Kinetics-600, Charades, and AViD [254] datasets. The model achieved strong performance on these benchmarks demonstrating the effectiveness of adaptive tokenisation for visual representation learning.
This study [255] proposed a shifted chunk transformer (SCT) for learning spatio-temporal representations in video-based action recognition. Image chunk self-attention blocks and shifted multi-head self-attention blocks are the two components of the SCT model. The image chunk self-attention block is used to capture local spatial dependencies within a chunk, while the shifted multi-head self-attention block captures long-range dependencies across different chunks. The model is evaluated on the Kinetics-400 and HMDB datasets. The SCT achieved competitive performance compared to existing methods, demonstrating its effectiveness for spatio-temporal representation learning in video action recognition.
Unary-pairwise transformer (UPT) which is a novel framework for human–object interaction (HOI) detection was introduced by the authors in reference [256]. The UPT processes each human–object pair individually and integrates unary and pairwise representations for enhanced interaction understanding. The UPT architecture incorporated a novel attention mechanism that attends to relevant regions in both the human and object features. The authors evaluated the model on two benchmark datasets, namely the HICO-DET dataset and V-COCO dataset. The results show that the UPT achieves state-of-the-art performance on both datasets.
Uncertainty-Guided Probabilistic Transformer (UGPT), another novel method introduced for complex action recognition using probabilistic modelling to quantify uncertainty in predictions, is introduced by the authors of [257]. The UGPT models attention scores as Gaussian random variables, capturing stochasticity in both data and predictions. The authors also proposed a novel training and inference strategy guided by this uncertainty, using majority and minority models to improve prediction accuracy and robustness. The method achieves state-of-the-art performance on the Breakfast Actions, MultiTHUMOS, and Charades datasets.
RegionViT [258], a novel vision transformer architecture, incorporated regional-to-local attention to enhance performance and efficiency in various computer vision tasks. Similar to Convolutional Neural Networks (CNNs), RegionViT utilised a pyramid structure to extract features at multiple scales. By limiting its scope while preserving the capability to capture global information, the regional-to-local attention mechanism reduced the computational burden of self-attention. The authors carried out experiments on multiple datasets including the ImageNet, CIFAR10, and CIFAR100 datasets. RegionViT achieved comparable performance across object detection and action recognition.
A novel video transformer architecture called RViT is introduced in this paper [241]. The proposed method used recurrent units to effectively process videos for action recognition. RViT incorporated a specifically designed attention gate mechanism to aggregate temporal attention recursively which processed both the current input frame and the hidden state from the previous frame, allowing the model to capture long-term temporal dependencies without being limited by video length. RViT operates on a frame-by-frame basis. The effectiveness of RViT is demonstrated through experiments on four benchmark datasets for human action recognition: Kinetics-400, Jester, Something-Something V2, and Charades.
DirecFormer is introduced in this work [259], which is also a transformer-based action recognition framework that incorporated a novel Directed Attention mechanism to address the issue of ordered temporal learning in videos. DirecFormer aimed to ensure that the model learns the correct order of actions in a video sequence, rather than being biased by background scenes or spurious correlations. The Directed Attention mechanism provides attention to human actions in the proper order, leading to improved robustness and generalisation. Experiments were carried out on the Jester, Something-Something V2, and Kinetics-400 datasets and DirecFormer achieved state-of-the-art results.
This paper [260] proposed UniFormer, a novel transformer architecture for efficient spatio-temporal representation learning in videos. UniFormer integrated 3D convolution and spatio-temporal self-attention within a unified transformer framework. The model utilised local multi-head relation aggregation (MHRA) in early layers to reduce computational complexity and global MHRA in deeper layers to capture long-range dependencies which addresses the redundancy and dependency issues inherent in video data. The model was evaluated on Kinetics-400/600 and Something-Something V1/V2. The results demonstrate that UniFormer achieves a good balance between accuracy and efficiency.
This paper [261] explored using image classifiers for the task of action recognition, challenging the traditional approach of using specialised temporal models. The methodology involves rearranging video frames into a “super image” based on a predefined spatial layout. This super image is then fed into an image classifier, treating action recognition as an image classification task. The paper focuses on action recognition and evaluates the approach on datasets such as Kinetics-400, Moments in Time, Something-Something V2, Jester, and Diving48. The authors experimented with Swin Transformer, a vision transformer model, and achieved state-of-the-art results, outperforming existing methods while being computationally efficient.
A novel Multiview Transformers approach to video recognition utilising multiple views of the input video to capture multi-resolution temporal context is presented in this work [262]. The model processed multiple temporal segments of the video in parallel, each at a different resolution, which effectively captured short-term and long-term temporal dependencies. This multiview strategy allowed the model to learn rich representations of the video content, leading to improved performance compared to single-view transformer architectures. The authors evaluated Multiview Transformers on six popular video classification datasets: Kinetics-400, Kinetics-600, Kinetics-700, Something-Something-V2, Moments in Time, and Epic Kitchens.
A self-supervised learning approach called Self-supervised Video Transformer (SVT) for training video transformers on unlabelled video data is presented in this study [239]. This method involves creating local and global spatio-temporal views of a video with varying spatial sizes and frame rates. A self-supervised objective is then used to match features from these diverse views, ensuring invariance to spatio-temporal variations in actions. The effectiveness of SVT is evaluated on four action recognition benchmarks: Kinetics-400, UCF-101, HMDB-51, and SSv2. The results show strong performance, particularly the ability to converge quickly with small batch sizes.
VideoMAE [263] is another a self-supervised video pre-training (SSVP) method based on masked autoencoders, specifically designed for video transformers. It employed an asymmetric encoder-decoder architecture where a high proportion of video cubes are masked, and the model is tasked with reconstructing these missing cubes. VideoMAE is evaluated on datasets such as Kinetics-400, Something-Something V2, UCF101, and HMDB51. The method achieves impressive results, particularly in data-efficient settings.
An extension of Masked Autoencoders (MAEs) for learning spatio-temporal representations from videos is presented in this work called the ST-MAE [264]. The approach, similar to ImageMAE, involves randomly masking out spacetime patches in videos and training an autoencoder to reconstruct them in pixels. The method is evaluated on action recognition benchmarks, including Kinetics-400, Kinetics-600, and Something-Something V1 and V2. The results demonstrate superior performance over supervised pre-training, achieving, for example, 82.9% top 1% accuracy on Kinetics-400 with ImageNet-1K pretraining.
Long-Short Temporal Contrastive Learning (LSTCL), a self-supervised pre-training technique for video transformers that leverages the models’ capability to capture long-range dependencies, is introduced in this work [265]. LSTCL creates positive pairs by contrasting the representation of a short video clip with that of a longer clip from the same video. The method encourages the model to learn clip-level representations that incorporate a broader temporal context. This paper focused on action recognition and evaluated LSTCL on datasets such as Kinetics-400, Kinetics-600, SomethingSomething V2, UCF101, and HMDB51, demonstrating performance on par with or surpassing supervised pre-training on large image datasets.
A novel approach for pre-training video transformers using a BERT-style objective named BEVT is introduced in this work [266], aiming to learn discriminative video representations. BEVT utilised a two-stream network, separating spatial representation learning and temporal dynamics learning, allowing the model to effectively capture both types of features. The authors investigated the importance of spatial and temporal clues in video recognition, highlighting their varying influence across different video samples. BEVT is evaluated on three challenging video benchmarks: Kinetics-400, Something-Something V2, and Diving48 and achieved satisfactory scores.
OmniMAE, a novel masked autoencoder framework for single model pre-training on images and videos, is proposed by Girdhar et al. [267]. OmniMAE utilised a unified architecture and training procedure for learning representations across multiple visual modalities that facilitated efficient knowledge transfer and generalisation for activity learning. The model leverages the complementary nature of image and video data by jointly training on ImageNet (IN1K) and Something-Something V2 (SSv2) datasets. This method allows the model to capture a wide range of visual concepts, including static appearance and dynamic motion patterns. The effectiveness of OmniMAE is demonstrated through transfer learning experiments on image datasets (iNaturalist-2018 and Places-365) and video datasets such as Kinetics-400 and EPIC-Kitchens-100.
This paper [268] introduced Masked Motion Encoding (MME) which is a novel self-supervised learning framework for video representation learning. This method masks out portions of motion trajectories within videos and tasks the model with predicting the missing motion information. This approach encourages the model to learn robust and discriminative representations of motion patterns, which are essential for various video understanding tasks. The effectiveness of MME is evaluated through transfer learning experiments on action recognition benchmarks, such as Kinetics-400 and Something-Something V2. The results demonstrate that MME pre-training significantly improves the performance of action recognition models compared to other self-supervised approaches.
A motion-guided masking algorithm (MGM) to enhance the effectiveness of masked autoencoders (MAE) for spatio-temporal representation learning in videos is suggested by Fan et al. [269]. Instead of random masking strategies used in image MAEs, MGM used motion vectors obtained from compressed video formats to dynamically mask spatio-temporally salient regions in videos. This method was evaluated on Kinetics-400 and Something-Something V2 datasets and showed significant improvements over existing video MAE approaches. Furthermore, MGM exhibits faster convergence and superior generalisation capabilities in downstream transfer learning and domain adaptation tasks on the UCF101, HMDB51, and Diving48 datasets.
One key advantage of attention-based networks is their versatility across different sensor modalities [258]. By dynamically attending to relevant sensor readings or visual features, transformers can effectively integrate information from diverse sources such as combining RGB camera data with depth and skeletal readings to achieve a more comprehensive understanding of human activities [269]. This multimodal fusion enhances recognition accuracy and robustness in real-world HAR applications, making transformers particularly well-suited for complex environments, including healthcare monitoring, autonomous driving, and human–computer interaction.
As shown above, recent research in attention-based models has led to several innovations in architecture design. Variants such as the Vision Transformer (ViT), TimeSformer, Video Swin Transformer (Swin-T), masked autoencoders (MAE), Masked Motion Encoding (MME), and motion-guided masking algorithm (MGM) have demonstrated exceptional performance in HAR tasks, particularly in video-based human activity datasets. These models have shown superior generalisation capabilities and scalability, making them an attractive choice for HAR systems that require real-time processing and the ability to adapt to various contexts. In addition to exploring new attention mechanisms, researchers are also investigating the benefits of transfer learning to leverage pre-trained models and fine-tune them for specific HAR tasks. By building on large, pre-existing datasets, HAR systems using transformers can achieve faster convergence and improved performance, even when dealing with smaller, domain-specific datasets. This makes transformers an attractive option for applications where collecting large-scale labelled data is challenging. Ongoing efforts are focused on optimising transformer architectures for computational efficiency, as transformers are known for their high computational and memory demands. Methods such as efficient attention mechanisms and lightweight transformer variants aim to make these models more accessible for real-time and resource-constrained HAR applications. Table S12 in Supplementary Material S1 summarises recent transformer-based models for HAR, showcasing their performance on popular datasets like Kinetics-400 (K400), UCF101, HMDB51, and Something-Something V2 (SSV2).

5.2.3. Reinforcement Learning-Based Networks

Human activity recognition (HAR) is a complex research area that focuses on the identification and recognition of movements or activities performed by individuals. Although people can quickly identify actions they see in videos, fully automating this process is challenging, yet essential given its wide range of real-world purposes [270]. HAR is presented as a learning problem in the literature on computer vision, and machine learning techniques are used to categorise and identify the activities. Prior to the introduction of deep learning (DL) models in activity detection methods, researchers primarily focused on using manually extracted features such as motion descriptors and interest points [271,272] for HAR. DL techniques were first applied to the field to extract efficient video representations and increase generalisation for video-based HAR as hand-crafted feature extraction is time-consuming and requires extensive domain knowledge. This led to a notable advancement in the field.
In contrast to supervised learning techniques, reinforcement learning (RL) is a branch of machine learning techniques that learns in an unsupervised manner by having one or more agents interact with the surroundings [273]. Depending on a pre-determined reward function, these interactions produce either positive or negative rewards. Without being aware of the ground truth, the agent learns how to maximise the predicted reward through trial and error as it acquires additional information from the environment. Deep Reinforcement Learning (DRL), a novel family of algorithms whose efficacy has been shown in numerous applications, was created more recently by combining reinforcement learning with deep learning [274,275]. Reinforcement Learning (RL) and Deep Reinforcement Learning (DRL) methods represent a growing frontier in human activity recognition (HAR) [270], offering innovative ways to address dynamic and complex activity recognition tasks across various sensing modalities, including RGB, mobile phones, wearable sensors, depth, and skeleton data. In contrast to traditional supervised learning models, that depend on fixed datasets for training, RL-based methods learn optimal decision-making policies by interacting with an environment [270]. This adaptability makes RL particularly well-suited for HAR scenarios where human activities are continually evolving, and where real-time feedback can help improve model performance.
In RL-based HAR, the system is modelled as an agent that learns to recognise activities by taking actions and observing outcomes. The agent receives rewards or penalties applied based on the accuracy of its predictions [273]. Over time, the agent learns to optimise its activity recognition strategy through a trial-and-error process by making the RL-based method a promising approach for tasks that involve sequential decision-making. It is particularly beneficial for situations such as continuous monitoring of activities and autonomous systems. RL-based methods can be effective for real-time interventions in healthcare settings [274]. Deep Reinforcement Learning (DRL) which combines RL and deep learning methods extends this paradigm by using deep neural networks to handle high-dimensional data inputs, such as video streams or multivariate sensor data, which are commonly used in HAR systems. DRL models are particularly effective in HAR applications that analyse complex spatio-temporal patterns [270]. DRL methods can be applied to recognise activities in crowded environments where existing methods may fail. DRL methods adapt to changes in sensor configurations and learn new activities over time without manual intervention which make it a powerful tool for enhancing the robustness of HAR systems.
A spatio-temporal deep Q-network (ST-DQN) for human activity localisation is presented by the authors in reference [276], which consist of recognising activity category labels by identifying their spatio-temporal locations in video sequences. The model is an integrated framework that uses temporal dependencies and contextual information to improve localisation accuracy while concurrently addressing spatial and temporal localisation. With a limited number of proposals, the ST-DQN achieves promising localisation performance. By incorporating temporal dependencies, it outperforms previous frame-based localisation methods, achieving higher mean average precision (mAP) scores across different intersection-over-union (IoU) thresholds on challenging datasets like UCF-Sports [277], sub-JHMDB [278], and UCF-101 [99].
Deep progressive reinforcement learning (DPRL) which is a new technique for skeleton-based action recognition, is suggested by the authors’ research presented in reference [279]. The proposed method concentrated on selecting the most relevant frames from a skeleton sequence while it eliminated ambiguous frames that effectively increased action accuracy. The authors used deep reinforcement learning to model frame selection as a progressive process. The quality of the nominated frames for action recognition and their association to the entire video sequence were the two main factors that the DPRL method used to iteratively update the frame selection. Additionally, the study employed a Graph-based Convolutional Neural Network (GCNN) that captured the inherent graph-like structure of human body joints. In GCNN, vertices represent joints and edges depict the relationships between them. The DPRL method achieved 79.6%, 98.5, and 89.8% accuracy on the SYSU-3D [280], UT-Kinect [64], and NTU RGB+D 60 [105] datasets, respectively, which were used for skeleton-based action recognition.
This study [281] proposed a multi-agent reinforcement learning (MARL) framework for untrimmed video-based action recognition. The framework focused on developing a learning-based frame sampling strategy to enhance both the accuracy and efficiency of video classification tasks, especially for untrimmed videos where frame-level saliency varies significantly. The paper considered frame sampling as numerous simultaneous Markov decision processes, each of which is tasked with identifying an informative frame or clip by improving an initial sample point, as an alternative to hand-crafted sampling techniques. The authors [281] employed a novel RNN-based context-aware observation network to model contextual information between nearby agents and incorporate historical states of individual agents. The MARL is comprised of this network, along with a policy network that determines action probabilities and a classification network for reward calculation and final recognition. The authors used ActivityNet [101] v1.2 and v1.3 for an action recognition task achieving a mAP score of 64.13% and 62.00%, respectively.
A novel framework for effective human activity classification in egocentric videos which focused on reducing computational costs while maintaining accuracy is introduced in this study [282]. The approach combined actor-critic reinforcement learning (RL) with a deep-shallow network architecture to identify and process regions of interest (ROI) within video frames. The scope of the work is to efficiently classify human activities captured from a first-person perspective using wearable cameras. The authors evaluated their approach on the Dataset of Multimodal Semantic Egocentric Video (DoM-SEV), focusing on five activities, which are walking, running, standing, being in conversation, and browsing. In comparison to a traditional 3D CNN (C3D), the results demonstrated a 36.4% reduction in processing time while achieving similar accuracy.
A unique method for human activity recognition (HAR) that uses a camera mounted on a mobile robot is presented in the author’s work [283]. This research aimed to improve HAR accuracy by optimising the robot’s position and viewing angle through Deep Reinforcement Learning (DRL) while reducing energy consumption related to the robot’s movement. The key idea was to train the robot to dynamically adjust the position of the robot based on the observed activity and its confidence in recognising the activity. The study used a virtual environment called the HoME Platform [284], to evaluate the proposed method. Twelve activities were conducted by virtual actors in the simulated environment and subsequently used for classification which achieved 75% accuracy.
A deep learning framework [285] for location- and person-independent activity recognition using WiFi signals was suggested in the work. The objective of the work was to develop a robust system that recognises activities accurately regardless of the position and orientation of WiFi devices or the identity of the person who is performing the activity. Given that WiFi signals are highly sensitive to environmental factors and individual variations, The proposed method addressed a key challenge in WiFi-based activity recognition. The authors used three Deep Neural Networks (DNNs), to recognise activities that are location- and person-independent. Firstly, a 2D CNN was used to extract location- and person-independent features from different perspectives of the CSI data. Secondly, 1D CNN captured temporal dependencies between successive CSI segments. Finally, a neural architecture search combined RNN with LSTM which acts as a reinforcement learning agent, responsible for optimising the neural architecture of the recognition algorithm. The Yousefi-2017 [286] and FallDeFi [287] datasets were used to evaluate the performance of the proposed method.
In order to identify action classes that have not been observed during training, this study [288] introduced CLASTER, a novel approach for zero-shot action recognition. The authors suggest that generalisation to unseen action categories may be hampered by the precise decision boundaries that typical neural networks acquire in supervised environments. To overcome this, CLASTER makes use of a cluster-based representation that encourages the formation of a representation that successfully generalises to new action cases while also regularising the learning process. Instead of using individual data points, CLASTER aims to learn a more general representation that is less susceptible to the variations present in individual instances by using cluster centroids. Reinforcement Learning (RL), more specifically the REINFORCE algorithm, is used to optimise the clustering. This allows direct optimisation of the cluster centroids based on the classification performance. The UCF101 [99], HMDB51 [248], and Olympic Sports [289] datasets were used for performance evaluation.
One of the key advantages of RL and DRL in HAR is their ability to learn from interactions with an environment, enabling systems to adjust to new activity patterns or sensor inputs without the need for extensive retraining. This makes RL-based HAR systems highly adaptable and scalable across different environments and applications, such as smart homes, autonomous driving, and healthcare monitoring. However, there are several challenges associated with RL-based HAR. The exploration of large state-action spaces in complex environments can require significant computational resources and time. Additionally, ensuring safety during the learning process is crucial, especially in applications like healthcare or autonomous systems, where incorrect decisions could have serious consequences. To address these issues, researchers are exploring various techniques, including reward shaping, transfer learning, and simulation-based training environments, to improve the efficiency and safety of RL-based HAR models. Despite these challenges, RL and DRL are gaining traction in HAR research, with ongoing developments in algorithmic efficiency, policy optimisation, and hybrid methods that combine RL with other machine learning paradigms to enhance performance. As RL techniques continue to evolve, they are expected to play a more prominent role in creating adaptive and autonomous HAR systems capable of handling complex, real-world scenarios. A summary of HAR methods using RL and DRL-based techniques is presented in Table S13 in Supplementary Material S1.

5.3. Machine Learning Algorithm Performance Comparison

Comparing the performance of various machine learning algorithms in human activity recognition (HAR) is crucial to identify the most effective approaches for specific application scenarios. The selection of algorithms often depends on factors such as accuracy, computational complexity, and scalability, which vary depending on the data modality and the target application. Each machine learning method brings its own strengths and limitations in addressing the diverse and dynamic nature of human activities.
Supervised learning algorithms like Support Vector Machines (SVM), Decision Trees (DT), Random Forests (RF), and K-Nearest Neighbours (K-NN) often achieve high accuracy, especially when most effective features are used and trained on large, well-labelled datasets, particularly in environments with clearly defined activity patterns as shown in the investigation [76], but these algorithms may encounter challenges when applied to recognising more complex, multi-step activities [290]. In contrast, semi-supervised learning techniques, which combine both labelled and unlabelled data, offer an effective alternative in scenarios where labelled data are scarce. However, these methods generally perform slightly lower than fully supervised approaches due to the challenges of effectively utilising unlabelled data [290].
On the other hand, deep learning methods, including CNNs and RNNs, have consistently outperformed traditional approaches due to their ability to automatically extract meaningful features from raw data, such as images or acceleration [291]. CNNs excel in processing spatial data like RGB images, while RNNs are particularly suited for temporal sequences, such as time-series data from wearable sensors, or skeletal joint movement data allowing superior activity recognition. Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) represent a significant advancement in HAR [215]. By automatically learning hierarchical and discriminative features from raw sensor data, these models achieve superior accuracy and generalisation across complex datasets. Attention-based networks and transformers further enhance accuracy by focusing on the most critical features within input sequences [292] by improving both interpretability and recognition performance [293]. By dynamically weighting relevant information, these networks improve the model’s ability to recognise complex activities with greater precision.
Reinforcement Learning (RL) methods, while less common in HAR, offer a unique advantage by enabling adaptive decision-making based on feedback from the environment [285]. Although Reinforcement Learning (RL) methods can also achieve high accuracy, they often require extensive exploration to optimise performance. RL-based approaches require extensive exploration and careful reward design, as they may present safety concerns in real-world applications where incorrect decisions could have serious consequences. RL-based HAR systems iteratively improve over time but may take longer to reach peak accuracy compared to supervised or deep learning models. To evaluate the performance of HAR methods, different metrics are used. In the section below, various evaluation metrics that are found in the literature used by authors to evaluate different methods are described.

5.3.1. Evaluation Metrics

In the context of human activity recognition (HAR), the terms True Positive ( T p ), False Positive ( F p ), False Negative ( F n ), and True Negative ( T n ) are used to evaluate the performance of a classification model. True Positives refer to the number of times an HAR model correctly predicts the activity as the correct class. False Positive indicates the number of times a model incorrectly classifies an activity as a certain class when it was actually a different activity [294]. False Negative signifies the number of times the method has failed to predict the correct activity and instead predicted a different activity. True Negatives denote the number of times an HAR classifier correctly identified that a certain activity did not occur. In the section below, commonly used metrics for HAR evaluation are listed with brief description.
  • Accuracy
As seen in the literature [41,190], accuracy is a widely used metric for evaluating the effectiveness of machine learning models which are also adapted to assess the performance of human activity recognition (HAR) systems. In the context of HAR, accuracy refers to the proportion of correctly identified activity samples compared to the total number of activity samples, across all classes and classes. Accuracy is defined as [294]
A c c u r a c y = N u m b e r   o f   C o r r e c t   P r e d i c t i o n s T o t a l   N u m b e r   o f   P r e d i c t i o n s = T p + T n T p + T n + F p + F n
where T p , T n , F p , and F n are True Positive and Negative, and False Positive and Negative, respectively.
  • Precision
Precision quantities the relationship of properly identified positive sample (True Positives) out of all the positive predictions, reflecting HAR model’s ability to minimize False Positives. Precision is defined as [294]
P r e c i s i o n = T p T p + F p
  • Recall (Sensitivity)
Recall or sensitivity measures the ratio of True Positive predictions out of all actual positive instances in the dataset. Recall is defined as [294]
P r e c i s i o n = T p T p + F n
  • F1-Score:
F1-Score measures the harmonic mean of precision and recall, providing a balance between the two. It is particularly useful when the classes are imbalanced. F1-Score is defined as [294]
F 1 S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
  • Confusion matrix:
A confusion matrix or error matrix is a table that provides a summary of the performance of a classification algorithm by showing the error made by the classifier including the True Positive, True Negative, False Positive, and False Negative values.
  • Intersection-over-Union (IoU):
IoU, also called Jaccard index, determines the accuracy of the detector on a specific dataset. When classifying human activity using sensors or video data, IoU can be treated as the overlap between predicted activity intervals (the time where a particular activity was recognised) and true activity intervals (the actual time when the activity occurred). IoU is defined as [295]
I o U = A r e a   o f   o v e r l a p A r e a   o f   u n i o n
  • Mean Average Precision (mAP):
In the context of information retrieval, object detection, and multi-label classification tasks, mean average precision (mAP) is a commonly used evaluation metric and it is also applied to evaluate vision-based HAR systems. Average Precision (AP) refers to the average precision scores estimated at different recall levels for a particular activity class. Average Precision is defined as [295]
A P = 1 N i = 1 N P r e c i s i o n   ( r i )
where N is the number of recall levels, and Precision ( r i ) is the precision at recall level r i .
  • Receiver Operating Characteristic (ROC) and Area Under the Curve (AUC):
ROC is popular graphical evaluation metric used in machine learning which is also applied in the evaluation of HAR systems. True Positive Rate (TPR) which measures the percentage of actual positive instances that are correctly identified by the model and False Positive Rate (FPR) which measures the amount of actual negative instances that are incorrectly classified as positive by the model are used to plot the ROC curve. Typically, TPR is plotted in the X-axis and FPR is plotted in the X-axis. TPR and FPR are calculated as [296]
T P R = T p T p + F n
F P R = F p F p + T n
The Area Under the Curve (AUC) refers to the area under the ROC curve which provides a single scalar value between 0 and 1 to summarise the performance of a classification model across all possible thresholds. The model’s ability to differentiate between positive and negative classes is measured by its AUC.
  • Internal Cluster Evaluation Metrics
When class labels are not available for the HAR dataset, typically unsupervised learning methods are applied to uncover the patterns within the data, so clustering methods such as K-means, Spectral, Hierarchical, and density-based clustering are often applied for that purpose [297]. Internal metrics assess the quality of clustering based purely on the clustering results themselves without using ground truth labels [83].
  • Silhouette Score
Silhouette Score is a frequently used matrix that evaluates a data point’s similarity to its own cluster in relation to other clusters. It defined as [137]
S i l h o u e t t e   s c o r e = b a     m a x ( a , b )
where a represents the mean distance between a data point and every other data point within the same cluster and b is the average distance from the point to all points in the nearest cluster.
  • Davies–Bouldin Index (DBI)
The Davies–Bouldin Index (DBI) calculates the average similarity ratio between each cluster and the cluster that is most comparable to it; typically, a lower DBI indicates better clustering. DBI is calculated as [137]
D B I = 1 k   i = 1 K m a x σ i + σ j d i j   w h e r e   i j
where K is the number of clusters. σ i is the average distance between samples in cluster i. d i j is the distance between the centroids of clusters i and j.
  • Dunn Index
The minimal inter-cluster distance divided by the maximum intra-cluster distance is measured by the Dunn Index. Better clustering is indicated by a higher Dunn Index. The Dunn Index is given by [137]
D u n n   I n e d x = m i n i j d i s t a n c e ( C i , C j ) m a x k d i s t a n c e ( C k )
where C i , C j are clusters i and j d i s t a n c e ( C k ) is the maximum distance between the points in cluster k.
  • External Evaluation Metrics
The clustering results are compared with ground truth labels (where available) using external measures. Expert-labelled activity data may be used to establish ground truth labels in unsupervised HAR, or heuristics or assumptions may be used to approximate them in the absence of expert labelling.
  • Adjusted Rand Index (ARI):
ARI evaluates the similarity of two clustering results to one another (e.g., unsupervised clusters vs. ground truth). It corrects for chance, which means it takes into consideration the possibility that some apparent resemblance could result from random clustering. ARI ranges between −1 and 1: 1 means perfect clustering, 0 means clustering is better than random clustering, while −1 denotes the worst clustering. ARI is defined as [137]
A R I = R I E [ R I ] max R I E [ R I ]
where RI is the Rand Index and E[RI] is the Expected Rand Index.
  • Normalised Mutual Information (NMI):
NMI evaluates the quantity of information that is shared between the ground truth and the clustering results. To guarantee that the outcome falls between 0 (no mutual information) and 1 (perfect agreement), it is normalised. It is defined as [137]
N M I = I ( U , V ) H ( U ) H ( V )
where the mutual information between the clustering and the ground truth is denoted by I(U, V). H(U) and H(V) represents the entropies of the clustering and the ground truth, respectively.
  • Fowlkes–Mallows Index (FMI):
The geometric average of the pairwise precision and recall between the ground truth and clustering is measured by FMI. Typically, it ranges between 0 and 1, with 1 signifying ideal clustering and values close to 0 mean bad clustering. FMI is calculated as [137]
F M I = T p ( T p + F p ) ( T p + F n )

5.3.2. Computational Complexity

Computational complexity refers to the number of computational resources—such as time, memory, and processing power—required to train and deploy machine learning algorithms for human activity recognition (HAR). This factor is particularly important in real-time or resource-constrained environments where processing must occur quickly and efficiently [282]. Machine learning models, such as SVM, KNN, and Decision Trees, tend to have lower computational complexity compared to deep learning methods. These models are generally faster to train and deploy, making them well-suited for applications where real-time processing or limited computational resources are a concern. However, as datasets grow in size or dimensionality, the time required to train these models can increase significantly, potentially limiting their scalability for large-scale HAR tasks. In contrast, deep learning models, including CNNs and RNNs, are computationally much more demanding. These models involve multiple layers and a large number of parameters, which makes training and inference resource-intensive [282]. Deep learning models often require GPUs or cloud-based resources to manage the high computational load, particularly when handling large-scale HAR datasets or complex activity patterns.
Attention-based methods, such as transformers, further increase computational complexity [282]. These methods calculate attention weights for each element in the input sequence, which can slow down both training and inference. Although more efficient attention mechanisms have been introduced in recent transformer architectures, attention-based models still require significant computational power to function effectively [282]. Reinforcement Learning (RL) methods are often the most computationally complex. RL models must explore large state-action spaces to learn optimal strategies for recognising human activities, a process that is both time-consuming and resource-intensive [281]. Training an RL agent in environments with high-dimensional data or complex activity patterns often demands substantial computational resources, making RL methods challenging to implement in real-time HAR systems without sufficient computational power. Methods such as AttentionClusters [281] may be explored for reducing complexity.

5.3.3. Scalability and Generalisation

Scalability refers to the ability of a machine learning algorithm to efficiently handle increasing amounts of data, complexity, or computational resources without significant performance degradation [298]. In the context of human activity recognition (HAR), scalability determines whether a model can adapt and perform effectively as the size of datasets, number of activity classes, or complexity of sensor environments grows [187].
Supervised learning algorithms, while typically effective for smaller datasets or less complex HAR tasks, may struggle with scalability when applied to larger or more diverse datasets [299]. As data volume increases, the time and computational resources required for training these models can grow exponentially. However, advancements in parallel processing and distributed computing, particularly in cloud environments, have enabled supervised learning algorithms to scale more effectively, making them viable for large-scale HAR applications [297]. Deep learning models, particularly Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs), are designed to scale well with large datasets due to their ability to learn complex hierarchical representations [291]. Their scalability allows them to capture intricate patterns in human activities from vast amounts of data, such as video sequences or time-series sensor data. However, this scalability comes at the cost of substantial computational resource requirements, limiting their deployment in environments with constrained hardware. Techniques like transfer learning [299] and model compression are increasingly used to improve the scalability of deep learning models, allowing them to function in resource-limited settings without significantly compromising accuracy.
Attention-based networks, such as transformers, can scale to larger datasets as the amount of the data and model grow, but they often encounter issues related to interpretability and increased computational complexity [292], making it harder to understand how the model is making decisions. Efforts are underway to optimise these architectures for larger-scale HAR applications, focusing on more efficient attention mechanisms that can handle increased data complexity without incurring excessive computational costs. Reinforcement Learning (RL) presents unique scalability challenges in HAR, particularly when the state-action space becomes too large [281]. RL algorithms must explore vast environments or datasets to learn optimal decision-making policies, which can make them computationally expensive and time-consuming to train. However, RL models can scale through efficient exploration and optimisation strategies [281]. In practice, this makes them more suitable for controlled environments or applications with well-defined activity patterns, as their scalability is often limited by the complexity of the environment and the availability of computational resources for extensive exploration [282].

6. Applications of Human Activity Recognition

Across a wide range of domains, human activity recognition (HAR) technology has proven its versatility and effectiveness turning it into a crucial tool for addressing various real-world problems [300]. The ability to effectively monitor, analyse, and recognise human actions using data from different types of sensors has opened opportunities for improving human lives. Fields such as safety, performance optimisation, enhancing everyday living [62], healthcare [301], sports tracking and analysis [302], security and surveillance, and smart environments [303] are among the many sectors where HAR has found practical applications.
Successful implementations of HAR systems across these domains showcase the broad applicability of this technology. For example, HAR systems can significantly improve fall detection and provide immediate response [304] in places like elderly care or individuals who are living alone. In sports, HAR can be used for performance monitoring and enhanced athletic training programs to facilitate training and prevent injury. Similarly, HAR systems can be deployed for intrusion detection and behaviour monitoring in public spaces to strengthened security measures. These are examples of real-world applications of HAR that emphasise the potential of the technology across a variety of fields, where it continues to advance rapidly.
The applications discussed above are not exhaustive. Researchers are continuously exploring new areas where HAR can be applied, especially as advancements in sensing modalities, learning algorithms, and computational power allow complex activity to be recognised with improved accuracy. From autonomous vehicles to workplace safety and beyond, HAR is poised to expand its impact, fostering innovative solutions to emerging challenges across a multitude of industries. In the section below, some of the HAR application are briefly described.

7. Challenges and Future Directions

As human activity recognition (HAR) continues to evolve, the research landscape is increasingly shaped by emerging trends and ongoing advancements in this field [91]. All the upcoming technologies such as multimodal sensor fusion, edge computing, IoT-based systems, and personalised adaptive models are becoming pivotal in expanding the scope and effectiveness of HAR applications. The increase in focus towards growing attention to ethical and social considerations are driving HAR towards more user-centric solutions.
However, the field faces several challenges that must be addressed for HAR systems to perform optimally. Technical hurdles introduced by sensors, limited availability of data segmentation methods, and model interpretability and transparency are needed for that purpose. Ensuring that HAR models can be generalised across diverse environments and transferred to new applications is crucial for scalable implementation. Moreover, human-centric design and evaluation frameworks are necessary to ensure HAR technologies align with user needs and societal expectations. The future of HAR research lies in addressing these challenges while embracing technological advancements, enabling more efficient, ethical, and human-centred systems.

7.1. Emerging Trends and Future Directions in Human Activity Recognition Research

Researchers are increasingly exploring innovative approaches to tackle current challenges and unlock new opportunities in human activity recognition (HAR). Future directions in HAR are focused on enhancing its accuracy and applicability across a variety of contexts. These approaches also take into consideration the ethical implications of widespread adoption.
One notable advancement within this domain is that of multimodal sensor fusion, in which HAR systems utilise data from different sensor modalities, such as RGB cameras or inertial and environmental sensors, to gain a better understanding of human activities [305,306,307]. Traditional HAR systems like the sensor-based and vision-based systems often face limitations due to factors like occlusion. Changes in lighting condition and environmental variability leads to lower accuracy. By combining data from multiple sensors, future HAR systems may offer improved accuracy and increase the reliability of HAR systems. The fusion of sensors are expected to overcome some of the existing challenges and allow fine-grained activity recognition across a variety of settings [308,309]. Multimodal sensor fusion techniques facilitate comprehensive analysis of human actions that capture complementary data from different sources to enhance the overall performance of HAR systems [309].
Another emerging trend is the integration of edge computing and Internet of Things (IoT)-based systems. These technologies capture and process sensor data directly at the edge of the network. Edge devices reduce latency and improve privacy as they do not need to send data to any centralised servers [310]. This is particularly useful in HAR applications that require real-time decision-making, such as healthcare monitoring or smart home systems. Edge computing also helps to conserve computational resources, leading to more scalable and efficient HAR deployments in IoT environments [310].
Adaptive models represent another area of growth in HAR research. Traditional HAR models often struggle with generalisation as they are designed to identify activities based on a fixed dataset. Those fixed models do not consider the variations of individual differences in activities. Future research aims to develop models that can adapt to the unique habits, preferences, and contexts of individual users [311,312]. By incorporating techniques such as reinforcement learning and user feedback, these personalised models may offer user-specific activity recognition that dynamically adjust to changes in activities and environments [214].
Ethical and social considerations are becoming increasingly important as HAR systems are gradually integrated into daily life. As these systems process sensitive data about human actions, privacy concerns and data security comes forward. Algorithmic bias and transparency of the model interoperability have come to the forefront [313]. Future HAR research will need to prioritise these issues to ensure that the technology is deployed responsibly considering data safety that protects the privacy of users [314]. Addressing these ethical and social challenges is essential for building trust to ensure the widespread adoption and acceptance of HAR technologies in various sectors.
These trends illustrate how future HAR research is moving towards creating more accurate, adaptable, and ethically responsible systems that can be applied across a broad spectrum of real-world scenarios.

7.2. Research Challenges and Potential Solutions

Despite significant advances in human activity recognition (HAR), numerous challenges persist that hinder the full realisation of its potential, particularly as HAR applications increasingly leverage artificial intelligence (AI), machine learning (ML), and deep learning (DL) methods. A wide range of obstacles, such as intraclass variations, view changes, meaningful feature extraction, the availability of large-scale datasets, multi-occupant activity recognition, computational costs, and privacy concerns, continue to pose challenges to the field, as identified in various surveys [315].
One prominent challenge is related to HAR modalities. Vision-based approaches, while the least intrusive, are sensitive to environmental factors and equipment limitations [305]. For instance, RGB camera-based HAR methods provide rich visual information but are highly susceptible to issues such as illumination variations and view changes, making it difficult to maintain consistent recognition accuracy across different environments. Additional challenges arise from intraclass variations, where activities that should be recognised as the same category exhibit different characteristics (e.g., falls that differ in their visual representation), self-occlusion, and the need to recognise activities that are differentiated by subtle movements. These challenges are illustrated in Figure 9, showing examples of intraclass variations of fall (Figure 9a–e).
Viewpoint variations are illustrated in Figure 10a,b, showing a person sitting from different angles, highlighting the impact of changes in viewing perspective on human activity recognition (HAR) accuracy. An example of synchronised activity is shown in Figure 10c, where a person is relaxing on the couch and reading a book, demonstrating how concurrent actions can complicate activity classification. Subtle differences in motion, such as running versus jogging, are depicted in Figure 10c and 10d, respectively, showcasing the challenge of distinguishing activities that only vary slightly in movement patterns. Lastly, Figure 10f illustrates the issue of self-occlusion, where parts of the body, like limbs or joints, are obscured by other body segments, complicating the recognition process [46].
Depth sensors, such as Microsoft Kinect, address challenges related to varying lighting conditions but come with inherent limitations, including a restricted field of view and limited portability, which constrain their application in more dynamic or expansive environments [316]. Motion capture (Mocap) systems, while providing highly accurate and detailed data suitable for controlled environments like those used in sports injury rehabilitation or entertainment, are often prohibitively expensive and impractical for everyday activity monitoring. In contrast, wearable sensors offer a robust, view-invariant approach to HAR, capable of recognising movements regardless of environmental factors. However, these sensors typically recognise only a limited range of activities and may cause user discomfort or even skin irritation when worn for extended periods, as demonstrated in Figure 11a,b.
Data segmentation and labelling represent another major challenge for HAR research [127]. The process of segmenting continuous data streams and accurately labelling activities is time-consuming, labour-intensive, and prone to errors, particularly for complex or overlapping activities and across diverse sensor modalities [305]. Various existing windowing methods commonly used by HAR researchers are portrayed in Figure 12. Manual annotation of data is often required, introducing inconsistencies and limiting scalability. In response, potential solutions include crowdsourcing annotation tasks to distribute the workload across a larger pool of participants, as well as developing automated and semi-automated annotation tools. The authors in reference [127] have proposed an autoencoder-based approach that use frame reconstruction errors to identify activity transition points. Though the method is intuitive, it requires few hyperparameter initialisations. More research in this field is required to fully automate the activity segmentation process. Additionally, transfer learning techniques can be used to adapt pre-trained models from similar tasks or domains, reducing the reliance on large manually labelled datasets while maintaining model accuracy and relevance.
Another critical issue lies in model interpretability and transparency [317]. As deep learning-based HAR systems become more prevalent, they are often viewed as “black boxes”, offering little insight into the decision-making process. This lack of transparency hampers user trust and accountability, particularly in high-stakes applications such as healthcare or security. To address this, researchers are increasingly focused on developing techniques that improve model interpretability [318], such as attention mechanisms that highlight which features or data points the model focuses on during decision-making, model visualisation tools that provide intuitive insights into neural network behaviour, and explanation methods that allow users to understand the underlying reasons behind specific predictions.
The transferability and generalisability of HAR models remain as significant challenges. Models trained on specific datasets and conditions often struggle to generalise effectively to new environments. Likewise for sensor configurations, unseen activity patterns discovery remains challenging [319]. This limits the deployment of HAR systems in real-world applications. The variability in human actions, bias introduced by sensor placements, and other environmental factors can be vast. To overcome these challenges, techniques like transfer learning and domain adaptation are being explored by HAR researchers. These methods allow HAR models to adapt to new conditions using knowledge from related tasks by artificially generating diverse training samples to simulate real-world variability.
Ultimately, ensuring that HAR systems are user-friendly and meet the requirements of the target audience is achieved by incorporating human-centric design and evaluation [318]. Involving users in the design and evaluation process through participatory design and user-centred evaluation methodologies helps ensure that HAR systems are inclusive, accessible, and capable of addressing real-world challenges effectively [319]. By prioritising human-centric design principles, researchers can create systems that not only perform well in technical terms but are also widely adopted and accepted by users in various domains, from healthcare to smart homes.
These challenges, while significant, offer opportunities for further innovation in the field of HAR. Creating scalable and user-friendly HAR systems that are capable of meeting the demands of real-world applications, HAR researchers can advance the field.

8. Conclusions

Human activity recognition (HAR) has surfaced as a crucial field of computer vision and pattern recognition research. Research in this field offers transformative potential across diverse applications across various fields including healthcare, sports, security, and smart environments. HAR technology uses various data modalities like vision sensors, depth sensors, motion capture devices, and wearable sensors. Regardless of sensing medium, the goal of any HAR system is to accurately classify human activities. Each modality comes with its advantages and challenges. For example, RGB-based systems offer rich visual data but are sensitive to environmental factors like lighting. RGB-based sensors have a longer range but are sensitive to occlusions. Depth sensors on the other hand provide reliable 3D spatial information but face limitations in range. Depth sensors are efficient for segmenting the foreground and background but some of the devices can only be used in indoor environments. Motion capture devices have high precision, but they are costly and impractical for everyday use. Wearable sensors are effective for real-time activity recognition in different contexts, but they cause user discomfort over long periods. Moreover, wearable sensors are limited in classifying a limited number of activities.
HAR has been transformed by the transition from conventional rule-based systems to advanced machine learning methods. Traditional methods relied heavily on handcrafted features and pre-defined rules were used. This specific domain knowledge limited their application in recognising simple activities that are inadequate for handling complex action. In contrast, modern machine learning approaches like the deep learning models such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs), have significantly improved the performance of HAR systems. These models automatically extract features and learn from raw data that allow them to adapt to diverse activity contexts and achieve high accuracy.
The pre-processing of data remains a crucial aspect of HAR. Data cleaning, normalisation, and feature engineering are often applied to ensure optimal model performance. Techniques like noise reduction, data augmentation and dimensionality reduction are employed to enhance data quality. Data labelling and segmentation is essential for HAR processes, but these steps remain less explored as they are labour-intensive and error-prone processes. Innovations in semi-supervised and crowdsourced data labelling are helping to reduce these challenges to a certain extent paving the way for efficient data preparation.
Trends in HAR research indicate that multimodal sensor fusion is becoming increasingly important. Hybrid methods integrate diverse sensors to capture comprehensive information about human activities. Edge computing and IoT-based systems allow real-time data processing while reducing latency. Personalised and adaptive models are being developed to cater to individual users’ preferences. The ethical and social implications of HAR technology like privacy concerns, algorithmic fairness, and data security are gaining attention. Researchers emphasise transparency and accountability in the design and deployment of HAR systems.
The practical applications of HAR technology are vast and impactful. In healthcare, HAR systems are used for fall detection, remote patient monitoring, and personalised rehabilitation. Advancements in HAR technology in the medical field can significantly improve patient outcomes and safety. In sports, HAR technology can be used to provide real-time feedback on athletes’ biomechanics, optimising training and reducing injury risks. Security applications benefit from HAR’s ability to detect intrusions and monitor suspicious behaviours, enhancing public safety. Meanwhile, smart environments leverage HAR to automate daily functions, improve energy efficiency, and respond to occupants’ needs in a personalised manner.
Despite these advancements, HAR research continues to face challenges, including variations in human behaviour, model interpretability and achieving reliable recognition in real-world settings. Addressing these issues requires continued innovation and interdisciplinary collaboration to fully realise the potential of HAR technology. Future studies should prioritise strengthening privacy protection with advanced methods such as federated learning to enable decentralised model training without exposing raw data and differential privacy to ensure robust anonymisation in shared datasets. Additionally, multimodal fusion strategies should be further refined, with a focus on adaptive fusion techniques that dynamically adjust sensor weightings based on environmental conditions and task requirements. To enhance real-time processing, research should also explore the development of lightweight deep learning architecture optimised for HAR and leverage edge computing to reduce latency and enhance computational efficiency. Moreover, enhancing model interpretability remains a critical challenge for real-world adoption. Future efforts should also incorporate explainable AI strategies, such as feature attribution methods and attention-based visualisation, to improve transparency and trust in HAR systems. Additionally, benchmarking HAR models across diverse and unstructured environments will be essential to improve their generalisability and robustness.
As HAR research progresses, its integration into smart environments, healthcare, autonomous systems, and security applications will continue to expand, making daily interactions more adaptive, safe, and efficient. This study underscores the significant impact of HAR and highlights the ongoing need for research and development to harness its full capabilities for societal benefit.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jimaging11030091/s1, Table S1: Vision-based HAR datasets; Table S2: Depth and skeleton-based HAR datasets; Table S3: Wearable and mobile accelerometer-based HAR datasets; Table S4: Motion capture (MoCap)-based HAR datasets; Table S5: Wi-Fi and radar-based HAR datasets; Table S6: Multimodal HAR datasets; Table S7: Other sensor modalities used in HAR; Table S8: Machine learning-based HAR using different modalities; Table S9: Unsupervised learning-based HAR methods; Table S10: Semi-supervised learning-based HAR methods; Table S11: Deep learning-based HAR methods; Table S12: Transformer-based HAR methods; Table S13: Reinforcement learning-based HAR methods.

Author Contributions

Conceptualization, M.A.H. and P.E.A.; methodology, M.A.H. and P.E.A.; validation, P.E.A.; formal analysis, M.A.H.; investigation, M.A.H.; resources, P.E.A.; data curation, M.A.H.; writing—original draft preparation, M.A.H.; writing—review and editing, P.E.A.; visualization, M.A.H.; supervision, P.E.A.; funding acquisition, P.E.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Universiti Brunei Darussalam grant number UBD/RSCH/1.3/FICBF(b)/2024/023.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article or Supplementary Material S1.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhou, X.; Liang, W.; Kevin, I.; Wang, K.; Wang, H.; Yang, L.T.; Jin, Q. Deep-Learning-Enhanced Human Activity Recognition for Internet of Healthcare Things. IEEE Internet Things J. 2020, 7, 6429–6438. [Google Scholar]
  2. Mekruksavanich, S.; Jitpattanakul, A. Lstm Networks Using Smartphone Data for Sensor-Based Human Activity Recognition in Smart Homes. Sensors 2021, 21, 1636. [Google Scholar] [CrossRef]
  3. Hsu, Y.-L.; Yang, S.-C.; Chang, H.-C.; Lai, H.-C. Human Daily and Sport Activity Recognition Using a Wearable Inertial Sensor Network. IEEE Access 2018, 6, 31715–31728. [Google Scholar]
  4. Beddiar, D.R.; Nini, B.; Sabokrou, M.; Hadid, A. Vision-Based Human Activity Recognition: A Survey. Multimed. Tools Appl. 2020, 79, 30509–30555. [Google Scholar] [CrossRef]
  5. He, J.; Zhang, C.; He, X.; Dong, R. Visual Recognition of Traffic Police Gestures with Convolutional Pose Machine and Handcrafted Features. Neurocomputing 2020, 390, 248–259. [Google Scholar]
  6. Bian, S.; Liu, M.; Zhou, B.; Lukowicz, P.; Magno, M. Body-Area Capacitive or Electric Field Sensing for Human Activity Recognition and Human-Computer Interaction: A Comprehensive Survey. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2024, 8, 1–49. [Google Scholar] [CrossRef]
  7. Al, G.A.; Estrela, P.; Martinez-Hernandez, U. Towards an Intuitive Human-Robot Interaction Based on Hand Gesture Recognition and Proximity Sensors. In Proceedings of the 2020 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems (MFI), Karlsruhe, Germany, 14–16 September 2020; pp. 330–335. [Google Scholar]
  8. Malche, T.; Tharewal, S.; Tiwari, P.K.; Jabarulla, M.Y.; Alnuaim, A.A.; Hatamleh, W.A.; Ullah, M.A. Artificial Intelligence of Things-(AIoT-) Based Patient Activity Tracking System for Remote Patient Monitoring. J. Healthc. Eng. 2022, 2022, 8732213. [Google Scholar]
  9. Zolfaghari, S.; Keyvanpour, M.R. SARF: Smart Activity Recognition Framework in Ambient Assisted Living. In Proceedings of the 2016 Federated Conference on Computer Science and Information Systems (FedCSIS), Gdansk, Poland, 11–14 September 2016; pp. 1435–1443. [Google Scholar]
  10. Hoang, V.-H.; Lee, J.W.; Piran, M.J.; Park, C. Advances in Skeleton-Based Fall Detection in RGB Videos: From Handcrafted to Deep Learning Approaches. IEEE Access 2023, 11, 92322–92352. [Google Scholar]
  11. Tien, P.W.; Wei, S.; Calautit, J.K.; Darkwa, J.; Wood, C. Vision-Based Human Activity Recognition for Reducing Building Energy Demand. Build. Serv. Eng. Res. Technol. 2021, 42, 691–713. [Google Scholar]
  12. Zhang, W.; Liu, Z.; Zhou, L.; Leung, H.; Chan, A.B. Martial Arts, Dancing and Sports Dataset: A Challenging Stereo and Multi-View Dataset for 3d Human Pose Estimation. Image Vis. Comput. 2017, 61, 22–39. [Google Scholar]
  13. Yu, S.J.; Koh, P.; Kwon, H.; Kim, D.S.; Kim, H.K. Hurst Parameter Based Anomaly Detection for Intrusion Detection System. In Proceedings of the 2016 IEEE International Conference on Computer and Information Technology (CIT), Nadi, Fiji, 8–10 December 2016; pp. 234–240. [Google Scholar]
  14. Lipton, Z.C.; Berkowitz, J.; Elkan, C. A Critical Review of Recurrent Neural Networks for Sequence Learning. arXiv 2015, arXiv:1506.00019. [Google Scholar]
  15. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar]
  16. Tay, Y.; Dehghani, M.; Bahri, D.; Metzler, D. Efficient Transformers: A Survey. ACM Comput. Surv. 2022, 55, 1–28. [Google Scholar]
  17. Khan, S.; Naseer, M.; Hayat, M.; Zamir, S.W.; Khan, F.S.; Shah, M. Transformers in Vision: A Survey. ACM Comput. Surv. 2022, 54, 1–41. [Google Scholar]
  18. Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent Advances in Convolutional Neural Networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar]
  19. Yu, Y.; Si, X.; Hu, C.; Zhang, J. A Review of Recurrent Neural Networks: LSTM Cells and Network Architectures. Neural Comput. 2019, 31, 1235–1270. [Google Scholar]
  20. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar]
  21. Zhang, S.; Tong, H.; Xu, J.; Maciejewski, R. Graph Convolutional Networks: A Comprehensive Review. Comput. Soc. Netw. 2019, 6, 11. [Google Scholar] [CrossRef] [PubMed]
  22. Wu, F.; Souza, A.; Zhang, T.; Fifty, C.; Yu, T.; Weinberger, K. Simplifying Graph Convolutional Networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6861–6871. [Google Scholar]
  23. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  24. Dey, R.; Salem, F.M. Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks. In Proceedings of the 2017 IEEE 60th International Midwest Symposium on Circuits and Systems (MWSCAS), Boston, MA, USA, 6–9 August 2017; pp. 1597–1600. [Google Scholar]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  26. Tasnim, N.; Islam, M.; Baek, J.H. Deep Learning-Based Action Recognition Using 3d Skeleton Joints Information. Inventions 2020, 5, 49. [Google Scholar] [CrossRef]
  27. Georis, B.; Maziere, M.; Bremond, F.; Thonnat, M. A Video Interpretation Platform Applied to Bank Agency Monitoring. In Proceedings of the IEE Intelligent Distributed Surveilliance Systems, London, UK, 23 February 2004; pp. 46–50. [Google Scholar]
  28. Turaga, P.; Chellappa, R.; Subrahmanian, V.S.; Udrea, O. Machine Recognition of Human Activities: A Survey. IEEE Trans. Circuits Syst. Video Technol. 2008, 18, 1473–1488. [Google Scholar] [CrossRef]
  29. Pawlyta, M.; Skurowski, P. A Survey of Selected Machine Learning Methods for the Segmentation of Raw Motion Capture Data into Functional Body Mesh. In Proceedings of the Information Technologies in Medicine: 5th International Conference (ITIB 2016), Kamień Śląski, Poland, 20–22 June 2016; Volume 2, pp. 321–336. [Google Scholar]
  30. Akula, A.; Shah, A.K.; Ghosh, R. Deep Learning Approach for Human Action Recognition in Infrared Images. Cogn. Syst. Res. 2018, 50, 146–154. [Google Scholar]
  31. Chapron, K.; Lapointe, P.; Bouchard, K.; Gaboury, S. Highly Accurate Bathroom Activity Recognition Using Infrared Proximity Sensors. IEEE J. Biomed. Health Inform. 2020, 24, 2368–2377. [Google Scholar] [CrossRef]
  32. Ghosh, R.; Gupta, A.; Nakagawa, A.; Soares, A.; Thakor, N. Spatiotemporal Filtering for Event-Based Action Recognition. arXiv 2019, arXiv:1903.07067. [Google Scholar]
  33. Wang, Y.; Xiao, Y.; Xiong, F.; Jiang, W.; Cao, Z.; Zhou, J.T.; Yuan, J. 3DV: 3D Dynamic Voxel for Action Recognition in Depth Video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 511–520. [Google Scholar]
  34. Singh, A.D.; Sandha, S.S.; Garcia, L.; Srivastava, M. Radhar: Human Activity Recognition from Point Clouds Generated through a Millimeter-Wave Radar. In Proceedings of the 3rd ACM Workshop on Millimeter-Wave Networks and Sensing Systems, Los Cabos, Mexico, 21–25 October 2019; pp. 51–56. [Google Scholar]
  35. Duan, H.; Zhao, Y.; Chen, K.; Lin, D.; Dai, B. Revisiting Skeleton-Based Action Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 2969–2978. [Google Scholar]
  36. Donahue, J.; Anne Hendricks, L.; Guadarrama, S.; Rohrbach, M.; Venugopalan, S.; Saenko, K.; Darrell, T. Long-Term Recurrent Convolutional Networks for Visual Recognition and Description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 2625–2634. [Google Scholar]
  37. Liang, D.; Thomaz, E. Audio-Based Activities of Daily Living (Adl) Recognition with Large-Scale Acoustic Embeddings from Online Videos. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2019, 3, 1–18. [Google Scholar]
  38. Wang, F.; Song, Y.; Zhang, J.; Han, J.; Huang, D. Temporal Unet: Sample Level Human Action Recognition Using Wifi. arXiv 2019, arXiv:1904.11953. [Google Scholar]
  39. Roche, J.; De-Silva, V.; Hook, J.; Moencks, M.; Kondoz, A. A Multimodal Data Processing System for LiDAR-Based Human Activity Recognition. IEEE Trans. Cybern. 2022, 52, 10027–10040. [Google Scholar] [CrossRef]
  40. Attal, F.; Mohammed, S.; Dedabrishvili, M.; Chamroukhi, F.; Oukhellou, L.; Amirat, Y. Physical Human Activity Recognition Using Wearable Sensors. Sensors 2015, 15, 31314–31338. [Google Scholar] [CrossRef]
  41. Gupta, A.; Gupta, K.; Gupta, K.; Gupta, K. A Survey on Human Activity Recognition and Classification. In Proceedings of the 2020 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 28–30 July 2020; pp. 915–919. [Google Scholar] [CrossRef]
  42. Zeng, M.; Nguyen, L.T.; Yu, B.; Mengshoel, O.J.; Zhu, J.; Wu, P.; Zhang, J. Convolutional Neural Networks for Human Activity Recognition Using Mobile Sensors. In Proceedings of the 6th International Conference on Mobile Computing, Applications and Services, Austin, TX, USA, 6–7 November 2014; pp. 197–205. [Google Scholar]
  43. Imran, J.; Raman, B. Evaluating Fusion of RGB-D and Inertial Sensors for Multimodal Human Action Recognition. J. Ambient Intell. Humaniz. Comput. 2020, 11, 189–208. [Google Scholar] [CrossRef]
  44. Aggarwal, J.K.; Cai, Q. Human Motion Analysis: A Review. Comput. Vis. Image Underst. 1999, 73, 428–440. [Google Scholar]
  45. Keyvanpour, M.R.; Vahidian, S.; Ramezani, M. HMR-Vid: A Comparative Analytical Survey on Human Motion Recognition in Video Data. Multimed. Tools Appl. 2020, 79, 31819–31863. [Google Scholar]
  46. Gupta, N.; Gupta, S.K.; Pathak, R.K.; Jain, V.; Rashidi, P.; Suri, J.S. Human Activity Recognition in Artificial Intelligence Framework: A Narrative Review. Artif. Intell. Rev. 2022, 55, 4755–4808. [Google Scholar]
  47. Ma, N.; Wu, Z.; Cheung, Y.; Guo, Y.; Gao, Y.; Li, J.; Jiang, B. A Survey of Human Action Recognition and Posture Prediction. Tsinghua Sci. Technol. 2022, 27, 973–1001. [Google Scholar] [CrossRef]
  48. Aggarwal, J.K.; Xia, L. Human Activity Recognition from 3D Data: A Review. Pattern Recognit. Lett. 2014, 48, 70–80. [Google Scholar] [CrossRef]
  49. Chen, L.; Wei, H.; Ferryman, J. A Survey of Human Motion Analysis Using Depth Imagery. Pattern Recognit. Lett. 2013, 34, 1995–2006. [Google Scholar] [CrossRef]
  50. Carvalho, L.I.; Sofia, R.C. A Review on Scaling Mobile Sensing Platforms for Human Activity Recognition: Challenges and Recommendations for Future Research. IoT 2020, 1, 451–473. [Google Scholar] [CrossRef]
  51. Homayounfar, S.Z.; Andrew, T.L. Wearable Sensors for Monitoring Human Motion: A Review on Mechanisms, Materials, and Challenges. SLAS Technol. Transl. Life Sci. Innov. 2020, 25, 9–24. [Google Scholar]
  52. Desmarais, Y.; Mottet, D.; Slangen, P.; Montesinos, P. A Review of 3D Human Pose Estimation Algorithms for Markerless Motion Capture. Comput. Vis. Image Underst. 2021, 212, 103275. [Google Scholar]
  53. Han, F.; Reily, B.; Hoff, W.; Zhang, H. Space-Time Representation of People Based on 3D Skeletal Data: A Review. Comput. Vis. Image Underst. 2017, 158, 85–105. [Google Scholar] [CrossRef]
  54. Singh, R.; Sonawane, A.; Srivastava, R. Recent Evolution of Modern Datasets for Human Activity Recognition: A Deep Survey. Multimed. Syst. 2020, 26, 83–106. [Google Scholar] [CrossRef]
  55. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human Action Recognition from Various Data Modalities: A Review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar]
  56. Pareek, P.; Thakkar, A. A Survey on Video-Based Human Action Recognition: Recent Updates, Datasets, Challenges, and Applications. Artif. Intell. Rev. 2021, 54, 2259–2322. [Google Scholar]
  57. Kong, Y.; Fu, Y. Human Action Recognition and Prediction: A Survey. Int. J. Comput. Vis. 2022, 130, 1366–1401. [Google Scholar]
  58. Azis, F.A.; Rijal, M.; Suhaimi, H.; Abas, P.E. Patent Landscape of Composting Technology: A Review. Inventions 2022, 7, 38. [Google Scholar] [CrossRef]
  59. Ratnayake, A.M.B.; Yasin, H.M.; Naim, A.G.; Abas, P.E. Buzzing through Data: Advancing Bee Species Identification with Machine Learning. Appl. Syst. Innov. 2024, 7, 62. [Google Scholar] [CrossRef]
  60. Hossen, M.A.; Abas, P.E. A Comparative Study of Supervised and Unsupervised Approaches in Human Activity Analysis Based on Skeleton Data. Int. J. Comput. Digit. Syst. 2023, 14, 10407–10421. [Google Scholar]
  61. Vallabh, P.; Malekian, R. Fall Detection Monitoring Systems: A Comprehensive Review. J. Ambient Intell. Humaniz. Comput. 2018, 9, 1809–1833. [Google Scholar] [CrossRef]
  62. Nweke, H.F.; Teh, Y.W.; Mujtaba, G.; Al-garadi, M.A. Data Fusion and Multiple Classifier Systems for Human Activity Detection and Health Monitoring: Review and Open Research Directions. Inf. Fusion 2019, 46, 147–170. [Google Scholar] [CrossRef]
  63. Frank, A.E.; Kubota, A.; Riek, L.D. Wearable Activity Recognition for Robust Human-Robot Teaming in Safety-Critical Environments via Hybrid Neural Networks. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 449–454. [Google Scholar]
  64. Xia, L.; Chen, C.C.; Aggarwal, J.K. View Invariant Human Action Recognition Using Histograms of 3D Joints. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 20–27. [Google Scholar]
  65. Vrigkas, M.; Nikou, C.; Kakadiaris, I.A. A Review of Human Activity Recognition Methods. Front. Robot. AI 2015, 2, 28. [Google Scholar] [CrossRef]
  66. Materzynska, J.; Berger, G.; Bax, I.; Memisevic, R. The Jester Dataset: A Large-Scale Video Dataset of Human Gestures. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  67. Fronteddu, G.; Porcu, S.; Floris, A.; Atzori, L. A Dynamic Hand Gesture Recognition Dataset for Human-Computer Interfaces. Comput. Netw. 2022, 205, 108781. [Google Scholar] [CrossRef]
  68. Minh Dang, L.; Min, K.; Wang, H.; Jalil Piran, M.; Hee Lee, C.; Moon, H. Sensor-Based and Vision-Based Human Activity Recognition: A Comprehensive Survey. Pattern Recognit. 2020, 108, 107561. [Google Scholar] [CrossRef]
  69. Zhang, S.; Wei, Z.; Nie, J.; Huang, L.; Wang, S.; Li, Z. A Review on Human Activity Recognition Using Vision-Based Method. J. Healthc. Eng. 2017, 2017, 3090343. [Google Scholar] [CrossRef] [PubMed]
  70. Bhatnagar, B.L.; Xie, X.; Petrov, I.A.; Sminchisescu, C.; Theobalt, C.; Pons-Moll, G. Behave: Dataset and Method for Tracking Human Object Interactions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 15935–15946. [Google Scholar]
  71. Lu, L.; Lu, Y.; Yu, R.; Di, H.; Zhang, L.; Wang, S. GAIM: Graph Attention Interaction Model for Collective Activity Recognition. IEEE Trans. Multimed. 2019, 22, 524–539. [Google Scholar] [CrossRef]
  72. Ibrahim, M.S.; Muralidharan, S.; Deng, Z.; Vahdat, A.; Mori, G. A Hierarchical Deep Temporal Model for Group Activity Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1971–1980. [Google Scholar]
  73. Qi, M.; Wang, Y.; Qin, J.; Li, A.; Luo, J.; Van Gool, L. StagNet: An Attentive Semantic RNN for Group Activity and Individual Action Recognition. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 549–565. [Google Scholar] [CrossRef]
  74. Tang, Y.; Lu, J.; Wang, Z.; Yang, M.; Zhou, J. Learning Semantics-Preserving Attention and Contextual Interaction for Group Activity Recognition. IEEE Trans. Image Process. 2019, 28, 4997–5012. [Google Scholar] [CrossRef] [PubMed]
  75. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 4338–4364. [Google Scholar] [CrossRef]
  76. Hossen, M.A.; Hong, O.W.; Caesarendra, W. Investigation of the Unsupervised Machine Learning Techniques for Human Activity Discovery. In Proceedings of the 2nd International Conference on Electronics, Biomedical Engineering, and Health Informatics, Surabaya, Indonesia, 3–4 November 2021; Triwiyanto, T., Rizal, A., Caesarendra, W., Eds.; Springer Nature: Singapore, 2022; pp. 499–514. [Google Scholar]
  77. Huynh, T.; Fritz, M.; Schiele, B. Discovery of Activity Patterns Using Topic Models. In Proceedings of the 10th International Conference on Ubiquitous Computing (UbiComp ’08), Seoul, Republic of Korea, 21–24 September 2008; ACM Press: New York, NY, USA, 2008; pp. 10–19. [Google Scholar]
  78. Kim, E.; Helal, S.; Cook, D. Human Activity Recognition and Pattern Discovery. IEEE Pervasive Comput. 2010, 9, 48–53. [Google Scholar] [CrossRef]
  79. Gjoreski, H.; Roggen, D. Unsupervised Online Activity Discovery Using Temporal Behaviour Assumption. In Proceedings of the 2017 ACM International Symposium on Wearable Computers (ISWC 2017), Maui, HI, USA, 11–15 September 2017; pp. 42–49. [Google Scholar] [CrossRef]
  80. Rashidi, P. Stream Sequence Mining for Human Activity Discovery; Elsevier Inc.: Amsterdam, The Netherlands, 2014; ISBN 9780123985323. [Google Scholar]
  81. Ye, J.; Fang, L.; Dobson, S. Discovery and Recognition of Unknown Activities. In Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct, Heidelberg, Germany, 12–16 September 2016; pp. 783–792. [Google Scholar] [CrossRef]
  82. Ong, W.H.; Koseki, T.; Palafox, L. Unsupervised Human Activity Detection with Skeleton Data from RGB-D Sensor. In Proceedings of the 2013 Fifth International Conference on Computational Intelligence, Communication Systems and Networks (CICSyN 2013), Madrid, Spain, 5–7 June 2013; pp. 30–35. [Google Scholar] [CrossRef]
  83. Steil, J.; Bulling, A. Discovery of Everyday Human Activities from Long-Term Visual Behaviour Using Topic Models. In Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing (UbiComp 2015), Osaka, Japan, 7–11 September 2015; pp. 75–85. [Google Scholar] [CrossRef]
  84. Wang, Y.; Jiang, H.; Drew, M.S.; Li, Z.N.; Mori, G. Unsupervised Discovery of Action Classes. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 1654–1661. [Google Scholar] [CrossRef]
  85. Fang, L.; Ye, J.; Dobson, S. Discovery and Recognition of Emerging Human Activities Using a Hierarchical Mixture of Directional Statistical Models. IEEE Trans. Knowl. Data Eng. 2019, 14, 1304–1316. [Google Scholar] [CrossRef]
  86. Rieping, K.; Englebienne, G.; Kröse, B. Behavior Analysis of Elderly Using Topic Models. Pervasive Mob. Comput. 2014, 15, 181–199. [Google Scholar] [CrossRef]
  87. Kwon, Y.; Kang, K.; Bae, C. Unsupervised Learning for Human Activity Recognition Using Smartphone Sensors. Expert Syst. Appl. 2014, 41, 6067–6074. [Google Scholar] [CrossRef]
  88. Veeraraghavan, A.; Roy-Chowdhury, A.K.; Chellappa, R. Matching Shape Sequences in Video with Applications in Human Movement Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1896–1909. [Google Scholar] [PubMed]
  89. Han, J.; Bhanu, B. Human Activity Recognition in Thermal Infrared Imagery. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05)-Workshops, San Diego, CA, USA, 20–25 June 2005; p. 17. [Google Scholar]
  90. Hossen, M.A.; Naim, A.G.; Abas, P.E. Evaluation of 2D and 3D Posture for Human Activity Recognition. AIP Conf. Proc. 2023, 2643, 40013. [Google Scholar] [CrossRef]
  91. Barnachon, M.; Bouakaz, S.; Boufama, B.; Guillou, E. Ongoing Human Action Recognition with Motion Capture. Pattern Recognit. 2014, 47, 238–247. [Google Scholar] [CrossRef]
  92. Xu, T.; An, D.; Jia, Y.; Yue, Y. A Review: Point Cloud-Based 3d Human Joints Estimation. Sensors 2021, 21, 1684. [Google Scholar] [CrossRef] [PubMed]
  93. Calabrese, E.; Taverni, G.; Awai Easthope, C.; Skriabine, S.; Corradi, F.; Longinotti, L.; Eng, K.; Delbruck, T. DHP19: Dynamic Vision Sensor 3D Human Pose Dataset. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  94. Zia Uddin, M.; Khaksar, W.; Torresen, J. A Thermal Camera-Based Activity Recognition Using Discriminant Skeleton Features and RNN. In Proceedings of the 2019 IEEE 17th International Conference on Industrial Informatics (INDIN), Helsinki, Finland, 22–25 July 2019; Volume 1, pp. 777–782. [Google Scholar]
  95. Jung, M.; Chi, S. Human Activity Classification Based on Sound Recognition and Residual Convolutional Neural Network. Autom. Constr. 2020, 114, 103177. [Google Scholar]
  96. Fioranelli, D.F.; Shah, D.S.A.; Li1, H.; Shrestha, A.; Yang, D.S.; Kernec, D.J. Le Radar Sensing for Healthcare: Associate Editor Francesco Fioranelli on the Applications of Radar in Monitoring Vital Signs and Recognising Human Activity Patterns. Electron. Lett. 2019, 55, 1022–1024. [Google Scholar]
  97. Yadav, S.K.; Sai, S.; Gundewar, A.; Rathore, H.; Tiwari, K.; Pandey, H.M.; Mathur, M. CSITime: Privacy-Preserving Human Activity Recognition Using WiFi Channel State Information. Neural Netw. 2022, 146, 11–21. [Google Scholar] [CrossRef]
  98. Straczkiewicz, M.; Onnela, J.-P. A Systematic Review of Human Activity Recognition Using Smartphones. arXiv 2019, arXiv:1910.03970. [Google Scholar]
  99. Soomro, K.; Zamir, A.R.; Shah, M. UCF101: A Dataset of 101 Human Actions Classes from Videos in the Wild. arXiv 2012, arXiv:1212.0402. [Google Scholar]
  100. Karpathy, A.; Toderici, G.; Shetty, S.; Leung, T.; Sukthankar, R.; Fei-Fei, L. Large-Scale Video Classification with Convolutional Neural Networks. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1725–1732. [Google Scholar]
  101. Caba Heilbron, F.; Escorcia, V.; Ghanem, B.; Carlos Niebles, J. Activitynet: A Large-Scale Video Benchmark for Human Activity Understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 961–970. [Google Scholar]
  102. Gorban, A.; Idrees, H.; Jiang, Y.-G.; Zamir, A.R.; Laptev, I.; Shah, M.; Sukthankar, R. THUMOS Challenge: Action Recognition with a Large Number of Classes; 2015. Available online: http://www.thumos.info (accessed on 14 March 2025).
  103. Sigurdsson, G.A.; Varol, G.; Wang, X.; Farhadi, A.; Laptev, I.; Gupta, A. Hollywood in Homes: Crowdsourcing Data Collection for Activity Understanding. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part I 14. Springer: Cham, Switzerland; pp. 510–526. [Google Scholar]
  104. Gao, C.; Du, Y.; Liu, J.; Lv, J.; Yang, L.; Meng, D.; Hauptmann, A.G. Infar Dataset: Infrared Action Recognition at Different Times. Neurocomputing 2016, 212, 36–47. [Google Scholar] [CrossRef]
  105. Shahroudy, A.; Liu, J.; Ng, T.T.; Wang, G. NTU RGB+D: A Large Scale Dataset for 3D Human Activity Analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1010–1019. [Google Scholar] [CrossRef]
  106. Amir, A.; Taba, B.; Berg, D.; Melano, T.; McKinstry, J.; Di Nolfo, C.; Nayak, T.; Andreopoulos, A.; Garreau, G.; Mendoza, M.; et al. A Low Power, Fully Event-Based Gesture Recognition System. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7243–7252. [Google Scholar]
  107. Jiang, Y.-G.; Wu, Z.; Wang, J.; Xue, X.; Chang, S.-F. Exploiting Feature and Class Relationships in Video Categorization with Regularized Deep Neural Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 352–364. [Google Scholar]
  108. Kay, W.; Carreira, J.; Simonyan, K.; Zhang, B.; Hillier, C.; Vijayanarasimhan, S.; Viola, F.; Green, T.; Back, T.; Natsev, P.; et al. The Kinetics Human Action Video Dataset. arXiv 2017, arXiv:1705.06950. [Google Scholar]
  109. Liu, C.; Hu, Y.; Li, Y.; Song, S.; Liu, J. Pku-Mmd: A Large Scale Benchmark for Continuous Multi-Modal Human Action Understanding. arXiv 2017, arXiv:1703.07475. [Google Scholar]
  110. Goyal, R.; Kahou, S.E.; Michalski, V.; Materzynska, J.; Westphal, S.; Kim, H.; Haenel, V.; Fruend, I.; Yianilos, P.; Mueller-Freitag, M.; et al. The “Something Something” Video Database for Learning and Evaluating Visual Common Sense. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5843–5851. [Google Scholar]
  111. Carreira, J.; Noland, E.; Banki-Horvath, A.; Hillier, C.; Zisserman, A. A Short Note about Kinetics-600. arXiv 2018, arXiv:1808.01340. [Google Scholar]
  112. Ji, Y.; Xu, F.; Yang, Y.; Shen, F.; Shen, H.T.; Zheng, W.-S. A Large-Scale RGB-D Database for Arbitrary-View Human Action Recognition. In Proceedings of the 26th ACM International Conference on Multimedia, Seoul, Republic of Korea, 22–26 October 2018; pp. 1510–1518. [Google Scholar]
  113. Martin, M.; Roitberg, A.; Haurilet, M.; Horne, M.; Reiß, S.; Voit, M.; Stiefelhagen, R. Drive & Act: A Multi-Modal Dataset for Fine-Grained Driver Behavior Recognition in Autonomous Vehicles. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2801–2810. [Google Scholar]
  114. Zhang, Y.; Cao, C.; Cheng, J.; Lu, H. Egogesture: A New Dataset and Benchmark for Egocentric Hand Gesture Recognition. IEEE Trans. Multimed. 2018, 20, 1038–1050. [Google Scholar]
  115. Monfort, M.; Andonian, A.; Zhou, B.; Ramakrishnan, K.; Bargal, S.A.; Yan, T.; Brown, L.; Fan, Q.; Gutfreund, D.; Vondrick, C.; et al. Moments in Time Dataset: One Million Videos for Event Understanding. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 502–508. [Google Scholar] [CrossRef]
  116. Liu, J.; Shahroudy, A.; Perez, M.; Wang, G.; Duan, L.-Y.; Kot, A.C. NTU RGB+D 120: A Large-Scale Benchmark for 3D Human Activity Understanding. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2684–2701. [Google Scholar] [CrossRef]
  117. Miech, A.; Alayrac, J.-B.; Laptev, I.; Sivic, J.; Zisserman, A. Rareact: A Video Dataset of Unusual Interactions. arXiv 2020, arXiv:2008.01018. [Google Scholar]
  118. Ghorbani, S.; Mahdaviani, K.; Thaler, A.; Kording, K.; Cook, D.J.; Blohm, G.; Troje, N.F. MoVi: A Large Multi-Purpose Human Motion and Video Dataset. PLoS ONE 2021, 16, e0253157. [Google Scholar]
  119. Li, T.; Liu, J.; Zhang, W.; Ni, Y.; Wang, W.; Li, Z. Uav-Human: A Large Benchmark for Human Behavior Understanding with Unmanned Aerial Vehicles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 16266–16275. [Google Scholar]
  120. Rai, N.; Chen, H.; Ji, J.; Desai, R.; Kozuka, K.; Ishizaka, S.; Adeli, E.; Niebles, J.C. Home Action Genome: Cooperative Compositional Action Understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 11184–11193. [Google Scholar]
  121. Grauman, K.; Westbury, A.; Byrne, E.; Chavis, Z.; Furnari, A.; Girdhar, R.; Hamburger, J.; Jiang, H.; Liu, M.; Liu, X.; et al. Ego4d: Around the World in 3000 Hours of Egocentric Video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 18995–19012. [Google Scholar]
  122. Damen, D.; Doughty, H.; Farinella, G.M.; Furnari, A.; Kazakos, E.; Ma, J.; Moltisanti, D.; Munro, J.; Perrett, T.; Price, W.; et al. Rescaling Egocentric Vision: Collection, Pipeline and Challenges for Epic-Kitchens-100. Int. J. Comput. Vis. 2022, 130, 33–55. [Google Scholar]
  123. Damen, D.; Doughty, H.; Farinella, G.M.; Fidler, S.; Furnari, A.; Kazakos, E.; Moltisanti, D.; Munro, J.; Perrett, T.; Price, W.; et al. Scaling Egocentric Vision: The Epic-Kitchens Dataset. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 720–736. [Google Scholar]
  124. Seneviratne, S.; Hu, Y.; Nguyen, T.; Lan, G.; Khalifa, S.; Thilakarathna, K.; Hassan, M.; Seneviratne, A. A Survey of Wearable Devices and Challenges. IEEE Commun. Surv. Tutor. 2017, 19, 2573–2620. [Google Scholar] [CrossRef]
  125. Gawande, U.; Hajari, K.; Golhar, Y. Scale-Invariant Mask R-CNN for Pedestrian Detection. Electron. Lett. Comput. Vis. Image Anal. 2020, 19, 98–117. [Google Scholar] [CrossRef]
  126. Ziaeefard, M.; Bergevin, R. Semantic Human Activity Recognition: A Literature Review. Pattern Recognit. 2015, 48, 2329–2345. [Google Scholar] [CrossRef]
  127. Hossen, M.A.; Naim, A.G.; Abas, P.E. Deep Learning for Skeleton-Based Human Activity Segmentation: An Autoencoder Approach. Technologies 2024, 12, 96. [Google Scholar] [CrossRef]
  128. Kim, S.-Y.; Kim, M.; Ho, Y.-S. Depth Image Filter for Mixed and Noisy Pixel Removal in RGB-D Camera Systems. IEEE Trans. Consum. Electron. 2013, 59, 681–689. [Google Scholar]
  129. Ahad, M.A.R.; Ahmed, M.; Das Antar, A.; Makihara, Y.; Yagi, Y. Action Recognition Using Kinematics Posture Feature on 3D Skeleton Joint Locations. Pattern Recognit. Lett. 2021, 145, 216–224. [Google Scholar] [CrossRef]
  130. Uddin, M.Z.; Thang, N.D.; Kim, J.T.; Kim, T.S. Human Activity Recognition Using Body Joint-Angle Features and Hidden Markov Model. ETRI J. 2011, 33, 569–579. [Google Scholar] [CrossRef]
  131. Franco, A.; Magnani, A.; Maio, D. Joint Orientations from Skeleton Data for Human Activity Recognition. In Proceedings of the International Conference on Image Analysis and Processing, Modena, Italy, 11–15 September 2017; Volume 2, pp. 152–162, ISBN 9783319685601. [Google Scholar]
  132. Wang, Q.; Guo, Y.; Yu, L.; Chen, X.; Li, P. Deep Q-Network-Based Feature Selection for Multisourced Data Cleaning. IEEE Internet Things J. 2020, 8, 16153–16164. [Google Scholar]
  133. Richert, W.; Coelho, L.P. Building Machine Learning Systems with Python, 2nd ed.; Packt Publishing: Birmingham, UK, 2015; ISBN 978-1-78439-277-2. [Google Scholar]
  134. Müller, P.N.; Rauterberg, F.; Achenbach, P.; Tregel, T.; Göbel, S. Physical Exercise Quality Assessment Using Wearable Sensors. In Multimedia Tools and Applications; Springer: Cham, Switzerland, 2021; Volume 80, pp. 229–243. [Google Scholar]
  135. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar]
  136. Vishwakarma, D.K. A Two-Fold Transformation Model for Human Action Recognition Using Decisive Pose. Cogn. Syst. Res. 2020, 61, 1–13. [Google Scholar] [CrossRef]
  137. Xu, D.; Tian, Y. A Comprehensive Survey of Clustering Algorithms. Ann. Data Sci. 2015, 2, 165–193. [Google Scholar] [CrossRef]
  138. Taherkhani, A.; Cosma, G.; McGinnity, T.M. AdaBoost-CNN: An Adaptive Boosting Algorithm for Convolutional Neural Networks to Classify Multi-Class Imbalanced Datasets Using Transfer Learning. Neurocomputing 2020, 404, 351–366. [Google Scholar]
  139. Ng, W.W.Y.; Zeng, G.; Zhang, J.; Yeung, D.S.; Pedrycz, W. Dual Autoencoders Features for Imbalance Classification Problem. Pattern Recognit. 2016, 60, 875–889. [Google Scholar]
  140. Dawar, N.; Ostadabbas, S.; Kehtarnavaz, N. Data Augmentation in Deep Learning-Based Fusion of Depth and Inertial Sensing for Action Recognition. IEEE Sens. Lett. 2018, 3, 7101004. [Google Scholar]
  141. Nan, M.; Florea, A.M. Fast Temporal Graph Convolutional Model for Skeleton-Based Action Recognition. Sensors 2022, 22, 7117. [Google Scholar] [CrossRef]
  142. Lopez-del Rio, A.; Martin, M.; Perera-Lluna, A.; Saidi, R. Effect of Sequence Padding on the Performance of Deep Learning Models in Archaeal Protein Functional Prediction. Sci. Rep. 2020, 10, 14634. [Google Scholar] [CrossRef]
  143. Lee, N.; Kim, J.-M. Conversion of Categorical Variables into Numerical Variables via Bayesian Network Classifiers for Binary Classifications. Comput. Stat. Data Anal. 2010, 54, 1247–1265. [Google Scholar]
  144. Angelov, P.; Filev, D.P.; Kasabov, N. Evolving Intelligent Systems: Methodology and Applications; John Wiley & Sons: Hoboken, NJ, USA, 2010; Volume 4, ISBN 9780470287194. [Google Scholar]
  145. Kumar, K.; Mishra, R.K. A Heuristic SVM Based Pedestrian Detection Approach Employing Shape and Texture Descriptors. Multimed. Tools Appl. 2020, 79, 21389–21408. [Google Scholar] [CrossRef]
  146. Fan, L.; Wang, Z.; Wang, H. Human Activity Recognition Model Based on Decision Tree. In Proceedings of the 2013 International Conference on Advanced Cloud and Big Data, Sanya, China, 16–18 December 2013; pp. 64–68. [Google Scholar]
  147. Nurwulan, N.R.; Selamaj, G. Human Daily Activities Recognition Using Decision Tree. J. Phys. Conf. Ser. 2021, 1833, 12039. [Google Scholar]
  148. Müller, P.N.; Rauterberg, F.; Achenbach, P.; Tregel, T.; Göbel, S. Physical Exercise Quality Assessment Using Wearable Sensors. In Proceedings of the Serious Games; Fletcher, B., Ma, M., Göbel, S., Baalsrud Hauge, J., Marsh, T., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 229–243. [Google Scholar]
  149. Lara, O.D.; Labrador, M.A. A Survey on Human Activity Recognition Using Wearable Sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar] [CrossRef]
  150. Pärkkä, J.; Cluitmans, L.; Ermes, M. Personalization Algorithm for Real-Time Activity Recognition Using PDA, Wireless Motion Bands, and Binary Decision Tree. IEEE Trans. Inf. Technol. Biomed. 2010, 14, 1211–1215. [Google Scholar] [CrossRef]
  151. Balli, S.; Sa\ugba\cs, E.A.; Peker, M. Human Activity Recognition from Smart Watch Sensor Data Using a Hybrid of Principal Component Analysis and Random Forest Algorithm. Meas. Control 2019, 52, 37–45. [Google Scholar]
  152. Dewi, C.; Chen, R.-C. Human Activity Recognition Based on Evolution of Features Selection and Random Forest. In Proceedings of the 2019 IEEE International Conference on Systems, Man and Cybernetics (SMC), Bari, Italy, 6–9 October 2019; pp. 2496–2501. [Google Scholar]
  153. Zhu, Y.; Chen, W.; Guo, G. Evaluating Spatiotemporal Interest Point Features for Depth-Based Action Recognition. Image Vis. Comput. 2014, 32, 453–464. [Google Scholar]
  154. Nunes, U.M.; Faria, D.R.; Peixoto, P. A Human Activity Recognition Framework Using Max-Min Features and Key Poses with Differential Evolution Random Forests Classifier. Pattern Recognit. Lett. 2017, 99, 21–31. [Google Scholar] [CrossRef]
  155. Nurwulan, N.R.; Selamaj, G. Random Forest for Human Daily Activity Recognition. J. Phys. Conf. Ser. 2020, 1655, 12087. [Google Scholar]
  156. Gan, L.; Chen, F. Human Action Recognition Using APJ3D and Random Forests. J. Softw. 2013, 8, 2238–2245. [Google Scholar] [CrossRef]
  157. Cippitelli, E.; Gasparrini, S.; Gambi, E.; Spinsante, S. A Human Activity Recognition System Using Skeleton Data from RGBD Sensors. Comput. Intell. Neurosci. 2016, 2016, 4351435. [Google Scholar] [CrossRef]
  158. Tran, D.N.; Phan, D.D. Human Activities Recognition in Android Smartphone Using Support Vector Machine. In Proceedings of the 2016 7th International Conference on Intelligent Systems, Modelling and Simulation (ISMS), Bangkok, Thailand, 26–28 January 2016; pp. 64–68. [Google Scholar]
  159. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. Human Activity Recognition on Smartphones Using a Multiclass Hardware-Friendly Support Vector Machine. In Ambient Assisted Living and Home Care, Proceedings of the 4th International Workshop (IWAAL 2012), Vitoria-Gasteiz, Spain, 3–5 December 2012; Proceedings 4; Springer: Berlin/Heidelberg, Germany, 2012; pp. 216–223. [Google Scholar]
  160. Kim, Y.; Ling, H. Human Activity Classification Based on Micro-Doppler Signatures Using a Support Vector Machine. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1328–1337. [Google Scholar] [CrossRef]
  161. Shuvo, M.M.H.; Ahmed, N.; Nouduri, K.; Palaniappan, K. A Hybrid Approach for Human Activity Recognition with Support Vector Machine and 1D Convolutional Neural Network. In Proceedings of the 2020 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Virtual, 13–15 October 2020; pp. 1–5. [Google Scholar]
  162. Nurhanim, K.; Elamvazuthi, I.; Izhar, L.I.; Ganesan, T. Classification of Human Activity Based on Smartphone Inertial Sensor Using Support Vector Machine. In Proceedings of the 2017 IEEE 3rd International Symposium in Robotics and Manufacturing Automation (ROMA), Kuala Lumpur, Malaysia, 19–21 September 2017; pp. 1–5. [Google Scholar]
  163. Mohsen, S.; Elkaseer, A.; Scholz, S.G. Human Activity Recognition Using K-Nearest Neighbor Machine Learning Algorithm. In Proceedings of the International Conference on Sustainable Design and Manufacturing, Virtual, 15–17 September 2021; pp. 304–313. [Google Scholar]
  164. Kaghyan, S.; Sarukhanyan, H. Activity Recognition Using K-Nearest Neighbor Algorithm on Smartphone with Tri-Axial Accelerometer. Int. J. Inform. Model. Anal. (IJIMA) ITHEA Int. Sci. Soc. Bulg. 2012, 1, 146–156. [Google Scholar]
  165. Al-Akam, R.; Paulus, D. RGBD Human Action Recognition Using Multi-Features Combination and k-Nearest Neighbors Classification. Int. J. Adv. Comput. Sci. Appl. 2017, 8, 383–389. [Google Scholar] [CrossRef]
  166. Mandong, A.; Munir, U. Smartphone Based Activity Recognition Using K-Nearest Neighbor Algorithm. In Proceedings of the International Conference on Engineering Technologies, Konya, Turkey, 26–28 October 2018; pp. 26–28. [Google Scholar]
  167. Zhang, L.; Wu, X.; Luo, D. Human Activity Recognition with HMM-DNN Model. In Proceedings of the 2015 IEEE 14th International Conference on Cognitive Informatics & Cognitive Computing (ICCI* CC), Beijing, China, 15–17 June 2015; pp. 192–197. [Google Scholar]
  168. Ronao, C.A.; Cho, S.-B. Human Activity Recognition Using Smartphone Sensors with Two-Stage Continuous Hidden Markov Models. In Proceedings of the 2014 10th International Conference on Natural Computation (ICNC), Xiamen, China, 19–21 August 2014; pp. 681–686. [Google Scholar]
  169. San-Segundo, R.; Montero, J.M.; Moreno-Pimentel, J.; Pardo, J.M. HMM Adaptation for Improving a Human Activity Recognition System. Algorithms 2016, 9, 60. [Google Scholar] [CrossRef]
  170. Shen, J.; Fang, H. Human Activity Recognition Using Gaussian Naive Bayes Algorithm in Smart Home. J. Phys. Conf. Ser. 2020, 1631, 12059. [Google Scholar] [CrossRef]
  171. Kose, M.; Incel, O.D.; Ersoy, C. Online Human Activity Recognition on Smart Phones. In Proceedings of the Workshop on Mobile Sensing: From Smartphones and Wearables to Big Data, Pittsburgh, PA, USA, 5–8 September 2012; Volume 16, pp. 11–15. [Google Scholar]
  172. Sarkar, A.M.J.; Lee, Y.-K.; Lee, S. A Smoothed Naive Bayes-Based Classifier for Activity Recognition. IETE Tech. Rev. 2010, 27, 107–119. [Google Scholar] [CrossRef]
  173. Chetty, G.; White, M.; Akther, F. Smart Phone Based Data Mining for Human Activity Recognition. Procedia Comput. Sci. 2015, 46, 1181–1187. [Google Scholar] [CrossRef]
  174. Zhang, W.; Zhao, X.; Li, Z. A Comprehensive Study of Smartphone-Based Indoor Activity Recognition via Xgboost. IEEE Access 2019, 7, 80027–80042. [Google Scholar] [CrossRef]
  175. Bukht, T.F.N.; Jalal, A. A Robust Model of Human Activity Recognition Using Independent Component Analysis and XGBoost. In Proceedings of the 2024 5th International Conference on Advancements in Computational Sciences (ICACS), Dubai, United Arab Emirates, 18–20 March 2024; pp. 1–7. [Google Scholar]
  176. O’Halloran, J.; Curry, E. A Comparison of Deep Learning Models in Human Activity Recognition and Behavioural Prediction on the MHEALTH Dataset. In Proceedings of the AICS, Galway, Ireland, 5–6 December 2019; pp. 212–223. [Google Scholar]
  177. Ayumi, V. Pose-Based Human Action Recognition with Extreme Gradient Boosting. In Proceedings of the 2016 IEEE Student Conference on Research and Development (SCOReD), Malacca, Malaysia, 13–14 December 2016; pp. 1–5. [Google Scholar]
  178. Ambati, L.S.; El-Gayar, O. Human Activity Recognition: A Comparison of Machine Learning Approaches. J. Midwest Assoc. Inf. Syst. 2021, 2021, 4. [Google Scholar]
  179. Sozinov, K.; Vlassov, V.; Girdzijauskas, S. Human Activity Recognition Using Federated Learning. In Proceedings of the 2018 IEEE International Conference on Parallel & Distributed Processing with Applications, Ubiquitous Computing & Communications, Big Data & Cloud Computing, Social Computing & Networking, Sustainable Computing & Communications (ISPA/IUCC/BDCloud/SocialCom/SustainC), Melbourne, Australia, 11–13 December 2018; pp. 1103–1111. [Google Scholar]
  180. Bustoni, I.A.; Hidayatulloh, I.; Ningtyas, A.M.; Purwaningsih, A.; Azhari, S.N. Classification Methods Performance on Human Activity Recognition. J. Phys. Conf. Ser. 2020, 1456, 12027. [Google Scholar] [CrossRef]
  181. Halim, N. Stochastic Recognition of Human Daily Activities via Hybrid Descriptors and Random Forest Using Wearable Sensors. Array 2022, 15, 100190. [Google Scholar] [CrossRef]
  182. Wang, L. Recognition of Human Activities Using Continuous Autoencoders with Wearable Sensors. Sensors 2016, 16, 189. [Google Scholar] [CrossRef]
  183. Shi, D.; Li, Y.; Ding, B. Unsupervised Feature Learning for Human Activity Recognition. Guofang Keji Daxue Xuebao/J. Natl. Univ. Def. Technol. 2015, 37, 128–134. [Google Scholar] [CrossRef]
  184. Jamel, A.A.M.; Akay, B. Human Activity Recognition Based on Parallel Approximation Kernel K-Means Algorithm. Comput. Syst. Sci. Eng. 2020, 35, 441. [Google Scholar] [CrossRef]
  185. Manzi, A.; Dario, P.; Cavallo, F. A Human Activity Recognition System Based on Dynamic Clustering of Skeleton Data. Sensors 2017, 17, 1100. [Google Scholar] [CrossRef]
  186. Gaglio, S.; Re, G.L.; Morana, M. Human Activity Recognition Process Using 3-D Posture Data. IEEE Trans. Human-Machine Syst. 2015, 45, 586–597. [Google Scholar] [CrossRef]
  187. Colpas, P.A.; Vicario, E.; De-La-Hoz-Franco, E.; Pineres-Melo, M.; Oviedo-Carrascal, A.; Patara, F. Unsupervised Human Activity Recognition Using the Clustering Approach: A Review. Sensors 2020, 20, 2702. [Google Scholar] [CrossRef]
  188. Wang, T.; Ng, W.W.Y.; Li, J.; Wu, Q.; Zhang, S.; Nugent, C.; Shewell, C. A Deep Clustering via Automatic Feature Embedded Learning for Human Activity Recognition. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 210–223. [Google Scholar]
  189. Mejia-Ricart, L.F.; Helling, P.; Olmsted, A. Evaluate Action Primitives for Human Activity Recognition Using Unsupervised Learning Approach. In Proceedings of the 2017 12th International Conference for Internet Technology and Secured Transactions (ICITST), London, UK, 11–14 December 2017; pp. 186–188. [Google Scholar]
  190. Dobbins, C.; Rawassizadeh, R. Towards Clustering of Mobile and Smartwatch Accelerometer Data for Physical Activity Recognition. Informatics 2018, 5, 29. [Google Scholar] [CrossRef]
  191. Piyathilaka, L.; Kodagoda, S. Gaussian Mixture Based HMM for Human Daily Activity Recognition Using 3D Skeleton Features. In Proceedings of the 2013 IEEE 8th Conference on Industrial Electronics and Applications (ICIEA), Melbourne, Australia, 19–21 June 2013; pp. 567–572. [Google Scholar]
  192. Cheng, X.; Huang, B. CSI-Based Human Continuous Activity Recognition Using GMM–HMM. IEEE Sens. J. 2022, 22, 18709–18717. [Google Scholar]
  193. Yang, X. Personalized Semi-Supervised Federated Learning for Human Activity Recognition. arXiv 2021, arXiv:2104.08094. [Google Scholar]
  194. Yang, S.-H.; Baek, D.-G.; Thapa, K. Semi-Supervised Adversarial Learning Using LSTM for Human Activity Recognition. Sensors 2022, 22, 4755. [Google Scholar] [CrossRef]
  195. Cook, D.; Crandall, A.; Thomas, B.; Krishnan, N. CASAS: A Smart Home in a Box. Computer 2013, 46, 62–69. [Google Scholar] [CrossRef]
  196. Oh, S.; Ashiquzzaman, A.; Lee, D.; Kim, Y.; Kim, J. Study on Human Activity Recognition Using Semi-Supervised Active Transfer Learning. Sensors 2021, 21, 2760. [Google Scholar] [CrossRef]
  197. Qu, Y.; Tang, Y.; Yang, X.; Wen, Y.; Zhang, W. Context-Aware Mutual Learning for Semi-Supervised Human Activity Recognition Using Wearable Sensors. Expert Syst. Appl. 2023, 219, 119679. [Google Scholar] [CrossRef]
  198. Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A Public Domain Dataset for Human Activity Recognition Using Smartphones. In Proceedings of the ESANN, Bruges, Belgium, 24–26 April 2013; Volume 3, p. 3. [Google Scholar]
  199. Wan, S.; Qi, L.; Xu, X.; Tong, C.; Gu, Z. Deep Learning Models for Real-Time Human Activity Recognition with Smartphones. Mob. Netw. Appl. 2020, 25, 743–755. [Google Scholar] [CrossRef]
  200. Banos, O.; Garcia, R.; Holgado-Terriza, J.A.; Damas, M.; Pomares, H.; Rojas, I.; Saez, A.; Villalonga, C. MHealthDroid: A Novel Framework for Agile Development of Mobile Health Applications. In Ambient Assisted Living and Daily Activities; Pecchia, L., Chen, L.L., Nugent, C., Bravo, J., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 91–98. [Google Scholar]
  201. Kim, D.; Han, S.; Son, H.; Lee, D. Human Activity Recognition Using Semi-Supervised Multi-Modal DEC for Instagram Data; Springer: Cham, Switzerland, 2020; pp. 869–880. ISBN 978-3-030-47425-6. [Google Scholar]
  202. Paola Patricia, A.C.; Rosberg, P.C.; Butt-Aziz, S.; Marlon Alberto, P.M.; Roberto-Cesar, M.O.; Miguel, U.T.; Naz, S. Semi-Supervised Ensemble Learning for Human Activity Recognition in Casas Kyoto Dataset. Heliyon 2024, 10, e29398. [Google Scholar] [CrossRef]
  203. Zeng, M.; Yu, T.; Wang, X.; Nguyen, L.T.; Mengshoel, O.J.; Lane, I. Semi-Supervised Convolutional Neural Networks for Human Activity Recognition. In Proceedings of the 2017 IEEE International Conference on Big Data (Big Data), Boston, MA, USA, 11–14 December 2017; pp. 522–529. [Google Scholar]
  204. Zilelioglu, H.; Khodabandelou, G.; Chibani, A.; Amirat, Y. Semisupervised Generative Adversarial Networks with Temporal Convolutions for Human Activity Recognition. IEEE Sens. J. 2023, 23, 12355–12369. [Google Scholar] [CrossRef]
  205. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3d Human Pose Estimation in the Wild Using Improved Cnn Supervision. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; pp. 506–516. [Google Scholar]
  206. Ronao, C.A.; Cho, S.-B. Human Activity Recognition with Smartphone Sensors Using Deep Learning Neural Networks. Expert Syst. Appl. 2016, 59, 235–244. [Google Scholar] [CrossRef]
  207. Zhang, S.; Li, Y.; Zhang, S.; Shahabi, F.; Xia, S.; Deng, Y.; Alshurafa, N. Deep Learning in Human Activity Recognition with Wearable Sensors: A Review on Advances. Sensors 2022, 22, 1476. [Google Scholar] [CrossRef]
  208. Wang, H.; Zhao, J.; Li, J.; Tian, L.; Tu, P.; Cao, T.; An, Y.; Wang, K.; Li, S. Wearable Sensor-Based Human Activity Recognition Using Hybrid Deep Learning Techniques. Secur. Commun. Netw. 2020, 2020, 2132138. [Google Scholar] [CrossRef]
  209. Park, S.U.; Park, J.H.; Al-Masni, M.A.; Al-Antari, M.A.; Uddin, M.Z.; Kim, T.S. A Depth Camera-Based Human Activity Recognition via Deep Learning Recurrent Neural Network for Health and Social Care Services. Procedia Comput. Sci. 2016, 100, 78–84. [Google Scholar] [CrossRef]
  210. Ren, B.; Liu, M.; Ding, R.; Liu, H. A Survey on 3D Skeleton-Based Action Recognition Using Learning Method. arXiv 2020, arXiv:2002.05907. [Google Scholar]
  211. Schuster, M.; Paliwal, K.K. Bidirectional Recurrent Neural Networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  212. Liu, Y.; Lu, Z.; Li, J.; Yang, T.; Yao, C. Global Temporal Representation Based CNNs for Infrared Action Recognition. IEEE Signal Process. Lett. 2018, 25, 848–852. [Google Scholar]
  213. Cardenas, E.J.E.; Chavez, G.C. Multimodal Human Action Recognition Based on a Fusion of Dynamic Images Using CNN Descriptors. In Proceedings of the SIBGRAPI, Foz do Iguaçu, Brazil, 7–10 October 2018. [Google Scholar]
  214. Bianchi, V.; Bassoli, M.; Lombardo, G.; Fornacciari, P.; Mordonini, M.; De Munari, I. IoT Wearable Sensor and Deep Learning: An Integrated Approach for Personalized Human Activity Recognition in a Smart Home Environment. IEEE Internet Things J. 2019, 6, 8553–8562. [Google Scholar]
  215. Herath, S.; Harandi, M.; Porikli, F. Going Deeper into Action Recognition: A Survey. Image Vis. Comput. 2017, 60, 4–21. [Google Scholar] [CrossRef]
  216. Nguyen, H.C.; Nguyen, T.H.; Scherer, R.; Le, V.H. Deep Learning for Human Activity Recognition on 3D Human Skeleton: Survey and Comparative Study. Sensors 2023, 23, 5121. [Google Scholar] [CrossRef]
  217. Wang, P.; Li, W.; Gao, Z.; Zhang, Y.; Tang, C.; Ogunbona, P. Scene Flow to Action Map: A New Representation for Rgb-d Based Action Recognition with Convolutional Neural Networks. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  218. Ke, Q.; An, S.; Bennamoun, M.; Sohel, F.; Boussaid, F. Skeletonnet: Mining Deep Part Features for 3-D Action Recognition. IEEE signal Process. lett. 2017, 24, 731–735. [Google Scholar] [CrossRef]
  219. Rahmani, H.; Bennamoun, M. Learning Action Recognition Model from Depth and Skeleton Videos. In Proceedings of the ICCV, Venice, Italy, 22–29 October 2017. [Google Scholar]
  220. Yan, S.; Xiong, Y.; Lin, D. Spatial Temporal Graph Convolutional Networks for Skeleton-Based Action Recognition. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence and Thirtieth Innovative Applications of Artificial Intelligence Conference and Eighth AAAI Symposium on Educational Advances in Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; AAAI Press: Washington, DC, USA, 2018. [Google Scholar]
  221. Wang, L.; Xiong, Y.; Wang, Z.; Qiao, Y.; Lin, D.; Tang, X.; Van Gool, L. Temporal Segment Networks for Action Recognition in Videos. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 2740–2755. [Google Scholar]
  222. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Actional-Structural Graph Convolutional Networks for Skeleton-Based Action Recognition. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 3590–3598. [Google Scholar]
  223. Song, Y.-F.; Zhang, Z.; Shan, C.; Wang, L. Richly Activated Graph Convolutional Network for Robust Skeleton-Based Action Recognition. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 1915–1925. [Google Scholar] [CrossRef]
  224. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-Stream Adaptive Graph Convolutional Networks for Skeleton-Based Action Recognition. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 12018–12027. [Google Scholar]
  225. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Skeleton-Based Action Recognition with Directed Graph Neural Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 7904–7913. [Google Scholar]
  226. Yang, H.; Yan, D.; Zhang, L.; Sun, Y.; Li, D.; Maybank, S.J. Feedback Graph Convolutional Network for Skeleton-Based Action Recognition. IEEE Trans. Image Process. 2022, 31, 164–175. [Google Scholar] [CrossRef]
  227. Cheng, K.; Zhang, Y.; He, X.; Chen, W.; Cheng, J.; Lu, H. Skeleton-Based Action Recognition with Shift Graph Convolutional Network. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 13–19 June 2020; pp. 180–189. [Google Scholar]
  228. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Decoupled Spatial-Temporal Attention Network for Skeleton-Based Action-Gesture Recognition. In Proceedings of the Asian Conference on Computer Vision, Macau, China, 1–4 December 2022; pp. 38–53. [Google Scholar]
  229. De Smedt, Q.; Wannous, H.; Vandeborre, J.-P.; Guerry, J.; Le Saux, B.; Filliat, D. Shrec’17 Track: 3d Hand Gesture Recognition Using a Depth and Skeletal Dataset. In Proceedings of the 3DOR—10th Eurographics Workshop on 3D Object Retrieval, Lyon, France, 26 April 2017; pp. 1–6. [Google Scholar]
  230. De Smedt, Q.; Wannous, H.; Vandeborre, J.-P. Skeleton-Based Dynamic Hand Gesture Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 27–30 June 2016; pp. 1–9. [Google Scholar]
  231. Liu, Z.; Zhang, H.; Chen, Z.; Wang, Z.; Ouyang, W. Disentangling and Unifying Graph Convolutions for Skeleton-Based Action Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 143–152. [Google Scholar]
  232. Kwon, H.; Kim, M.; Kwak, S.; Cho, M. Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 13045–13055. [Google Scholar]
  233. Davoodikakhki, M.; Yin, K. Hierarchical Action Classification with Network Pruning. In Advances in Visual Computing, Proceedings of the 15th International Symposium (ISVC 2020), San Diego, CA, USA, 5–7 October 2020; Proceedings, Part I 15; Springer International Publishing: New York, NY, USA, 2020; pp. 291–305. [Google Scholar]
  234. Das, S.; Dai, R.; Yang, D.; Bremond, F. Vpn++: Rethinking Video-Pose Embeddings for Understanding Activities of Daily Living. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 9703–9717. [Google Scholar]
  235. Das, S.; Dai, R.; Koperski, M.; Minciullo, L.; Garattoni, L.; Bremond, F.; Francesca, G. Toyota Smarthome: Real-World Activities of Daily Living. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  236. Liu, Z.; Ning, J.; Cao, Y.; Wei, Y.; Zhang, Z.; Lin, S.; Hu, H. Video Swin Transformer. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 3192–3201. [Google Scholar]
  237. Duan, H.; Zhao, Y.; Xiong, Y.; Liu, W.; Lin, D. Omni-Sourced Webly-Supervised Learning for Video Recognition. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 670–688. [Google Scholar]
  238. Cook, D.; Feuz, K.D.; Krishnan, N.C. Transfer Learning for Activity Recognition: A Survey. Knowl. Inf. Syst. 2013, 36, 537–556. [Google Scholar] [CrossRef] [PubMed]
  239. Ranasinghe, K.; Naseer, M.; Khan, S.; Khan, F.S.; Ryoo, M.S. Self-Supervised Video Transformer. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  240. Bertasius, G.; Wang, H.; Torresani, L. Is Space-Time Attention All You Need for Video Understanding? In Proceedings of the ICML, Virtual, 18–24 July 2021. [Google Scholar]
  241. Yang, J.; Dong, X.; Liu, L.; Zhang, C.; Shen, J.; Yu, D. Recurring the Transformer for Video Action Recognition. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  242. Carreira, J.; Zisserman, A. Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. In Proceedings of the CVPR, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  243. Girdhar, R.; Joao Carreira, J.; Doersch, C.; Zisserman, A. Video Action Transformer Network. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 244–253. [Google Scholar]
  244. Li, S.; Cao, Q.; Liu, L.; Yang, K.; Liu, S.; Hou, J.; Yi, S. GroupFormer: Group Activity Recognition with Clustered Spatial-Temporal Transformer. In Proceedings of the ICCV, Virtual, 11–17 October 2021. [Google Scholar]
  245. Choi, W.; Shahid, K.; Savarese, S. What Are They Doing?: Collective Activity Classification Using Spatio-Temporal Relationship among People. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision Workshops (ICCV Workshops), Kyoto, Japan, 27 September–4 October 2009; pp. 1282–1289. [Google Scholar] [CrossRef]
  246. Neimark, D.; Bar, O.; Zohar, M.; Asselmann, D. Video Transformer Network. In Proceedings of the ICCV, Virtual, 11–17 October 2021. [Google Scholar]
  247. Zhang, Y.; Li, X.; Liu, C.; Shuai, B.; Zhu, Y.; Brattoli, B.; Chen, H.; Marsic, I.; Tighe, J. Vidtr: Video Transformer Without Convolutions. In Proceedings of the ICCV, Virtual, 11–17 October 2021. [Google Scholar]
  248. Kuehne, H.; Jhuang, H.; Garrote, E.; Poggio, T.; Serre, T. HMDB: A Large Video Database for Human Motion Recognition. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2556–2563. [Google Scholar]
  249. Bulat, A.; Perez Rua, J.M.; Sudhakaran, S.; Martinez, B.; Tzimiropoulos, G. Space-Time Mixing Attention for Video Transformer. Adv. Neural Inf. Process. Syst. 2021, 34, 19594–19607. [Google Scholar]
  250. Patrick, M.; Campbell, D.; Asano, Y.; Misra, I.; Metze, F.; Feichtenhofer, C.; Vedaldi, A.; Henriques, J.F. Keeping Your Eye on the Ball: Trajectory Attention in Video Transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12493–12506. [Google Scholar]
  251. Fan, H.; Xiong, B.; Mangalam, K.; Li, Y.; Yan, Z.; Malik, J.; Feichtenhofer, C. Multiscale Vision Transformers. In Proceedings of the ICCV, Virtual, 11–17 October 2021. [Google Scholar]
  252. Arnab, A.; Dehghani, M.; Heigold, G.; Sun, C.; Lučić, M.; Schmid, C. Vivit: A Video Vision Transformer. In Proceedings of the ICCV, Virtual, 11–17 October 2021. [Google Scholar]
  253. Ryoo, M.S.; Piergiovanni, A.J.; Arnab, A.; Dehghani, M.; Angelova, A. TokenLearner: What Can 8 Learned Tokens Do for Images and Videos? arXiv 2021, arXiv:2106.11297. [Google Scholar]
  254. Piergiovanni, A.J.; Ryoo, M. Avid Dataset: Anonymized Videos from Diverse Countries. Adv. Neural Inf. Process. Syst. 2020, 33, 16711–16721. [Google Scholar]
  255. Zha, X.; Zhu, W.; Xun, L.T.; Yang, S.; Liu, J. Shifted Chunk Transformer for Spatio-Temporal Representational Learning. In Proceedings of the NeurIPS, Virtual, 6–14 December 2021. [Google Scholar]
  256. Zhang, F.Z.; Campbell, D.; Gould, S. Efficient Two-Stage Detection of Human-Object Interactions with a Novel Unary-Pairwise Transformer. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  257. Guo, H.; Wang, H.; Ji, Q. Uncertainty-Guided Probabilistic Transformer for Complex Action Recognition. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  258. Chen, C.-F.; Panda, R.; Fan, Q. Regionvit: Regional-to-Local Attention for Vision Transformers. In Proceedings of the ICLR, Virtual, 25–29 April 2022. [Google Scholar]
  259. Truong, T.-D.; Bui, Q.-H.; Duong, C.N.; Seo, H.-S.; Phung, S.L.; Li, X.; Luu, K. Direcformer: A Directed Attention in Transformer Approach to Robust Action Recognition. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  260. Li, K.; Wang, Y.; Gao, P.; Song, G.; Liu, Y.; Li, H.; Qiao, Y. Uniformer: Unified Transformer for Efficient Spatiotemporal Representation Learning. In Proceedings of the ICLR, Virtual, 25–29 April 2022. [Google Scholar]
  261. Fan, Q.; Chen, C.-F.; Panda, R. Can an Image Classifier Suffice For Action Recognition? In Proceedings of the ICLR, Virtual, 25–29 April 2022. [Google Scholar]
  262. Yan, S.; Xiong, X.; Arnab, A.; Lu, Z.; Zhang, M.; Sun, C.; Schmid, C. Multiview Transformers for Video Recognition. In Proceedings of the CVPR, New Orleans, LA, USA, 19–24 June 2022. [Google Scholar]
  263. Tong, Z.; Song, Y.; Wang, J.; Wang, L. Videomae: Masked Autoencoders Are Data-Efficient Learners for Self-Supervised Video Pre-Training. Adv. Neural Inf. Process. Syst. 2022, 35, 10078–10093. [Google Scholar]
  264. Feichtenhofer, C.; Li, Y.; He, K. Masked Autoencoders as Spatiotemporal Learners. Adv. Neural Inf. Process. Syst. 2022, 35, 35946–35958. [Google Scholar]
  265. Wang, J.; Bertasius, G.; Tran, D.; Torresani, L. Long-Short Temporal Contrastive Learning of Video Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 14010–14020. [Google Scholar]
  266. Wang, R.; Chen, D.; Wu, Z.; Chen, Y.; Dai, X.; Liu, M.; Jiang, Y.-G.; Zhou, L.; Yuan, L. Bevt: Bert Pretraining of Video Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 14733–14743. [Google Scholar]
  267. Girdhar, R.; El-Nouby, A.; Singh, M.; Alwala, K.V.; Joulin, A.; Misra, I. Omnimae: Single Model Masked Pretraining on Images and Videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 10406–10417. [Google Scholar]
  268. Sun, X.; Chen, P.; Chen, L.; Li, C.; Li, T.H.; Tan, M.; Gan, C. Masked Motion Encoding for Self-Supervised Video Representation Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 2235–2245. [Google Scholar]
  269. Fan, D.; Wang, J.; Liao, S.; Zhu, Y.; Bhat, V.; Santos-Villalobos, H.; MV, R.; Li, X. Motion-Guided Masking for Spatiotemporal Representation Learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Vancouver, BC, Canada, 18–22 June 2023; pp. 5619–5629. [Google Scholar]
  270. Nikpour, B.; Sinodinos, D.; Armanfard, N. Deep Reinforcement Learning in Human Activity Recognition: A Survey and Outlook. IEEE Trans. Neural Networks Learn. Syst. 2024, 36, 4267–4278. [Google Scholar]
  271. Laptev, I.; Marszalek, M.; Schmid, C.; Rozenfeld, B. Learning Realistic Human Actions from Movies. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  272. Klaser, A.; Marszałek, M.; Schmid, C. A Spatio-Temporal Descriptor Based on 3d-Gradients. In Proceedings of the BMVC; 2008. [Google Scholar]
  273. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  274. Liu, Z.; Yao, C.; Yu, H.; Wu, T. Deep Reinforcement Learning with Its Application for Lung Cancer Detection in Medical Internet of Things. Futur. Gener. Comput. Syst. 2019, 97, 1–9. [Google Scholar] [CrossRef]
  275. Shao, K.; Tang, Z.; Zhu, Y.; Li, N.; Zhao, D. A Survey of Deep Reinforcement Learning in Video Games. arXiv 2019, arXiv:1912.10944. [Google Scholar]
  276. Xu, W.; Yu, J.; Miao, Z.; Wan, L.; Ji, Q. Spatio-Temporal Deep Q-Networks for Human Activity Localization. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 2984–2999. [Google Scholar] [CrossRef]
  277. Rodriguez, M.D.; Ahmed, J.; Shah, M. Action Mach a Spatio-Temporal Maximum Average Correlation Height Filter for Action Recognition. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  278. Wang, L.; Qiao, Y.; Tang, X. Video Action Detection with Relational Dynamic-Poselets. In Computer Vision—ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13; Springer: Zurich, Switzerland, 2014; pp. 565–580. [Google Scholar]
  279. Tang, Y.; Tian, Y.; Lu, J.; Li, P.; Zhou, J. Deep Progressive Reinforcement Learning for Skeleton-Based Action Recognition. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  280. Hu, J.-F.; Zheng, W.-S.; Lai, J.; Zhang, J. Jointly Learning Heterogeneous Features for RGB-D Activity Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5344–5352. [Google Scholar]
  281. Wu, W.; He, D.; Tan, X.; Chen, S.; Wen, S. Multi-Agent Reinforcement Learning Based Frame Sampling for Effective Untrimmed Video Recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6222–6231. [Google Scholar]
  282. Lu, Y.; Li, Y.; Velipasalar, S. Efficient Human Activity Classification from Egocentric Videos Incorporating Actor-Critic Reinforcement Learning. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 564–568. [Google Scholar]
  283. Kumrai, T.; Korpela, J.; Maekawa, T.; Yu, Y.; Kanai, R. Human Activity Recognition with Deep Reinforcement Learning Using the Camera of a Mobile Robot. In Proceedings of the 2020 IEEE International Conference on Pervasive Computing and Communications (PerCom), Austin, TX, USA, 23–27 March 2020; pp. 1–10. [Google Scholar]
  284. Brodeur, S.; Perez, E.; Anand, A.; Golemo, F.; Celotti, L.; Strub, F.; Rouat, J.; Larochelle, H.; Courville, A. Home: A Household Multimodal Environment. arXiv 2017, arXiv:1711.11017. [Google Scholar]
  285. Ma, Y.; Arshad, S.; Muniraju, S.; Torkildson, E.; Rantala, E.; Doppler, K.; Zhou, G. Location-and Person-Independent Activity Recognition with WiFi, Deep Neural Networks, and Reinforcement Learning. ACM Trans. Internet Things 2021, 2, 1–25. [Google Scholar]
  286. Yousefi, S.; Narui, H.; Dayal, S.; Ermon, S.; Valaee, S. A Survey on Behavior Recognition Using WiFi Channel State Information. IEEE Commum. Mag. 2017, 55, 98–104. [Google Scholar] [CrossRef]
  287. Palipana, S.; Rojas, D.; Agrawal, P.; Pesch, D. FallDeFi: Ubiquitous Fall Detection Using Commodity Wi-Fi Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 1, 1–25. [Google Scholar]
  288. Gowda, S.N.; Sevilla-Lara, L.; Keller, F.; Rohrbach, M. Claster: Clustering with Reinforcement Learning for Zero-Shot Action Recognition. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 187–203. [Google Scholar]
  289. Niebles, J.C.; Chen, C.-W.; Fei-Fei, L. Modeling Temporal Structure of Decomposable Motion Segments for Activity Classification. In Proceedings of the Computer Vision—ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part II 11. Springer: Berlin/Heidelberg, Germany, 2010; pp. 392–405. [Google Scholar]
  290. Li, Q.; Gravina, R.; Li, Y.; Alsamhi, S.H.; Sun, F.; Fortino, G. Multi-User Activity Recognition: Challenges and Opportunities. Inf. Fusion 2020, 63, 121–135. [Google Scholar] [CrossRef]
  291. Wang, J.; Chen, Y.; Hao, S.; Peng, X.; Hu, L. Deep Learning for Sensor-Based Activity Recognition: A Survey. Pattern Recognit. Lett. 2019, 119, 3–11. [Google Scholar] [CrossRef]
  292. Zhang, H.; Zhang, L.; Qui, X.; Li, H.; Torr, P.H.S.; Koniusz, P. Few-Shot Action Recognition with Permutation-Invariant Attention. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  293. Wei, Y.; Liu, H.; Xie, T.; Ke, Q.; Guo, Y. Spatial-Temporal Transformer for 3D Point Cloud Sequences. In Proceedings of the WACV, Waikoloa, HI, USA, 4–8 January 2022. [Google Scholar]
  294. Dalianis, H. Evaluation Metrics and Evaluation. In Clinical Text Mining; Springer International Publishing: Cham, Switzerland, 2018; pp. 45–53. [Google Scholar]
  295. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  296. Davis, J.; Goadrich, M. The Relationship between Precision-Recall and ROC Curves. In Proceedings of the 23rd International Conference on Machine Learning—ICML ’06, Pittsburgh, PA, USA, 25–29 June 2006; ACM Press: New York, New York, USA, 2006; pp. 233–240. [Google Scholar]
  297. Arbelaitz, O.; Gurrutxaga, I.; Muguerza, J.; Pérez, J.M.; Perona, I. An Extensive Comparative Study of Cluster Validity Indices. Pattern Recognit. 2013, 46, 243–256. [Google Scholar] [CrossRef]
  298. Wijekoon, A.; Wiratunga, N.; Sani, S.; Cooper, K. A Knowledge-Light Approach to Personalised and Open-Ended Human Activity Recognition. Knowl.-Based Syst. 2020, 192, 105651. [Google Scholar] [CrossRef]
  299. Abdallah, Z.S.; Gaber, M.M.; Srinivasan, B.; Krishnaswamy, S. Activity Recognition with Evolving Data Streams: A Review. ACM Comput. Surv. 2018, 51, 1–36. [Google Scholar] [CrossRef]
  300. Saleem, G.; Raza, R.H. Toward Human Activity Recognition: A Survey; Springer: London, UK, 2022; Volume 4, ISBN 0123456789. [Google Scholar]
  301. Hassan, M.M.; Ullah, S.; Hossain, M.S.; Alelaiwi, A. An End-to-End Deep Learning Model for Human Activity Recognition from Highly Sparse Body Sensor Data in Internet of Medical Things Environment. J. Supercomput. 2021, 77, 2237–2250. [Google Scholar] [CrossRef]
  302. Long, X.; Yin, B.; Aarts, R.M. Single-Accelerometer-Based Daily Physical Activity Classification. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC 2009), Minneapolis, MN, USA, 3–6 September 2009; pp. 6107–6110. [Google Scholar] [CrossRef]
  303. Fahad, L.G.; Tahir, S.F.; Rajarajan, M. Activity Recognition in Smart Homes Using Clustering Based Classification. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014; pp. 1348–1353. [Google Scholar]
  304. Ramirez, H.; Velastin, S.A.; Meza, I.; Fabregas, E.; Makris, D.; Farias, G. Fall Detection and Activity Recognition Using Human Skeleton Features. IEEE Access 2021, 9, 33532–33542. [Google Scholar]
  305. Yadav, S.K.; Tiwari, K.; Pandey, H.M.; Akbar, S.A. A Review of Multimodal Human Activity Recognition with Special Emphasis on Classification, Applications, Challenges and Future Directions. Knowl.-Based Syst. 2021, 223, 106970. [Google Scholar] [CrossRef]
  306. De, D.; Bharti, P.; Das, S.K.; Chellappan, S. Multimodal Wearable Sensing for Fine-Grained Activity Recognition in Healthcare. IEEE Internet Comput. 2015, 19, 26–35. [Google Scholar] [CrossRef]
  307. Chen, C.; Jafari, R.; Kehtarnavaz, N. A Survey of Depth and Inertial Sensor Fusion for Human Action Recognition. Multimed. Tools Appl. 2017, 76, 4405–4425. [Google Scholar] [CrossRef]
  308. Majumder, S.; Kehtarnavaz, N. Vision and Inertial Sensing Fusion for Human Action Recognition: A Review. IEEE Sens. J. 2020, 21, 2454–2467. [Google Scholar]
  309. Wei, H.; Kehtarnavaz, N. Simultaneous Utilization of Inertial and Video Sensing for Action Detection and Recognition in Continuous Action Streams. IEEE Sens. J. 2020, 20, 6055–6063. [Google Scholar] [CrossRef]
  310. Islam, M.M.; Nooruddin, S.; Karray, F.; Muhammad, G. Multi-Level Feature Fusion for Multimodal Human Activity Recognition in Internet of Healthcare Things. Inf. Fusion 2023, 94, 17–31. [Google Scholar] [CrossRef]
  311. Rokni, S.A.; Nourollahi, M.; Ghasemzadeh, H. Personalized Human Activity Recognition Using Convolutional Neural Networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar]
  312. Ferrari, A.; Micucci, D.; Mobilio, M.; Napoletano, P. On the Personalization of Classification Models for Human Activity Recognition. IEEE Access 2020, 8, 32066–32079. [Google Scholar] [CrossRef]
  313. Lendínez, A.M.; Linares, C.; Perandres, A.; Cruz, A.; Ruiz, J.L.L.; Nugent, C.; Espinilla, M. Transforming Elderly Care Through Ethical and Social Evaluation of Intelligent Activity Recognition Systems in Nursing Homes. In Proceedings of the International Conference on Ubiquitous Computing and Ambient Intelligence, Salamanca, Spain, 28 November–1 December 2023; pp. 221–227. [Google Scholar]
  314. Jyothi, A.P.; Kumar, M.; Pravallika, V.; Saha, R.; Sanjana, K.; Varshitha, V.; Shankar, A.; Narayan, A. Face and Action Detection for Crime Prevention: Balancing Promise and Ethical Considerations. In Proceedings of the 2023 International Conference on Integrated Intelligence and Communication Systems (ICIICS), Bengaluru, India, 14–16 December 2023; pp. 1–7. [Google Scholar]
  315. Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep Learning for Sensor-Based Human Activity Recognition: Overview, Challenges, and Opportunities. ACM Comput. Surv. 2021, 54, 1–40. [Google Scholar]
  316. Zhu, G.; Zhang, L.; Shen, P.; Song, J. An Online Continuous Human Action Recognition Algorithm Based on the Kinect Sensor. Sensors 2016, 16, 161. [Google Scholar] [CrossRef] [PubMed]
  317. Uddin, M.Z.; Soylu, A. Human Activity Recognition Using Wearable Sensors, Discriminant Analysis, and Long Short-Term Memory-Based Neural Structured Learning. Sci. Rep. 2021, 11, 16455. [Google Scholar] [CrossRef]
  318. Babaians, E.; Korghond, N.K.; Ahmadi, A.; Karimi, M.; Ghidary, S.S. Skeleton and Visual Tracking Fusion for Human Following Task of Service Robots. In Proceedings of the 2015 3rd RSI International Conference on Robotics and Mechatronics (ICROM), Tehran, Iran, 7–9 October 2015; pp. 761–766. [Google Scholar] [CrossRef]
  319. Sargano, A.B.; Wang, X.; Angelov, P.; Habib, Z. Human Action Recognition Using Transfer Learning with Deep Representations. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 463–469. [Google Scholar]
Figure 1. The PRISMA flow diagram used for the systematic identification, screening, eligibility, and inclusion of publications.
Figure 1. The PRISMA flow diagram used for the systematic identification, screening, eligibility, and inclusion of publications.
Jimaging 11 00091 g001
Figure 2. Application of HAR in various fields.
Figure 2. Application of HAR in various fields.
Jimaging 11 00091 g002
Figure 3. Different type of activities from the literature.
Figure 3. Different type of activities from the literature.
Jimaging 11 00091 g003
Figure 4. Stages of Human Activity Recognition (HAR).
Figure 4. Stages of Human Activity Recognition (HAR).
Jimaging 11 00091 g004
Figure 5. Activity recognition.
Figure 5. Activity recognition.
Jimaging 11 00091 g005
Figure 6. Activity discovery.
Figure 6. Activity discovery.
Jimaging 11 00091 g006
Figure 7. Categories of HAR based on sensor modalities.
Figure 7. Categories of HAR based on sensor modalities.
Jimaging 11 00091 g007
Figure 8. Activities performed in a sequence. Top row depicts skeleton-based representation of multiple activities performed sequentially while the bottom row represents hand joint movements over time.
Figure 8. Activities performed in a sequence. Top row depicts skeleton-based representation of multiple activities performed sequentially while the bottom row represents hand joint movements over time.
Jimaging 11 00091 g008
Figure 9. Intraclass variations of fall (a) Fall due to imbalanced leg (b) Fall from a ladder (c) Fall from a slope (d) Fall from stairs (e) Fall due to slippery surface (f) Fall while going downstairs.
Figure 9. Intraclass variations of fall (a) Fall due to imbalanced leg (b) Fall from a ladder (c) Fall from a slope (d) Fall from stairs (e) Fall due to slippery surface (f) Fall while going downstairs.
Jimaging 11 00091 g009
Figure 10. Challenges of vision based HAR (a) Sitting viewed from different angles (b) Sitting viewed from the front (c) Simultaneously sitting and reading (d) Running (e) Jogging (f) Self-occlusion.
Figure 10. Challenges of vision based HAR (a) Sitting viewed from different angles (b) Sitting viewed from the front (c) Simultaneously sitting and reading (d) Running (e) Jogging (f) Self-occlusion.
Jimaging 11 00091 g010
Figure 11. (a) Commercially available wearable device attached to user’s wrist to monitor daily exercise. (b) Skin irritation caused due to long exposure to wearable devices.
Figure 11. (a) Commercially available wearable device attached to user’s wrist to monitor daily exercise. (b) Skin irritation caused due to long exposure to wearable devices.
Jimaging 11 00091 g011
Figure 12. Existing windowing methods commonly applied for segmenting human activity data.
Figure 12. Existing windowing methods commonly applied for segmenting human activity data.
Jimaging 11 00091 g012
Table 1. Unsupervised activity discovery methods.
Table 1. Unsupervised activity discovery methods.
MethodsYearModalityTop AccuracyDatasets Used
Estimation-based [77]2008Wearable72.7[77]
UnADevs using temporal behaviour assumption [73] 2017Wearable52, 61.0JSI-ADL, REALDISP
Stream sequence mining [79,80]2014Smart Home-CASAS
Discover unknown activities [81]2016Smart Home68.01CASAS
Incremental clustering using K-means [82]2013Skeleton71.9CAAD60
Particle swarm [83] 2023Skeleton80.1, 58.10, 64.1, 45.1, 40.1CAD-60, UTK, F3D, KARD, MSR
Long-term visual behaviour [84]2015RGB74.75Gaze [84]
Unsupervised discovery using spectral clustering [85]2006RGB50.01, 52.1, 49.5Figure Skating [85], Baseball, Basketball
HMDSM [86]2019Smart Home91.5UASH [86]
Topic models [87]2014Smart Home85.1[87]
Evaluation of multiple clustering algorithms [76] K-means, Spectral, GMM, 2023Skeletons70.1, 68.05, 74.5CAD60, UTK, UBDKinect
DBSCAN [68]2014Smart phone80.4[68]
Table 2. Various HAR modalities from literature with summarised with pros and cons.
Table 2. Various HAR modalities from literature with summarised with pros and cons.
ModalityDevicesSampleProsCons
Silhouette [88]Different CamerasJimaging 11 00091 i001
  • Provides basic humanoid figure and posture information
  • Computationally efficient
  • Clear representation of human motion
  • Lacks detailed texture and object information
  • Sensitive to occlusions and background noise
  • Challenges in handling variability
RGB Standard RGB camerasJimaging 11 00091 i002
  • Rich visuals with detailed features
  • Widely available dataset and sensing mediums
  • Can be directly fed to deep learning methods for rich feature learning and scene analysis
  • Sensitive to occlusions and background noise and change of view
  • Limited spatial and temporal resolution
  • Invasive to privacy
Infrared [89]Infrared CamerasJimaging 11 00091 i003
  • Suitable for low dark/light environments
  • Less intrusive in terms of privacy compared to visible light cameras
  • Can penetrate through certain materials and surfaces that may hinder visible light such as fog
  • Sensitive to temperature and daylight exposure
  • Lacks detailed texture and object information
  • Costlier than visible light cameras
Depth [90]Depth Cameras such as Microsoft Kinect/Kinect Azure/Jimaging 11 00091 i004
  • Privacy preserving and provides rich 3D structural information.
  • Efficient for separating background and foreground objects
  • Invariant to changes in lighting conditions and partial occlusion
  • Sensitive to daylight
  • Limited range and may produce noisy data, at longer distances.
  • Lacks detailed texture and object information
Skeletons [34]Derived Sensors/RGB CamerasJimaging 11 00091 i005
  • Representing human motion using a simplified skeletal model
  • Invariant to background clutter and variations in lighting conditions and change of view
  • Captures the underlying structure of human motion, including joint movements and temporal sequences
  • Heavily reliant on accurate estimation of joint positions
  • Lacks texture, scene, and object information
  • Skeleton-based methods may struggle to handle missing or inaccurate joint detections
Mocap [91]Commercial Mocap systemsJimaging 11 00091 i006
  • Provides highly accurate and detailed motion data
  • Capturing the movement of individual body parts with precision
  • Action recognition can be used for animation and visualizations of human movement
  • Mocap devices are expensive and costly to maintain
  • Requires individuals to attach markers to the body
  • Multiple sensors are required for capturing data, mostly suitable for controlled environments
Point cloud [92] Depth sensors (like Kinect, LiDAR, range sensors), stereo camerasJimaging 11 00091 i007
  • Captures the 3D spatial information of the scene and subject
  • Robust to occlusions caused by obstacles
  • Point cloud data can be generated from LiDAR sensors or depth sensors
  • May exhibit data sparsity and irregularity
  • Processing and analysing point cloud is computationally intensive
  • Sensitive to calibration and environmental disturbances
Event stream [93]Specialised event camerasJimaging 11 00091 i008
  • High temporal resolutions with reduced visual redundancy
  • Longer viewing and tracking range and insensitive to motion blur
  • Data are generated only when there is a change in the scene
  • Tracking devices are expensive and costly to maintain
  • Difficult to model activities with less motions
  • Multiple sensors are required for capturing data, suitable for controlled environments
Thermal [94] Thermal CamerasJimaging 11 00091 i009
  • Suitable for low dark/light environments and able to penetrate through certain surfaces
  • Less intrusive in terms of privacy compared to visible light cameras
  • Can detect anomalies in human body temperature, such as fever
  • Provides lower spatial resolution compared to visible light cameras and sensitive to temperature difference
  • Lacks detailed texture and object information
  • Costlier than visible light cameras
Audio [95]Microphones, audio recording devicesJimaging 11 00091 i010
  • Offers non-intrusive monitoring of human activities
  • Provides complementary information to visual and motion-based sensing modalities
  • Can be applied in a wide range of environments and contexts
  • Lacks the discriminative power required to distinguish between different activities
  • May raise privacy concerns
  • Susceptible to environmental factors such as background noise, and acoustic interference
Radar [96]Radar devicesJimaging 11 00091 i011
  • Can penetrate through obstacles such as walls and furniture
  • Less affected by environmental conditions such as lighting, weather, and ambient noise
  • High spatial and temporal resolution, capturing detailed information about the position
  • Expensive and complex to deploy compared to other sensing modalities
  • Lacks the discriminative power required to distinguish between different activities
  • May raise privacy concerns
WiFi CSI [97] Standard WiFi routers, network interface cardsJimaging 11 00091 i012
  • Non-intrusive method for monitoring of human activities
  • Provides fine-grained information about the propagation of wireless signals
  • Can be used in ubiquitous indoor environments
  • Limited range and coverage
  • Sensitive to environmental changes such as multipath reflections, obstacle locations
  • May raise privacy concerns
Inertial sensors [98] Smartwatches, fitness trackers, smartphones with IMUs (accelerometers, gyroscopes, magnetometers)Jimaging 11 00091 i013
  • Sensors are compact and lightweight, provided multimodal data such as speed and angle
  • Low-cost devices, fine grained signals, privacy protecting
  • Can capture motion data in real-time, allowing for immediate feedback and analysis of human activities.
  • Lacks contextual information about the environment or activity context
  • Signals are noisy, multiple sensors are required for high label activity recognition
  • Need to attach to body and cause user discomfort
Environmental sensor [39]Sensors embedded in floors or wearables; various proximity sensor devicesJimaging 11 00091 i014
  • Consume low power compared to other sensing modalities
  • Non-intrusive monitoring of human activities
  • Provide contextual information about the surrounding environment
  • Limited discriminative power for distinguishing activities
  • Performance is affected by obstacles and other moving objects
  • Requires cautious planning and installation to ensure optimal sensor placement and coverage
Table 3. Summary of key publicly available HAR datasets.
Table 3. Summary of key publicly available HAR datasets.
DatasetYearModality#Class#Subject#Sample#Viewpoint
UCF101 [99]2012RGB101-13,320-
Sports-1M [100]2014RGB487-1,113,158-
ActivityNet [101]2015RGB203-27,801-
THUMOS Challenge 15 [102]2015RGB101-24,017-
Charades [103]2016RGB1572679848-
InfAR [104]2016IR12406002
NTU RGB+D RGB+D [105]2016RGB, S, D, IR604056,88080
DvsGesture [106]2017Event Stream1729--
FCVID [107]2017RGB239-91,233-
Kinetics-400 [108]2017RGB400-306,245-
PKU-MMD [109]2017RGB, S, D, IR516610763
Something-Something-v1 [110]2017RGB174-108,499-
Kinetics-600 [111]2018RGB600-495,547-
RGB-D Varying-view [112]2018RGB, S, D4011825,6008 + 1 (360°)
DHP19 [94]2019ESS3317-
Drive&Act [113]2019RGB, S, D, IR8315-6
Egogesture [114]2018RGB, D83502,953,224
Moments in time [115]2019RGB339-∼1,000,000-
NTU RGB+D 120 [116]2019RGB, S, D, IR120106114,480155
RareAct [117]2020RGB122-905-
MoVi [118]2021RGB, Mocap21907,344,000+4
UAV-Human [119]2021RGB, S, D, IR, etc.15511967,428-
HOMAGE [120]2021RGB, IR, Au, Ac, Gy86 26,000+5
Ego4D [121]2022RGB, Au, Ac, etc.-923-Egocentric
EPIC-KITCHENS-100 [122]2022RGB, Au, Ac-4589,979Egocentric
EPIC-KITCHENS-55 [123]2023RGB-3972,000,000+Egocentric
Table keys, Gy = Gyroscope, S = Skeleton, D = Depth, Ir = Infrared, Au = Audio, Ac = Acceleration, Ps = Pressure.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossen, M.A.; Abas, P.E. Machine Learning for Human Activity Recognition: State-of-the-Art Techniques and Emerging Trends. J. Imaging 2025, 11, 91. https://doi.org/10.3390/jimaging11030091

AMA Style

Hossen MA, Abas PE. Machine Learning for Human Activity Recognition: State-of-the-Art Techniques and Emerging Trends. Journal of Imaging. 2025; 11(3):91. https://doi.org/10.3390/jimaging11030091

Chicago/Turabian Style

Hossen, Md Amran, and Pg Emeroylariffion Abas. 2025. "Machine Learning for Human Activity Recognition: State-of-the-Art Techniques and Emerging Trends" Journal of Imaging 11, no. 3: 91. https://doi.org/10.3390/jimaging11030091

APA Style

Hossen, M. A., & Abas, P. E. (2025). Machine Learning for Human Activity Recognition: State-of-the-Art Techniques and Emerging Trends. Journal of Imaging, 11(3), 91. https://doi.org/10.3390/jimaging11030091

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop