1. Introduction
Lower limb exoskeletons represent a significant advancement in rehabilitation engineering and assistive technology, serving as wearable robotic devices that can augment, restore, or enhance human locomotor function. These devices have become increasingly important in addressing mobility challenges for individuals with spinal cord injuries, stroke, and other neurological conditions [
1]. Lower limb exoskeletons can be broadly categorized into three types: medical exoskeletons for rehabilitation and mobility assistance, military exoskeletons for load-carrying and endurance enhancement, and industrial exoskeletons for worker support and injury prevention [
2].
In medical applications, these devices demonstrate promise in gait rehabilitation, showing improved outcomes in walking speed, balance, and functional independence among patients [
3]. Recent technological developments have led to lighter, more energy-efficient designs incorporating advanced control strategies like EMG-based systems and adaptive algorithms, making them more practical for daily use [
4]. The applications of these devices extend beyond medical rehabilitation to include performance enhancement in military operations, support for industrial workers in physically demanding tasks, and assistance for elderly individuals in maintaining mobility and independence [
5,
6].
Researchers use different types of vision-based and sensor-based input data for HAR for data collection. Although many research works have talked about the advantages of sensor-based data compared with vision-based data, most state-of-the-art studies still use video cameras (i.e., vision-based) for HAR due to their high accuracy [
7]. Vision-based data collection approaches can be classified into two types: videos and images. For the videos in the HAR literature, collected from CCTV or smartphone devices are used, but for vision-based HAR, social media and camera images are used. On the other hand, mobile and wearable body sensors are the two types of sensor-based data sources that are found in the existing literature. Despite such importance, vision-based data are larger in size and take more processing than sensor-based data. One feasible option is increasing data density through sensors; however, while the cost of sensors has come down significantly over time compared to the alternatives (e.g., vision-based data capturing devices), it remains much more expensive. However, the computability of the body sensor system limits the execution of such a complex algorithm. Thus, sensor-based data are somehow advantageous.
HAR is the human activity detection and identification challenge in different state-of-the-art techniques [
8]. Activities are mainly Activities of Daily Living (ADLs) like walking, jogging, going up or down the stairs or ramp walking. Data availability and its nature are essential parts of a reliable HAR. It has been developed as an essential research domain in ubiquitous computing and human–computer interaction, offering significant implications for healthcare monitoring, assisted living, and rehabilitation applications. The integration of wearable sensors, particularly inertial measurement units (IMUs) comprising accelerometers, gyroscopes, and magnetometers, has revolutionized the ability to accurately detect, classify, and monitor human movements in real-time environments [
9].
Modern HAR systems leverage these sensor networks to capture complex motion patterns and physiological signals, enabling the detection of both basic activities (walking, sitting, standing) and more intricate movements, as can be seen in
Figure 1 [
10].
The advancement of machine learning and deep learning techniques has substantially improved the robustness and accuracy of activity recognition systems [
11]. The authors of [
12] proposed methods for identifying human activities based on a decision tree classifier. However, the classification accuracy rate is considered unsatisfactory. Cheng et al. [
13] proposed three distinct classification methods, such as hidden Markov model, support vector machine, and artificial neural network, to categorize body activities. While these methods deliver acceptable performance, they are either constrained in handling significant intraclass variations or hindered by the complexity of adjusting model parameters. Furthermore, the integration of contextual information and multi-modal sensor fusion techniques has enhanced the system’s ability to distinguish between similar activities and detect transitions between different movement states, making HAR systems increasingly reliable for real-world applications.
Furthermore, Mekruksavanich et al. have introduced a novel deep learning classifier for gym activities named CNN-ResBiGRU. They collected raw EMG and IMU signals from 10 healthy subjects and achieved a classification accuracy of 97.29% [
14]. Zhu et al. [
15] have introduced a load-free hand rehabilitation system based on virtual reality (VR) made from ionic hydrogels. The system can identify 14 hand gestures with an accuracy of 97.9%. Another activity recognition system is developed by Lu et al. [
16]. As they have produced a 5G Narrowband Internet of Things (NB-IoT) system, it is developed for human healthcare data collection, transmission, and reproduction together. The system is integrated with a bionic crack-spring fiber sensor (CSFS) inspired by Cirrus and Spider Structures. This system is characterized by its high sensitivity and long sensing range.
Another study is presented by Mengarelli et al. [
17], this study investigates the feasibility of estimating the vertical component of the ground reaction force (VGRF) using only EMG signals from the thigh and shank muscles. Two deep learning models were used across three experimental setups. The findings demonstrate that EMG signals can be effectively leveraged to estimate VGRF during walking. Tigrini et al. [
18] has proposed a new phasor-based feature extraction approach (PHASOR) that captures spatial myoelectric features to improve the performance of LDA and SVM in gait phase recognition. A publicly available dataset was used to evaluate PHASOR. Additionally, data-driven deep learning architectures, such as Rocket and Mini-Rocket, were included for comparison.
Moreover, myoelectric activity of muscles was used to estimate ankle kinematics as proposed by Mobarak et al. [
19]. sEMG signals were recorded for a total of 288 gait cycles. Two feature sets were extracted from sEMG signals in the time domain (TD) and wavelet (WT) and compared. Then, they were used for feeding three machine learning models (artificial neural networks, random forest, and least squares support vector machine (LS-SVM)).
However, the usage of such highly complex and high processing-based classifiers is considered a costly process and needs a lot of raw data for developing a dependable classifier. At the same time, accurate classification and signal conditioning of raw and complex data such as EMG and IMU signals are a necessity. Normally, open issues and challenges that still exist in previous works related to HAR can be categorized into five categories: data collection, data pre-processing, hardware/sensors used, complex activity discovery, and non-overlap between activities.
The continuous evolution of sensor technology, coupled with sophisticated data processing algorithms, has made HAR an indispensable tool for applications ranging from fall detection in elderly care to performance analysis in sports science [
20]. Recently, to achieve more flexibility in data recording, smartphones and wearables such as wristbands and smartwatches are integrated with sensors to improve the flexibility of data recording further. Such devices were used for daily health and sport users. These HAR systems can also be used for assistive and biomedical devices. For controlling an exoskeleton or artificial limb, EMGs or IMUs are typically the most convenient sensors that can be used for HAR systems. But to utilize these two types, typically a complete framework of data collection, preprocessing, feature extraction and activity classification is utilized.
For the usage of EMGs, the authors in research [
21] proposed a data acquisition system for measuring EMG signals for human lower limb activity recognition. Five leg activities have been accomplished to measure EMG signals from two lower limb muscles to validate the developed hardware. Five subjects were chosen to acquire EMG signals during these activities. The raw EMG signal was first denoised using a hybrid of Wavelet Decomposition with Ensemble Empirical Mode Decomposition (WD-EEMD) approach to classify the recorded EMG dataset. Then, eight time domain (TD) features were extracted using the overlapping windowing technique.
Another example of a complete frame for activity recognition is proposed in [
22] with IMU sensors. The authors have collected data from four classes of body movement datasets, namely stand-up, sit-down, run, and walk. Wearable inertial measurement unit (IMU) sensors were used for sensing and data sampling of human activity. Then, data pre-processing and feature analysis were performed by PCA and the minimum redundancy–maximum relevance (mRMR) feature selection algorithm. Finally, activity recognition was performed by traditional machine learning, deep neural networks, transfer learning and hyperparameter optimization methods. Hence, in our research, the idea of presenting a complete framework for activity recognition for assistive devices is sought after.
The first step is to provide biomechanical data from the HAR sensors. Normally, open-source datasets are used for the verification of classification and signal conditioning techniques. This can be considered the first step to producing real-time experimental hardware for exoskeleton mechanisms. Several open-source datasets were collected for HAR applications.
Moore et al. [
23] presented a dataset with 15 healthy subjects. They were four females and eleven males with an average age of 24 ± 4 years, height of 1.75 ± 0.09 m, and mass of 74 ± 13 kg. The recorded activities are walking at three different speeds (0.8 m s
−1, 1.2 m s
−1 and 1.6 m s
−1). A total of approximately 1.5 h of normal walking and 6 h of perturbed walking are included in this dataset. The trials were performed on an R-Mill treadmill, which has dual six-degree-of-freedom force plates and independent belts for each foot. A USB 6255 card was used as a data acquisition unit. Four ADXL330 three-axis accelerometers were used as wearable sensors. An Osprey camera is used as the motion capture system.
Another open-source dataset is presented by Hu et al. [
24], in which data were selected from 10 healthy subjects. There were seven males and three females. Their biometrics were 25.5 ± 2 years; 174 ± 12 cm; and 70 ± 14 kg for age, height, and weight, respectively.
The utilized sensors were sEMG, IMUs and goniometers. The sEMG (DE2.1 Delsys) was fixed on the following seven muscles in each leg: tibialis anterior (TA), soleus (SOL), vastus lateralis (VL), medial gastrocnemius (MG), rectus femoris (RF), semitendinosus (ST), and biceps femoris (BF). These signals were amplified by 1000×, band-pass filtered between 20 and 450 Hz and sampled at 1 kHz. Additionally, a 6-DOF inertial measurement unit named MPU 9250 IMUs was placed bilaterally on the subjects’ thigh (below RF) and shank (adjacent to TA) and sampled at 500 Hz.
All data were recorded by a 16-bit DAQ unit. The performed activities were a complete circuit of sitting (S), LW, ascending/descending a ramp with a 10° slope (RA/RD), standing (St), and ascending/descending a four-step staircase (SA/SD) step-over-step. A larger dataset was presented by Lencioni et al. [
25], where data were collected from 50 healthy subjects. There were 25 males and 25 females. Their age range, mass and height were 6–72 years, 18.2–110 kg, and 116.6–187.5 cm, respectively. An eight-channel wireless sEMG (ZeroWirePlus) was used. Their signals were band-pass filtered at 10–400 Hz and sampled at 800 Hz, 960 Hz and 1000 Hz. They were applied on the following muscles: tibialis anterior (TA), gastrocnemius medialis (GM), soleus (SO), rectus femoris (RF), peroneus longus (PL), vastus medialis (VM), gluteus maximus (GMax) and biceps femoris (BF). Additional utilized sensors were a 9-camera motion capture system (SMART system) and two force plates (Kistler). The performed activities were walking at different speeds, toe-walking (T), heel-walking (H), step ascending (U) and step descending (D).
Another dataset was presented by Schreiber et al. [
26] with 50 healthy subjects. There were 26 males and 24 females. Their age range, height and weight were 37.0 ± 13.6 years, 1.74 ± 0.09 m, and 71.0 ± 12.3 kg, respectively. A 10-camera optoelectronic system (OQUS4, Qualisys) is used for data acquisition, and the data were sampled at 100 Hz. Ground forces and moments were recorded using two force plates (OR6-5-AMTI). Their data were sampled at 1500 Hz. Eight wireless versions of sEMG (Desktop DTS—Noraxon) were used to collect muscle data from the right leg, and the utilized muscles were gluteus maximus, gluteus medius, vastus medialis, rectus femoris, gastrocnemius medialis, semitendinosus, soleus, and tibialis anterior. Bandpass filtering was applied to these data between 30 and 300 Hz. During a single session, the following exercises were carried out: walking at five different speeds on a level, straight walkway. These speeds were 0–0.4 m. s
−1, 0.4–0.8 m. s
−1, and 0.8–1.2 m. s
−1 in addition to other faster speeds. In total, 1143 trials were completed for all subjects and all activities.
From previous literature, it was proven the necessity of developing data classification and signal conditioning techniques and testing them on open-source datasets. In this way, new algorithms can be tested on previously validated data. Consequently, the new techniques can be applied to new experimental datasets and to human subjects. This paper presents a new methodology for the classification of ADLs using sEMG and IMUs with the intention of achieving high accuracy and low-speed classification.
In this work, our objective is to present a novel autocepstrum-based framework for studying lower limb locomotion. The remarkable characteristic of autocepstrum analysis is enhancing the significant features representing a specific activity while suppressing noise such as additive Gaussian noise according to homomorphic filtering capabilities. The proposed work captures and extracts information from different lower limb muscles to accurately recognize human movement. Indeed, many transfemoral amputees have had their lower limbs removed entirely below the knee because of illness or an accident; our proposed work is greatly inspired by this fact. An open-source dataset has been selected for testing the proposed approach to reduce the complexity of hardware preparations and facilitate algorithm testing. From our observation, wearing many sEMG or IMU sensors may make the wearer uncomfortable and require a lot of data processing and hinder the portability of assistive devices. Hence, we aim to choose between the employment of sEMG and IMU based on the obtained classification accuracy. To ensure recognition accuracy, the number of sensors must be kept to a minimum. This can be fulfilled by deciding on the muscles that have the most effective contribution in identifying a specific activity.
The methodology of this work is presented in
Section 2. Different signal conditioning techniques are presented and applied to IMU and EMG signals in
Section 3. The main proposed framework for activity recognition is shown in
Section 4. The obtained results and their discussion and analysis are presented in
Section 5. The paper is finalized with the conclusion and future work in
Section 6.
5. Data Classification, Results and Discussion
The last step in the proposed framework is data classification for activity recognition. The utilized hardware is a computer with an Intel i7 core processor, 2.6 GHz speed and 16 GB RAM. The utilized benchmark dataset, ‘Georgia Tech dataset’ [
27], consists of three main activities (walking, stairs and ramp) with a total of 2511 samples collected from 22 able-bodied adults for multiple locomotion modes. The sensors are applied over 11 muscles using EMG sensors sampled at 1000 Hz. Inertial measurement unit data are collected from four different muscles: trunk, thigh, shank, and foot segments sampled at 200 Hz. The samples are divided into training and testing samples for different activities, as can be shown in
Table 2. The proportion of training samples to testing samples is 70% to 30%.
The applied classifiers are K-nearest neighbor, neural networks and random forest. For KNN, the number of neighbors is three, equal distance weighting is used and the distance metric technique is Euclidean distance. The utilized neural network is a feedforward NN with 1 hidden layer, ReLU is the activation function, 100 epochs and stochastic gradient descent as an optimizer. The third classifier, random forest, has several trees equal to seven. For results analysis and assessment,
precision (
P),
recall (
R), and
F-measure are calculated. Their equations are shown in Equations (3)–(5) as follows:
where
TP represents the number of correctly detected activities,
FN represents the number of undetected activities, and
FP represents the number of incorrectly detected activities. The full obtained results are shown for EMGs and IMUs in the
Appendix A (
Table A1 and
Table A2), respectively.
For further validation of the obtained results, a threefold K-fold cross-validation technique is used. The dataset is split up into three folds. Each iteration uses a single fold as testing data and the other folds as training data. Consequently, the procedure is repeated until every dataset has been assessed. The mean score of the evaluation metrics values is typically used to represent the K-fold results. The classification results obtained for EMGs and IMUs are shown in
Table 3 and
Table 4, respectively, where the highest five sensors results in terms of P, R and F are displayed with respect to other sensors.
The results for EMGs in
Table 3 show that the random forest algorithm provides higher results than the other techniques. The precision (P) varies from 95.4 up to 98.56%, the recall (R) varies from 94.84 up to 96.75% and the F-measure (F) varies from 95 up to 97.63%. The highest assessment values were achieved for the soleus muscle, where the F-score value was 97.63%. The soleus muscle produces the best results when employing a random forest classifier.
Figure 13a displays the confusion matrix of three activities (walking, ramping, and stairs) using an EMG sensor. According to the results, all samples were accurately recognized for the third activity (stairs); however, a few samples were incorrectly detected for the first and second activities (walking and ramp).
Similarly to EMG, the random forest classifier outperforms the other two classifiers in terms of the aforementioned metrics when utilizing an IMU sensor. The precision (P) varies from 95.03 up to 99.29%, the recall (R) varies from 94.98 up to 98.05% and the F-measure (F) varies from 95.05 up to 98.52%. The best results can be obtained by extracting the gyroscope signal in the Y direction from the shank muscle.
The classification results of the three gyroscope signal activities in the Y direction that were taken from the IMU sensor using a random forest classifier are displayed in the confusion matrix in
Figure 13b. The findings showed that while all samples were correctly identified for the third activity (stairs), several samples were misidentified for the first and second activities (walking and ramp). The results are shown in
Table 4.
The previous results indicate that the best discrimination between the different activities can be achieved by using an EMG signal collected from the soleus muscle or by using a signal extracted from an IMU sensor located on the shank muscle. These obtained results can be compared to other HAR approaches that are mentioned in the literature, as can be seen in
Table 5. Hence, the results achieved can be seen as comparable and successful to other research results.
6. Conclusions and Future Work
In this paper, we have proposed an activity recognition framework based on signal segmentation, decomposition, and feature extraction from sEMG and IMU sensors. This approach is applied to two different types of signals extracted from EMG and IMU sensors. This methodology relies heavily on deep signal conditioning of IMU and EMG signals to pave the way for easily implemented machine learning classifiers. Autocepstrum analysis (ACA) was chosen for signal conditioning after several trials with other techniques like WT, EMD and PCA. Three machine learning classifiers were chosen, and they were able to achieve superior accuracy for locomotion activities, particularly in gait rehabilitation applications. The resultant data were assessed by confusion matrices, precision, recall and F-measure indicators. The data were later validated by the K-fold validation technique.
The results indicate that the random forest classifier performs better than KNN and neural networks across all muscle groups. As for EMG signals, the most accurate results were obtained from the soleus, gracilis, and vastus medialis muscles, with F-scores of 97.63%, 97.11% and 96.66%, respectively. On the other hand, shank and foot signals achieved the highest the F-scores, with 98.52% and 97.63%, respectively. These findings validate the necessity of sensor-based HAR for rehabilitation robotic devices.
Future work will focus on obtaining a new custom dataset for real-time data collection and deep learning-based classification. Further research will include the development of a new data acquisition system with the previously recommended classification framework to be integrated with an actual rehabilitation device to enhance mobility assistance for individuals with neuromuscular impairments. Indeed, high-performance sensors with advanced specifications, such as precise calibration, reliable data transmission, and seamless integration with lower limb rehabilitation devices, must be carefully considered to ensure the accuracy and practicality of sensory data acquisition in real-world applications. Moreover, clinical validation of the acquired data with future rehabilitation devices should be monitored carefully to ensure its effectiveness, safety, and applicability in real-world therapeutic scenarios, particularly for patients with varying levels of mobility impairments. Finally, further investigation of using multimodal sensory data for signal processing and classification is planned.