Next Article in Journal
MONET: The Minor Body Generator Tool at DART Lab
Previous Article in Journal
Linewidth Measurement of a Narrow-Linewidth Laser: Principles, Methods, and Systems
Previous Article in Special Issue
Biomechanical Effects of Using a Passive Exoskeleton for the Upper Limb in Industrial Manufacturing Activities: A Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Effect of Sensor Feature Inputs on Joint Angle Prediction across Simple Movements

1
Department of Mechanical Engineering, Auburn University, Auburn, AL 36849, USA
2
Department of Industrial & Systems Engineering, Auburn University, Auburn, AL 36849, USA
3
Department of Industrial & Systems Engineering and Engineering Management, University of Alabama-Huntsville, Huntsville, AL 35899, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(11), 3657; https://doi.org/10.3390/s24113657
Submission received: 8 May 2024 / Revised: 31 May 2024 / Accepted: 3 June 2024 / Published: 5 June 2024
(This article belongs to the Special Issue Wearable Sensors for Biomechanics Applications—2nd Edition)

Abstract

:
The use of wearable sensors, such as inertial measurement units (IMUs), and machine learning for human intent recognition in health-related areas has grown considerably. However, there is limited research exploring how IMU quantity and placement affect human movement intent prediction (HMIP) at the joint level. The objective of this study was to analyze various combinations of IMU input signals to maximize the machine learning prediction accuracy for multiple simple movements. We trained a Random Forest algorithm to predict future joint angles across these movements using various sensor features. We hypothesized that joint angle prediction accuracy would increase with the addition of IMUs attached to adjacent body segments and that non-adjacent IMUs would not increase the prediction accuracy. The results indicated that the addition of adjacent IMUs to current joint angle inputs did not significantly increase the prediction accuracy (RMSE of 1.92° vs. 3.32° at the ankle, 8.78° vs. 12.54° at the knee, and 5.48° vs. 9.67° at the hip). Additionally, including non-adjacent IMUs did not increase the prediction accuracy (RMSE of 5.35° vs. 5.55° at the ankle, 20.29° vs. 20.71° at the knee, and 14.86° vs. 13.55° at the hip). These results demonstrated how future joint angle prediction during simple movements did not improve with the addition of IMUs alongside current joint angle inputs.

1. Introduction

The use of machine learning for human intent recognition in health-related applications, such as exercise monitoring, human–robotic assistance, and rehabilitation, has grown considerably in recent years [1]. Human movement intent prediction (HMIP), such as future joint angle prediction, is a challenging task, where information of user activity is utilized to determine or estimate future body motion. Oftentimes, HMIP incorporates information about user activity in the form of time-series signals from wearable sensors to approximate limb motion 100 ms into the future [2]. HMIP has applications for human–robot interaction, exoskeleton control, and rehabilitation [3,4]. Determining sensor combinations that optimize HMIP, however, poses a grand challenge in the field of biomechanics [5,6,7,8,9]. There are several factors that influence joint angle prediction from wearable sensors, such as the number of sensors, the types of sensors, and the placement of sensors [10]. Although multiple sensors may be needed to characterize human motion, too many sensors may burden the user, increase the setup time, and require additional computation [10]. Furthermore, an excessive number of physical sensors could alter the user’s natural movement, resulting in movement strategies that differ from their ordinary daily lives.
Wearable sensors, including inertial measurement units (IMUs), provide a portable means of accessing user motion outside the confines of a lab [11,12]. IMUs are widely used to estimate joint, segment, and muscle kinematics and kinetics from accelerations, angular velocities, and local magnetic fields [10,13]. Although IMUs offer a popular and modular solution to HMIP, excessive computational complexity due to additional sensors may delay intent prediction. This can be problematic for real-time exoskeleton applications since exoskeletons must react faster than the onset of physiological reactive joint moments of about 130 ms [14]. Furthermore, real-time controllers of assistive devices have been shown to reduce in functional performance as delays exceed 125 ms and be considered unacceptable for real-time applications with delays above 300 ms [15,16]. Including additional sensors may increase the computational complexity, which may increase the runtime for predicting user intent. Conversely, too few sensors may inadequately capture the user’s intent during complex tasks or task transitions [17,18].
Along with sensor count and type, precise sensor location for extracting relevant physiological signals is essential for effective HMIP [19]. For example, sensors placed on the foot and shank can adequately measure ankle kinematics (r2 of 0.97 for ankle angle) [20]. Sensors are often positioned proximal and distal to the neighboring joint of interest (e.g., thigh–shank for knee kinematics and thigh–waist for hip kinematics) [20,21]. For example, Molinaro et al. (2020) estimated sagittal plane hip torque using data from rotary encoders mounted on the sagittal plane of the hip and bilateral thigh IMUs (3D accelerometer and gyroscope) [21]. Additionally, Dey et al. (2020) used thigh kinematics (angles, angular velocity, and angular acceleration) as inputs to a Random Forest regression to estimate joint angles and moments at the ankle and knee [20].
Previous work using machine learning algorithms from wearable sensors has shown promise for accurate HMIP. For example, Hollinger et al. (2023) used a Random Forest, long short-term memory (LSTM), and bidirectional LSTM to predict lower-limb joint angles 100 ms into the future using EMG, IMU, and joint kinematic inputs. Although the bidirectional LSTM resulted in a mean root mean squared error (RMSE) of 1.42–5.71°, the study was limited to level-ground walking and did not explore feature combinations to reduce the number of input signals [22]. Kazemimoghadam and Fey (2021) examined the effect of training task combinations for task recognition accuracy of transitions and found that including data of the target task in the training data enhanced the prediction accuracy on unanticipated locomotion tasks [17]. Mundt et al. (2020) used an artificial neural network using IMU sensors to predict joint angles with a mean RMSE < 4.8° [23]. Xiong et al. (2019) used an artificial neural network to predict joint angles and moments using EMG and joint angle input features and reported a lower-limb joint moment normalized RMSE (NRMSE) < 7.89% [24]. Yun et al. (2014) used a statistical stochastic approach called Gaussian process regression with anthropometry (14 body parameters) as input features to predict joint motions with a mean error < 4.3° at the ankle [25]. Farmer et al. (2014) used a nonlinear autoregressive model with EMG and kinematic inputs to estimate the future prosthetic ankle angle in three transtibial amputees [26]. The prior work mentioned here clearly shows how machine learning can accurately predict joint angles. However, what remains unknown is the optimal set of feature combinations from wearable sensors that would result in the greatest prediction accuracy while simultaneously limiting the number of attached sensors for the user.
Oftentimes, it is preferred to use a subset of sensor inputs if the reduced number of inputs provides a similar performance. For example, joint angle inputs from the leading or trailing limb have been reported to result in the greatest prediction accuracy and did not improve with the addition of gyroscope and acceleration signals for predicting level-ground straight walking, turns, and transitioning from walking to stair ascent [18]. Certain clinical applications may also have limited time and resources and could benefit from using a sensor configuration that can be easily collected in the clinic [27]. Bao and Intille (2004) found that two accelerometers placed on the wrist and thigh resulted in comparable accuracy to five accelerometers placed on the hip, wrist, arm, ankle, and thigh for classifying twenty everyday tasks [28]. Therefore, HMIP is often a balance between achieving high-level accuracy while simultaneously minimizing the number of sensors worn to maximize practical adoption.
A hybrid approach using a combination of subject-independent and subject-dependent training data has also shown promise for lower-limb joint-level prediction [29]. Combining subject-dependent (individual-specific) and subject-independent (inter-individual) data from shank, knee, and hip kinematics and anthropometry as inputs for predicting ankle torque also highlights how machine learning algorithm accuracy is improved by the inclusion of multiple participants’ data. Dey et al. (2021) found that incorporating personalized data for training a machine learning algorithm resulted in the best accuracy compared to the baseline user-independent model [29]. However, Dey et al. (2021) did not examine the feature combinations of the input data, which could offer more granular-level information for HMIP. Kazemimoghadam and Fey (2021) observed signals of the lower limb and torso but did not compare the combinations of input signals to optimize prediction accuracy [17]. They did, however, recommend future studies to identify top input signals contributing to accurate prediction of the user’s intended tasks. Despite a plethora of research studies dedicated to using machine learning for HMIP, it is still unclear if the number and location of sensors optimize the prediction accuracy. Therefore, the objective of this study was to analyze various combinations of input signals that maximize the machine learning prediction accuracy of joint angles for multiple actions, focusing on simple movement tasks, such as knee flexion, knee extension, ankle plantarflexion, ankle dorsiflexion, hip flexion, and hip extension, while minimizing the required number of sensors to do so. In this study, we trained and evaluated a Random Forest algorithm to predict user intent 100 ms into the future across simple movement tasks from various combinations of sensor feature inputs. Unlike prior studies, which largely focused on prediction accuracy with various algorithms, this study seeks to determine the optimal set of feature combinations from wearable sensors to maximize the prediction accuracy while minimizing the number of sensors. This study begins with simple movements as a starting point and differs from studies that only analyze complex tasks. This approach allows for a more controlled evaluation of how sensor number and location affect joint angle prediction. Although deep learning models, such as LSTM and CNN, have been shown to produce excellent prediction accuracy for HMIP, their complexity makes it challenging to interpret results and use them for real-time applications where low latency is crucial [30,31,32]. Therefore, we decided to use Random Forest for HMIP because it has an excellent tradeoff of speed, accuracy, interpretability, and simplicity [32,33].
Determining IMU quantity and placement that accurately predict joint angles has several practical applications for improving functional outcomes during rehabilitation. Specifically, stroke survivors may use joint angle prediction for personalized rehabilitation protocols for enhancing healthy gait biomechanics [34]. Athletes recovering from sports injuries can also benefit from accurate joint angle prediction of their movements to detect the risk of future injury from compensatory movement strategies [35]. Finally, since hip fractures were reported in 1–2% of falls among the elderly population, accurate joint angle prediction can potentially improve exercise protocols that enhance mobility and mitigate the risk of falling [36].
As mentioned by Chen et al. (2023), sensor fusion methods based on static and quasi-static conditions are still needed for estimating joint angles [37]. This study, therefore, focused on simplistic movements to assess the initial capabilities of wearable sensors for HMIP under quasi-static conditions. We aimed to start with these basic motions to lay the foundation for the system’s performance before estimating joint angles in more complex scenarios, such as walking, jogging, and stair climbing. We hypothesized that the prediction accuracy of ankle, knee, and hip angles across six isolated movements would increase with the addition of IMUs attached to adjacent body segments. We also hypothesized that the addition of IMUs attached to non-adjacent body segments would decrease the prediction accuracy.

2. Materials and Methods

Twenty-eight participants (16 males, 12 females, aged 21.96 ± 2.83 years; height 1.73 ± 0.10 m; weight 71.80 ± 11.86 kg) provided written informed consent to volunteer in the study at the Auburn University Biomechanical Engineering (AUBE) Laboratory. The test protocol was approved by the Auburn University IRB (protocol no. 17-279-MR 1707) and followed the same experimental protocol as Hollinger et al. (2023) [22]. However, in this study, we analyzed eight IMUs (Trigno IM Sensors, version 3.5, Delsys Inc., Boston, MA, USA) attached bilaterally to segments of the torso, thigh, shank, and dorsal aspect of the foot (Figure 1). The IMU sensors recorded three-dimensional acceleration and angular velocity sampled at 148 Hz. Delsys Trigno IM Sensors were chosen because they provide reliable measurements of triaxial accelerometry and angular velocity, which are essential for accurate joint angle prediction. Since these sensors are commonly used in industry, clinical, and academic research, they improved the overall comparability of our study with existing literature and ensured reproducibility for future studies.
Synchronization between motion capture and IMU sensors was performed similar to the methodology explained by Hollinger et al. (2023) [22]. The overall testing workflow is shown in Figure 2, where joint angles derived from motion capture were treated as ground truth values and compared to joint angle predictions (Figure 2). The Random Forest predictor was fed 50 ms of input data to forecast joint angles 100–150 ms later. Following a prediction, the input data shifted in increments of 50 ms and the prediction process was repeated until the end of the movement trial. Random Forest was chosen for its high estimation accuracy and robustness to noise and outliers, which make it ideal for handling nonlinear relationships between inertial data and joint angles [38]. Since Random Forest is an ensemble of decision trees with efficient parallelization, the model is computationally efficient for making quick predictions [38]. This aspect is an important consideration for real-time applications where fast predictions are essential. Participants performed three repetitions of six joint-level movements. The movements were ankle plantarflexion, ankle dorsiflexion, knee flexion, knee extension, hip flexion, and hip extension. We performed subject-dependent training and testing by randomly selecting a single limb and two repetitions of a simple movement for training the Random Forest predictor. The predictor then performed testing on the third trial, which was excluded from the training set of the same participant. For 28 participants and 6 simple movements, this approach resulted in 168 trained models. The Random Forest hyperparameters were chosen from default values provided by Python’s scikit-learn library version 1.4.2 (n_estimators = 100, criterion = ‘squared_error’0, max_depth = none, min_samples_split = 2, and min_samples_leaf = 1). The default settings were chosen so we could analyze the most simplistic scenario of a Random Forest configuration without the influence of hyperparameter tuning.
Sensor features were systematically selected as inputs to the Random Forest algorithm. The selected features varied by number and type. The number of sensors varied to include a full set of IMUs on the ipsilateral single side of the joint (four IMUs) versus a reduced set of IMUs (two IMUs; Figure 3). Since we randomly tested on a single side for joint angle prediction, the maximum sensor configuration consisted of four IMUs. The rationale for comparing a full set of IMUs on the ipsilateral side to a reduced set of IMUs was to strategically compare sensors per body segment with one sensor per adjacent body segment. Although the addition of IMUs attached to non-adjacent segments did not exhibit the same amount of motion as IMUs attached to adjacent segments neighboring the joint, some movement still occurred. For instance, the foot IMU moves as the user lifts their foot in the air during knee extension and flexion and hip extension and flexion. Therefore, even though it may seem as though non-adjacent IMUs did not provide any relevance for predicting joint angles, information may potentially be lost by ignoring the IMUs attached to these segments.
The types of features in each feature set were also varied based on three levels: joint angles (θ), IMU, and IMU + θ. The overall number of feature inputs ranged from 1 input (joint angles, θ) to 25 input signals comprised of 4 IMU sensors (Acceleration X, Acceleration Y, Acceleration Z, Angular Velocity X, Angular Velocity Y, and Angular Velocity Z per IMU sensor) and joint angles (θ). Although obtaining joint angles from optical motion capture is not practical for real-time applications, the data could be derived through an onboard potentiometer or rotary encoder. RMSE was the chosen evaluation metric for prediction performance because the metric penalizes outliers more heavily compared to the mean absolute error (MAE). This is because RMSE is an L2 norm of the error vector measured between the predicted and the true values. A lower RMSE means the predictor is more accurate compared to a higher RMSE, as shown in the following equation:
R M S E = i = 1 N Y i Y ^ 2 N
where N represents the number of predictions, Y ^ represents the predicted angle, and Y i represents the measured angle at value i .
Cohen’s d was calculated to evaluate the effect size among pairwise comparisons of the sensor combinations and was analyzed for small (d = 0.2), medium (d = 0.5), and large (d  0.8) effects [39]. Bland–Altman plots with Pearson correlation coefficients were implemented to assess the strength of the linear relationship between the differences between the predicted and measured joint angles and the measured joint angle, as shown in the following equation:
r = ( x m x ) ( y m y ) x m x 2 y m y 2  
where x and y are two vectors of length N , and m x and m y are the means of x and y .
The plots were set with upper and lower bounds capturing a 95% level of agreement. To test our hypotheses, we performed a one-way analysis of variance (ANOVA) to compare the various sensor inputs for predicting the joint angles shown in Figure 3. Significant ANOVA tests prompted post hoc Tukey’s honest significant difference (HSD) tests. The Tukey–Kramer method was used to adjust p-values from multiple pairwise comparisons of inputs θ, 2IMU, 2IMU + θ, 4IMU, and 4IMU + θ. The significance threshold for statistical testing was set to αi = 0.05 but was adjusted following multiple comparisons to avoid Type I error. All statistical tests were performed using the Statsmodels module in Python programming language version 3.10.11.

3. Results

The RMSE values comparing the predictions of the Random Forest algorithm based on altering the sensor feature inputs are shown in Figure 4, where θa, θk, and θh represent the joint angles of the ankle, knee, and hip, respectively. We ran three independent ANOVAs (one each for the ankle, knee, and hip) and reported degrees of freedom (df), sum of squares (SS), and mean sum of squares (MS), with corresponding F-statistics and p-values. The ANOVA tests showed that the main effect of sensor combinations was statistically significant for algorithm prediction at the ankle (F4,279 = 11.07, p < 0.001), knee (F4,279 = 15.69, p < 0.01), and hip (F4,279 = 18.92, p < 0.001), prompting post hoc Tukey HSD pairwise comparisons (Table 1). Tukey HSD pairwise comparisons showed significant differences (p < 0.05) in RMSE values (Table 2, Table 3 and Table 4). Table 2, Table 3 and Table 4 show Tukey HSD pairwise comparisons for the ankle, knee, and hip to highlight how different sensor configurations led to more accurate predictions for specific joint angles. Pairwise comparisons resulted in the mean difference in RMSE values between sensor conditions, an adjusted p-value (p-adj) for each comparison, lower and upper bounds of the 95% confidence interval for the mean difference, and Cohen’s d, which shows the effect size as the difference between sensor conditions. The prediction of RMSE (mean ± one standard deviation) for one input (current joint angle, θ) resulted in the lowest predicted RMSE values of 1.92 ± 2.48°, 8.78 ± 2.96°, and 5.48 ± 1.47° for the ankle, knee, and hip, respectively (Figure 4). Two IMUs resulted in predicted RMSE values of 5.35 ± 4.28°, 20.29 ± 13.71°, and 14.86 ± 8.17° for the ankle, knee, and hip, which were not significantly different from the four IMU inputs to predict ankle (5.55 ± 4.10°, p < 0.99), knee (20.71 ± 12.41°, p < 0.99), and hip angles (13.55 ± 8.41°, p < 0.80; Table 2, Table 3 and Table 4).
The qualitative assessments of the Bland–Altman plots showed that as the measured angle of the joint increased, the difference between the predicted and measured angles also increased (Figure 5, Figure 6 and Figure 7).

4. Discussion

The purpose of this study was to evaluate various combinations of input signals that maximize machine learning prediction accuracy for simple movements while minimizing the required number of sensors. Our initial hypothesis that prediction accuracy would increase with the addition of IMUs attached to adjacent body segments was rejected. In fact, the prediction RMSE (mean ± one standard deviation) for one input (current joint angle, θ) resulted in the lowest predicted RMSE values of 1.92 ± 2.48°, 8.78 ± 2.96°, and 5.48 ± 1.47° for the ankle, knee, and hip, respectively (Figure 4). In essence, the addition of IMUs neighboring the joint of interest did not improve the joint angle predictions compared to joint angle inputs alone. In other words, starting with only joint angle inputs, incorporating two IMU sensors adjacent to the joint segment into the Random Forest predictor increased the RMSE at the ankle, knee, and hip by an average of 1.40°, 3.76°, and 5.19°, respectively. One plausible explanation for this discrepancy may come from the fact that adjacent IMUs alone (i.e., the two IMU groups) exhibited higher RMSE values compared to joint angle inputs alone (i.e., θa, θk, and θh groups, as shown in Figure 4). Therefore, adding adjacent IMUs to the joint angle inputs (θ vs. θ + two IMU groups) introduced new uncertainty in the estimates, resulting in a statistically significant worse accuracy at the ankle (RMSE of 1.92 ± 2.48° vs. 3.32 ± 2.59°), knee (RMSE of 8.78 ± 2.96° vs. 12.54 ± 6.28°), and hip (RMSE of 5.48 ± 1.47° vs. 9.67 ± 4.93°).
Our second hypothesis that the addition of IMUs attached to non-adjacent body segments would not increase prediction accuracy was confirmed because there was not a statistically significant difference between two IMUs vs. four IMUs to predict ankle (5.35 ± 4.28 vs. 5.55 ± 4.10°, p < 0.99), knee (20.29 ± 13.71 vs. 20.71 ± 12.41°, p < 0.99), and hip angles (14.86 ± 8.17° vs. 13.55 ± 8.41°, p < 0.80; Table 2, Table 3 and Table 4). Furthermore, adding non-adjacent segment IMUs to joint angle inputs (four IMUs + θ vs. two IMUs + θ) did not significantly reduce the RMSE values for the ankle, knee, and hip (Table 2, Table 3 and Table 4). This means that the increase in the number of IMU inputs to the Random Forest algorithm from 12 input signals (two IMU sensors each with 3D acceleration and 3D angular velocity per IMU sensor) to 24 input signals (four IMU sensors), or 13 input signals (two IMUs + θ) to 25 input signals (four IMUs + θ), did not reduce RMSE when predicting joint angles. This result suggests that using two IMUs on the adjacent segments of the joint is preferred over four IMUs on the torso, thigh, shank, and foot when predicting ankle, knee, and hip angles. These results align with the notion that classification performance significantly improves when sensors are positioned directly on the body region engaged in the specific movement of interest [40]. Since the IMUs were evaluated on segments exhibiting static and dynamic motion, there is a combination of uncertainty in the initial velocity from raw IMU signals. Furthermore, non-adjacent IMUs (e.g., torso) exhibiting quasi-static behavior may impose a greater amount of uncertainty for predicting dynamic motions (e.g., knee flexion) [41]. In optimal control and estimation theory, the quality of the estimate depends on the reliability and consistency of the sensor data [42]. In this study, uncertainties likely propagated through the estimation algorithm whenever new sources of noise and uncertainty from the IMUs were introduced, leading to increased estimation errors [41].
Perhaps the inclusion of four IMU sensors did not enhance the prediction accuracy compared to the two-IMU-sensor configuration because the two additional IMU sensors were not placed proximal and distal to the neighboring joint. This result is noteworthy, as it is more practical to setup two sensors as opposed to four sensors, and removing the number of attached sensors would likely be more convenient for the user while still maintaining a comparable prediction accuracy (Figure 4). The results also agreed with prior work showing how even a single IMU can produce a reasonable classification accuracy for certain manual material handling (MMH) tasks [30]. However, limiting the sensor configuration to a single IMU sensor should be used with caution, as accuracy can substantially decrease if the sensor momentarily drops the signal connection. Also, Porta et al. (2021) showed that prediction accuracy from a single sensor depends on certain MMH tasks, and additional sensors may be needed for tasks where multiple sensors can enhance the prediction accuracy (e.g., push/pull tasks) [30,31]. We also speculated that the addition of IMU sensors did not improve the prediction accuracy because the IMU sensors were not feature-engineered to predict future joint angles. The algorithm predictions, therefore, lagged behind the motion of the limb, as evident by the increased differences as a function of measured joint angles from the Bland–Altman plots (Figure 5, Figure 6 and Figure 7). The Bland–Altman plots showed positive Pearson correlations for 24 out of 30 scenarios, suggesting a general trend of increased differences as the measured joint angle increased (Figure 5, Figure 6 and Figure 7). With the exception of Figure 5f,g,i, and Figure 7b,d,f, the Pearson correlation coefficient revealed that the difference between the Random Forest prediction and the measured joint angle increased as the measured joint angle increased. The Bland–Altman plots also showed that these values were within the 95% limits of agreement, suggesting that the errors were consistently biased in the direction of the measured slope. Cohen’s d scores also showed large effect sizes (d ≥ 0.8) for a single joint angle input (θ) when compared to IMU inputs for 11 out of 12 comparisons. The only scenario where a large effect did not occur for a single joint angle input (θ) was at the knee (θk vs. two IMUs + θk), which resulted in a medium effect (d = −0.77).
The result of adding relevant types of input sensors to improve the prediction accuracy was in line with prior research [43,44]. For instance, Nurse et al. (2023) showed how the combination of pressure insole data with trunk motion increased low back disorder risk estimates from a range of r = 0.20–0.56 to r = 0.93–0.98 [43]. Substituting trunk IMU with thigh or pelvis IMU did not significantly reduce the prediction accuracy; however, removing more relevant sensors, such as force estimates from pressure insoles, significantly reduced low-back-loading estimation accuracy during MMH [45]. Smartphone camera data have also been shown to improve the prediction accuracy of knee adduction moment and knee flexion moment when combined with IMU sensors compared to models that used IMU or cameras alone [46]. Furthermore, optimizing the sensor configuration for HMIP for lower-limb joints aligns with previous research that validated the accurate prediction of ankle, knee, and hip angles [47]. This offers potential for the practical application of industrial exoskeletons during lifting activity [47]. However, the success of predictors not only depends on application but also on algorithm development and may still require sophisticated deep learning algorithms, such as LSTM, to obtain accurate estimates [30,48]. Nevertheless, this study points to implementing strategies for real-world applications where the tradeoff between the number of sensors and prediction accuracy is essential. Such applications include predicting ankle, knee, and hip mobility during rehabilitation, where attaching sensors can be time-consuming and inconvenient for users. Patients recovering from injury may require precise monitoring of joint angles across simple movements, and this study showed how a single sensor measuring joint angles can accurately predict ankle, knee, and hip angles, with predicted RMSE values of 1.92 ± 2.48°, 8.78 ± 2.96°, and 5.48 ± 1.47°. These RMSE values are a small percentage of the typical sagittal plane range of motion for the healthy adult population at the ankle (70–80°), (55–60°), knee (140–155°), and hip (130°) [49].
It is also worth noting that machine learning algorithms trained with a single feature input of joint angles resulted in statistically lower predicted RMSE value compared to algorithms with multiple IMUs along with joint angle inputs (Table 2, Table 3 and Table 4). It is possible that this result occurred because the IMU signals were not feature-engineered. Although the input signals were normalized, we did not perform other feature engineering techniques to extract relevant information. Additional feature extraction techniques that could have enhanced prediction accuracy using IMU sensor inputs include sensor fusion, strapdown integration of the gyroscope signal to acquire displacement, and statistical features across a moving time window, such as mean, minimum, maximum, standard deviation, and kurtosis [21]. Due to the complexities and limitations of feature engineering techniques, such as drift error, this study did not perform feature engineering techniques, apart from normalization. More complex deep learning models, such as artificial neural networks, autoencoders, and long short-term memory, are capable of extracting relevant features from raw data and may have an improved prediction accuracy with the addition of adjacent raw IMU inputs. However, it is important to consider factors such as model complexity, runtime, and hyperparameter tuning when implementing complex deep learning models.
This study consisted of some limitations. First, the recorded activities were simple movements performed in isolation. The study did not evaluate the algorithm during multi-joint movement tasks, nor was it tested during transitions between tasks. Activities of daily living are rarely performed in isolation, as human motion is a sequence of one action to another. Since the Random Forest algorithm was not trained with the transition from one task to another, it may incur prediction errors when predicting joint angles between tasks. As a result, the algorithm may be incapable of predicting multiple actions in a fluid-like manner. Therefore, future studies should include more complex actions and even transitions between actions in the training data to avoid overfitting to isolated tasks. Nonetheless, we believe this study serves as an initial baseline for future studies to effectively predict joint angles during complex movements. Another limitation was that this was an offline analysis. Future work should consider how runtime may alter prediction accuracy in an online scenario, where delays are critical for predicting user intent. Finally, this study was only performed on a healthy young adult population. As a result, additional data collection from diverse user populations, such as the elderly or amputees, is recommended for HMIP tailored to a specific population. Additionally, a mixture of training data from multiple populations could enhance the overall generalizability of the machine learning model, which can be used for activities that were not performed in this study, such as fall detection and running [50,51,52,53].
In order to offset the prediction bias, the Random Forest algorithm may benefit from additional sensor inputs, such as EMG signals. Since EMG signals occur approximately 120 ms prior to limb motion, this input information may help combat the lagging behavior of the Random Forest algorithm. As shown by Coker et al. (2020), including joint angles with EMG inputs resulted in RMSE 2.04° for predicting knee flexion angles 100 ms into the future [2]. Furthermore, Hollinger et. al. (2023) used EMG signals in addition to IMU and kinematic inputs for future joint angle prediction, which did not result in a prediction bias during level-ground walking [22]. However, the predictive algorithms by Hollinger et al. (2023) and Coker et al. (2020) were limited to level-ground walking. Therefore, EMG signals, feature engineering of IMUs, or a combination of both may help with joint angle prediction across diverse actions and may be worth exploring in future studies. Although this study focused on the growing topic of machine learning for HMIP, the applications of HMIP have a wide range of potential applications in monitoring rehabilitation, controlling a wearable exoskeleton, and identifying risk factors for work-related musculoskeletal disorders [44].

5. Conclusions

The results from this study indicated that machine learning prediction accuracy did not necessarily improve with additional IMU sensors. For instance, the addition of non-adjacent IMUs compared to two IMU sensors attached to adjacent segments did not increase the accuracy of predicting ankle, knee, and hip angles. This study also showed that using only kinematic inputs (joint angles) resulted in the best prediction accuracy of the Random Forest algorithm for the ankle, knee, and hip. This study, therefore, demonstrated how future joint angle prediction is largely determined by previous joint angles and is the most practical input signal for maximizing the joint angle prediction accuracy for multiple simple movements.

Author Contributions

Conceptualization, D.H. and M.Z.; methodology, D.H. and M.Z.; formal analysis, D.H.; investigation, D.H. and M.Z.; resources, M.Z.; data curation, D.H.; writing—original draft preparation, D.H.; writing—review and editing, M.C.S.J., H.C. and M.Z.; supervision, M.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by United States Army Combat Capabilities and Development Command, Grant Number: W15QKN-17-9-1025, PAO Number: PR2024-534. Approved for public release, distribution unlimited.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of Auburn University (#17-279 MR 1707, 06/08/2023).

Informed Consent Statement

Written informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon request.

Acknowledgments

The authors would like to thank Taylor Troutman for her assistance with data collection during this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human Activity Recognition Using Inertial, Physiological and Environmental Sensors: A Comprehensive Survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef] [PubMed]
  2. Coker, J.; Chen, H.; Schall, M.C.; Gallagher, S.; Zabala, M. EMG and Joint Angle-Based Machine Learning to Predict Future Joint Angles at the Knee. Sensors 2021, 21, 3622. [Google Scholar] [CrossRef] [PubMed]
  3. McGhan, C.; Nasir, A.; Atkins, E. Human Intent Prediction Using Markov Decision Processes. In Infotech@Aerospace 2012; American Institute of Aeronautics and Astronautics: Garden Grove, CA, USA, 2012. [Google Scholar] [CrossRef]
  4. Cervantes, C.; De Mesa, M.; Ramos, J.; Singer, S.; Del Carmen, D.J.; Cajote, R.D. Multi-Stage Hybrid-CNN Transformer Model for Human Intent-Prediction. In Proceedings of the TENCON 2023—2023 IEEE Region 10 Conference (TENCON), Chiang Mai, Thailand, 31 October 2023; IEEE: Chiang Mai, Thailand; pp. 1151–1156. [Google Scholar] [CrossRef]
  5. Bi, L.; Feleke, A.G.; Guan, C. A review on EMG-based motor intention prediction of continuous human upper limb motion for human-robot collaboration. Biomed. Signal Process. Control 2019, 51, 113–127. [Google Scholar] [CrossRef]
  6. Tang, G.; Wang, H.; Tian, Y. sEMG-Based Estimation of Knee Joint Angles and Motion Intention Recognition. In Proceedings of the 2017 9th International Conference on Intelligent Human-Machine Systems and Cybernetics (IHMSC), Hangzhou, China, 26–27 August 2017; Volume 2, pp. 390–393. [Google Scholar] [CrossRef]
  7. Liu, Y.; Liu, X.; Wang, Z.; Yang, X.; Wang, X. Improving performance of human action intent recognition: Analysis of gait recognition machine learning algorithms and optimal combination with inertial measurement units. Comput. Biol. Med. 2023, 163, 107192. [Google Scholar] [CrossRef] [PubMed]
  8. Hu, B.; Rouse, E.; Hargrove, L. Fusion of Bilateral Lower-Limb Neuromechanical Signals Improves Prediction of Locomotor Activities. Front. Robot. AI 2018, 5, 78. [Google Scholar] [CrossRef] [PubMed]
  9. Anwary, A.R.; Yu, H.; Vassallo, M. Optimal Foot Location for Placing Wearable IMU Sensors and Automatic Feature Extraction for Gait Analysis. IEEE Sens. J. 2018, 18, 2555–2567. [Google Scholar] [CrossRef]
  10. Gurchiek, R.D.; Cheney, N.; McGinnis, R.S. Estimating Biomechanical Time-Series with Wearable Sensors: A Systematic Review of Machine Learning Techniques. Sensors 2019, 19, 5227. [Google Scholar] [CrossRef] [PubMed]
  11. Rana, M.; Mittal, V. Wearable Sensors for Real-Time Kinematics Analysis in Sports: A Review. IEEE Sens. J. 2021, 21, 1187–1207. [Google Scholar] [CrossRef]
  12. Weizman, Y.; Tirosh, O.; Fuss, F.K.; Tan, A.M.; Rutz, E. Recent State of Wearable IMU Sensors Use in People Living with Spasticity: A Systematic Review. Sensors 2022, 22, 1791. [Google Scholar] [CrossRef]
  13. Prathivadi, Y.; Wu, J.; Bennett, T.R.; Jafari, R. Robust activity recognition using wearable IMU sensors. In Proceedings of the 2014 IEEE SENSORS, Valencia, Spain, 2–5 November 2014; pp. 486–489. [Google Scholar] [CrossRef]
  14. Exoskeletons Need to React Faster than Physiological Responses to Improve Standing Balance. Available online: https://www.science.org/doi/10.1126/scirobotics.adf1080 (accessed on 10 July 2023).
  15. Novak, D.; Riener, R. A survey of sensor fusion methods in wearable robotics. Robot. Auton. Syst. 2015, 73, 155–170. [Google Scholar] [CrossRef]
  16. Farrell, M.T.; Herr, H. A method to determine the optimal features for control of a powered lower-limb prostheses. In Proceedings of the 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011; pp. 6041–6046. [Google Scholar] [CrossRef]
  17. Kazemimoghadam, M.; Fey, N.P. Continuous Classification of Locomotion in Response to Task Complexity and Anticipatory State. Front. Bioeng. Biotechnol. 2021, 9. [Google Scholar] [CrossRef] [PubMed]
  18. Kazemimoghadam, M.; Fey, N.P. Biomechanical Signals of Varied Modality and Location Contribute Differently to Recognition of Transient Locomotion. Sensors 2020, 20, 5390. [Google Scholar] [CrossRef] [PubMed]
  19. Niswander, W.; Wang, W.; Kontson, K. Optimization of IMU Sensor Placement for the Measurement of Lower Limb Joint Kinematics. Sensors 2020, 20, 5993. [Google Scholar] [CrossRef] [PubMed]
  20. Dey, S.; Yoshida, T.; Foerster, R.H.; Ernst, M.; Schmalz, T.; Schilling, A.F. Continuous Prediction of Joint Angular Positions and Moments: A Potential Control Strategy for Active Knee-Ankle Prostheses. IEEE Trans. Med. Robot. Bionics 2020, 2, 347–355. [Google Scholar] [CrossRef]
  21. Molinaro, D.D.; Kang, I.; Camargo, J.; Young, A.J. Biological Hip Torque Estimation using a Robotic Hip Exoskeleton. In Proceedings of the 2020 8th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics (BioRob), New York, NY, USA, 29 November–1 December 2020; Volume 2020, pp. 791–796. [Google Scholar] [CrossRef]
  22. Hollinger, D.; Schall, M.; Chen, H.; Bass, S.; Zabala, M. The Influence of Gait Phase on Predicting Lower-Limb Joint Angles. IEEE Trans. Med. Robot. Bionics 2023, 5, 343–352. [Google Scholar] [CrossRef]
  23. Mundt, M.; Koeppe, A.; David, S.; Witter, T.; Bamer, F.; Potthast, W.; Markert, B. Estimation of Gait Mechanics Based on Simulated and Measured IMU Data Using an Artificial Neural Network. Front. Bioeng. Biotechnol. 2020, 8, 41. [Google Scholar] [CrossRef] [PubMed]
  24. Xiong, B.; Zeng, N.; Li, H.; Yang, Y.; Li, Y.; Huang, M.; Shi, W.; Du, M.; Zhang, Y. Intelligent Prediction of Human Lower Extremity Joint Moment: An Artificial Neural Network Approach. IEEE Access 2019, 7, 29973–29980. [Google Scholar] [CrossRef]
  25. Yun, Y.; Kim, H.-C.; Shin, S.Y.; Lee, J.; Deshpande, A.D.; Kim, C. Statistical method for prediction of gait kinematics with Gaussian process regression. J. Biomech. 2014, 47, 186–192. [Google Scholar] [CrossRef] [PubMed]
  26. Farmer, S.; Silver-Thorn, B.; Voglewede, P.; Beardsley, S.A. Within-socket myoelectric prediction of continuous ankle kinematics for control of a powered transtibial prosthesis. J. Neural Eng. 2014, 11, 056027. [Google Scholar] [CrossRef]
  27. Rokhmanova, N.; Kuchenbecker, K.J.; Shull, P.B.; Ferber, R.; Halilaj, E. Predicting knee adduction moment response to gait retraining with minimal clinical data. PLoS Comput. Biol. 2022, 18, e1009500. [Google Scholar] [CrossRef]
  28. Bao, L.; Intille, S.S. Activity Recognition from User-Annotated Acceleration Data. In Pervasive Computing; Ferscha, A., Mattern, F., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2004; Volume 3001, pp. 1–17. ISBN 978-3-540-21835-7. [Google Scholar]
  29. Dey, S.; Yoshida, T.; Foerster, R.H.; Ernst, M.; Schmalz, T.; Carnier, R.M.; Schilling, A.F. A hybrid approach for dynamically training a torque prediction model for devising a human-machine interface control strategy. arXiv 2021, arXiv:2110.03085. [Google Scholar] [CrossRef]
  30. Porta, M.; Kim, S.; Pau, M.; Nussbaum, M.A. Classifying diverse manual material handling tasks using a single wearable sensor. Appl. Ergon. 2021, 93, 103386. [Google Scholar] [CrossRef]
  31. Kim, K.; Cho, Y.K. Effective inertial sensor quantity and locations on a body for deep learning-based worker’s motion recognition. Autom. Constr. 2020, 113, 103126. [Google Scholar] [CrossRef]
  32. Zhao, H.; Qiu, Z.; Peng, D.; Wang, F.; Wang, Z.; Qiu, S.; Shi, X.; Chu, Q. Prediction of Joint Angles Based on Human Lower Limb Surface Electromyography. Sensors 2023, 23, 5404. [Google Scholar] [CrossRef] [PubMed]
  33. Hutter, F.; Xu, L.; Hoos, H.H.; Leyton-Brown, K. Algorithm runtime prediction: Methods & evaluation. Artif. Intell. 2014, 206, 79–111. [Google Scholar] [CrossRef]
  34. Laudanski, A.; Brouwer, B.; Li, Q. Measurement of lower limb joint kinematics using inertial sensors during stair ascent and descent in healthy older adults and stroke survivors. J. Healthc. Eng. 2013, 4, 555–576. [Google Scholar] [CrossRef] [PubMed]
  35. Ramachandran, A.K.; Pedley, J.S.; Moeskops, S.; Oliver, J.L.; Myer, G.D.; Lloyd, R.S. Changes in Lower Limb Biomechanics Across Various Stages of Maturation and Implications for ACL Injury Risk in Female Athletes: A Systematic Review. Sports Med. 2024. [Google Scholar] [CrossRef] [PubMed]
  36. Fuller, G.F. Falls in the Elderly. Am. Fam. Physician 2000, 61, 2159–2168. [Google Scholar] [PubMed]
  37. Chen, H.; Schall, M.C.; Martin, S.M.; Fethke, N.B. Drift-Free Joint Angle Calculation Using Inertial Measurement Units without Magnetometers: An Exploration of Sensor Fusion Methods for the Elbow and Wrist. Sensors 2023, 23, 7053. [Google Scholar] [CrossRef] [PubMed]
  38. Segal, M. Machine Learning Benchmarks and Random Forest Regression; Technical Report; Center for Bioinformatics & Molecular Biostatistics, University of California: San Francisco, CA, USA, 2003. [Google Scholar]
  39. Sullivan, G.M.; Feinn, R. Using Effect Size—Or Why the P Value Is Not Enough. J. Grad. Med. Educ. 2012, 4, 279–282. [Google Scholar] [CrossRef]
  40. Atallah, L.; Lo, B.; King, R.; Yang, G.-Z. Sensor Positioning for Activity Recognition Using Wearable Accelerometers. IEEE J. Mag. 2011, 5, 320–329. Available online: https://ieeexplore.ieee.org/document/5951802 (accessed on 17 April 2024). [CrossRef] [PubMed]
  41. Garreis, S.R. Optimal Control under Uncertainty: Theory and Numerical Solution with Low-Rank Tensors. Doctoral Dissertation, Technische Universität München, München, Germany, 2019. [Google Scholar]
  42. Hippe, P.; Deutscher, J. (Eds.) Optimal Control and Estimation. In Design of Observer-Based Compensators: From the Time to the Frequency Domain; Springer: London, UK, 2009; pp. 167–184. ISBN 978-1-84882-537-6. [Google Scholar] [CrossRef]
  43. Nurse, C.A.; Elstub, L.J.; Volgyesi, P.; Zelik, K.E. How Accurately Can Wearable Sensors Assess Low Back Disorder Risks during Material Handling? Exploring the Fundamental Capabilities and Limitations of Different Sensor Signals. Sensors 2023, 23, 2064. [Google Scholar] [CrossRef]
  44. Chan, V.C.H.; Ross, G.B.; Clouthier, A.L.; Fischer, S.L.; Graham, R.B. The role of machine learning in the primary prevention of work-related musculoskeletal disorders: A scoping review. Appl. Ergon. 2022, 98, 103574. [Google Scholar] [CrossRef] [PubMed]
  45. Matijevich, E.S.; Volgyesi, P.; Zelik, K.E. A Promising Wearable Solution for the Practical and Accurate Monitoring of Low Back Loading in Manual Material Handling. Sensors 2021, 21, 340. [Google Scholar] [CrossRef] [PubMed]
  46. Tan, T.; Wang, D.; Shull, P.B.; Halilaj, E. IMU and Smartphone Camera Fusion for Knee Adduction and Knee Flexion Moment Estimation During Walking. IEEE Trans. Ind. Inform. 2023, 19, 1445–1455. [Google Scholar] [CrossRef]
  47. Woo, J.-H.; Sahithi, K.K.; Kim, S.-T.; Choi, G.-R.; Kim, B.-S.; Shin, J.-G.; Kim, S.-H. Machine Learning Based Recognition of Elements in Lower-Limb Movement Sequence for Proactive Control of Exoskeletons to Assist Lifting. IEEE Access 2023, 11, 127107–127118. [Google Scholar] [CrossRef]
  48. Lim, S. Exposures to Several Risk Factors can be Estimated from a Continuous Stream of Inertial Sensor Measurements during a Variety of Lifting-Lowering Tasks. SSRN Electron. J. 2024, 1–16. [Google Scholar] [CrossRef]
  49. Roaas, A.; Andersson, G.B.J. Normal Range of Motion of the Hip, Knee and Ankle Joints in Male Subjects, 30–40 Years of Age. Acta Orthop. Scand. 1982, 53, 205–208. [Google Scholar] [CrossRef]
  50. Zabala, M.; Hollinger, D. Enhancing Model Generalizability via Transfer Learning for Multi-action Intent Recognition. In Proceedings of the 2023 American Society of Biomechanics Meeting, Knoxville, TN, USA, 8–11 August 2023. [Google Scholar]
  51. Donahue, S.R.; Hahn, M.E. Estimation of gait events and kinetic waveforms with wearable sensors and machine learning when running in an unconstrained environment. Sci. Rep. 2023, 13, 2339. [Google Scholar] [CrossRef]
  52. Aziz, O.; Musngi, M.; Park, E.J.; Mori, G.; Robinovitch, S.N. A comparison of accuracy of fall detection algorithms (threshold-based vs. machine learning) using waist-mounted tri-axial accelerometer signals from a comprehensive set of falls and non-fall trials. Med. Biol. Eng. Comput. 2017, 55, 45–55. [Google Scholar] [CrossRef]
  53. Ahmed, M.; Mehmood, N.; Nadeem, A.; Mehmood, A.; Rizwan, K. Fall Detection System for the Elderly Based on the Classification of Shimmer Sensor Prototype Data. Healthc. Inform. Res. 2017, 23, 147–158. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Experimental setup. (a) IMU placement of the lower limb and foot. (b) Frontal view of the placement of retroreflective markers and IMU sensors. (c) IMU sensor axes of flexion/extension, abduction/adduction, and internal/external rotation following the x-z-y coordinate system.
Figure 1. Experimental setup. (a) IMU placement of the lower limb and foot. (b) Frontal view of the placement of retroreflective markers and IMU sensors. (c) IMU sensor axes of flexion/extension, abduction/adduction, and internal/external rotation following the x-z-y coordinate system.
Sensors 24 03657 g001
Figure 2. Lab-based data analysis and algorithm overview. Trials of multiple actions (action-agnostic) were trained with sensory information input for a continuous Random Forest predictor. Joint angle predictions were compared to angles obtained from motion capture using inverse kinematics.
Figure 2. Lab-based data analysis and algorithm overview. Trials of multiple actions (action-agnostic) were trained with sensory information input for a continuous Random Forest predictor. Joint angle predictions were compared to angles obtained from motion capture using inverse kinematics.
Sensors 24 03657 g002
Figure 3. Input signal combinations for predicting ankle, knee, and hip angles. The blue rectangles on the person represent the evaluated IMU sensor configuration. Two IMUs were evaluated to the neighboring joint as a reduced sensor set. The two IMU sensors were located at the shank and foot to predict ankle angles, the thigh and the shank to predict knee angles, and the torso and thigh to predict hip angles. An identical set of four IMU sensors was evaluated at the foot, shank, thigh, and torso to predict ankle, knee, and hip angles.
Figure 3. Input signal combinations for predicting ankle, knee, and hip angles. The blue rectangles on the person represent the evaluated IMU sensor configuration. Two IMUs were evaluated to the neighboring joint as a reduced sensor set. The two IMU sensors were located at the shank and foot to predict ankle angles, the thigh and the shank to predict knee angles, and the torso and thigh to predict hip angles. An identical set of four IMU sensors was evaluated at the foot, shank, thigh, and torso to predict ankle, knee, and hip angles.
Sensors 24 03657 g003
Figure 4. Prediction of RMSE of joint angles 100 ms into the future, showing significant results denoted as * p < 0.05, ** p < 0.01, and *** p < 0.001.
Figure 4. Prediction of RMSE of joint angles 100 ms into the future, showing significant results denoted as * p < 0.05, ** p < 0.01, and *** p < 0.001.
Sensors 24 03657 g004
Figure 5. Bland–Altman plots corresponding to the difference between the predicted and measured ankle angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Figure 5. Bland–Altman plots corresponding to the difference between the predicted and measured ankle angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Sensors 24 03657 g005
Figure 6. Bland–Altman plots corresponding to the difference between the predicted and measured knee angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Figure 6. Bland–Altman plots corresponding to the difference between the predicted and measured knee angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Sensors 24 03657 g006
Figure 7. Bland–Altman plots corresponding to the difference between the predicted and measured hip angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Figure 7. Bland–Altman plots corresponding to the difference between the predicted and measured hip angles. Sensor combinations are displayed for ankle dorsiflexion (ae) and ankle plantarflexion (fj). The dotted red line represents the slope of the differences between the predicted and measured angles. The scatterplot colors are darkened for additional number sensor inputs and purple is shown for 4 IMUs. The 95% prediction limits and Pearson correlation coefficients assessing the proportional bias are displayed in the upper left corner of each plot. Error bars represent the standard error.
Sensors 24 03657 g007
Table 1. Analysis of variance results for the ankle, knee, and hip. A bold p-value indicates a statistically significant result (p < 0.05).
Table 1. Analysis of variance results for the ankle, knee, and hip. A bold p-value indicates a statistically significant result (p < 0.05).
Sourcedf SSMSFp
Ankle between groups43921.21980.30
Ankle within groups514069.7879.79
Ankle total2797991.00None11.07<0.001
Knee between groups457,979.0614,494.76
Knee within groups5138,835.67761.48
Knee total27996,814.73None15.69<0.001
Hip between groups428,201.667050.41
Hip within groups5117,439.06341.94
Hip total27945,640.7None18.92<0.001
Table 2. Post hoc pairwise Tukey’s honest significant difference—ankle condition.
Table 2. Post hoc pairwise Tukey’s honest significant difference—ankle condition.
Group 1Group 2 Mean Diff.p-adj.LowerUpperCohen’s d
2 IMU2 IMU + θa−2.03<0.01−3.80−0.27−0.57
2 IMU4 IMU0.200.99−1.571.960.05
2 IMU4 IMU + θa−1.690.070−3.460.070−0.46
2 IMUθa−3.43<0.001−5.20−1.67−0.98
2 IMU + θa4 IMU2.23<0.010.474.000.65
2 IMU + θa4 IMU + θa0.340.98−1.422.100.12
2 IMU + θaθa−1.400.19−3.160.36−0.55
4 IMU4 IMU + θa−1.890.029−3.66−0.13−0.53
4 IMUθa−3.63<0.01−5.40−1.87−1.07
4 IMU + θaθa−1.740.055−3.500.024−0.64
Table 3. Post hoc pairwise Tukey’s honest significant difference—knee condition.
Table 3. Post hoc pairwise Tukey’s honest significant difference—knee condition.
Group 1Group 2 Mean Diff.p-adj.LowerUpperCohen’s d
2 IMU2 IMU + θk−7.76<0.001−12.76−2.74−0.73
2 IMU4 IMU0.410.99−4.595.420.03
2 IMU4 IMU + θk−5.480.023−10.49−0.47−0.49
2 IMUθk−11.51<0.001−16.52−6.50−1.16
2 IMU + θk4 IMU8.17<0.0013.1613.180.83
2 IMU + θk4 IMU + θk2.270.72−2.737.280.31
2 IMU + θkθk−3.750.24−8.761.25−0.77
4 IMU4 IMU + θk−5.900.011−10.91−0.89−0.56
4 IMUθk−11.92<0.001−16.93−6.91−1.32
4 IMU + θkθk−6.02<0.01−11.03−1.01−0.98
Table 4. Post hoc pairwise Tukey’s honest significant difference—hip condition.
Table 4. Post hoc pairwise Tukey’s honest significant difference—hip condition.
Group 1Group 2 Mean Diff.p-adj.LowerUpperCohen’s d
2 IMU2 IMU + θh−5.19<0.001−8.47−1.90−0.77
2 IMU4 IMU−1.310.80−4.591.97−0.16
2 IMU4 IMU + θh−4.79<0.001−8.07−1.50−0.68
2 IMUθh−9.38<0.001−12.66−6.091.60
2 IMU + θh4 IMU3.870.0110.597.160.56
2 IMU + θh4 IMU + θh0.390.99−2.883.680.07
2 IMU + θhθh−4.19<0.01−7.47−0.901.15
4 IMU4 IMU + θh−3.470.031−6.76−0.19−0.48
4 IMUθh−8.07<0.001−11.35−4.781.34
4 IMU + θhθh−4.59<0.01−7.87−1.301.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hollinger, D.; Schall, M.C., Jr.; Chen, H.; Zabala, M. The Effect of Sensor Feature Inputs on Joint Angle Prediction across Simple Movements. Sensors 2024, 24, 3657. https://doi.org/10.3390/s24113657

AMA Style

Hollinger D, Schall MC Jr., Chen H, Zabala M. The Effect of Sensor Feature Inputs on Joint Angle Prediction across Simple Movements. Sensors. 2024; 24(11):3657. https://doi.org/10.3390/s24113657

Chicago/Turabian Style

Hollinger, David, Mark C. Schall, Jr., Howard Chen, and Michael Zabala. 2024. "The Effect of Sensor Feature Inputs on Joint Angle Prediction across Simple Movements" Sensors 24, no. 11: 3657. https://doi.org/10.3390/s24113657

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop