Next Article in Journal
The Feasibility of Modelling the Crown Profile of Larix olgensis Using Unmanned Aerial Vehicle Laser Scanning Data
Next Article in Special Issue
Fatigue Monitoring in Running Using Flexible Textile Wearable Sensors
Previous Article in Journal
Video-Based Stress Detection through Deep Learning
Previous Article in Special Issue
Development of a Smart Ball to Evaluate Locomotor Performance: Application in Adolescents with Intellectual Disabilities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning in Gait Parameter Prediction for OA and TKA Patients Wearing IMU Sensors

Center for Orthopaedic Biomechanics, University of Denver, Denver, CO 80208, USA
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(19), 5553; https://doi.org/10.3390/s20195553
Submission received: 1 September 2020 / Revised: 23 September 2020 / Accepted: 23 September 2020 / Published: 28 September 2020
(This article belongs to the Collection Sensors for Gait, Human Movement Analysis, and Health Monitoring)

Abstract

:
Quantitative assessments of patient movement quality in osteoarthritis (OA), specifically spatiotemporal gait parameters (STGPs), can provide in-depth insight into gait patterns, activity types, and changes in mobility after total knee arthroplasty (TKA). A study was conducted to benchmark the ability of multiple deep neural network (DNN) architectures to predict 12 STGPs from inertial measurement unit (IMU) data and to identify an optimal sensor combination, which has yet to be studied for OA and TKA subjects. DNNs were trained using movement data from 29 subjects, walking at slow, normal, and fast paces and evaluated with cross-fold validation over the subjects. Optimal sensor locations were determined by comparing prediction accuracy with 15 IMU configurations (pelvis, thigh, shank, and feet). Percent error across the 12 STGPs ranged from 2.1% (stride time) to 73.7% (toe-out angle) and overall was more accurate in temporal parameters than spatial parameters. The most and least accurate sensor combinations were feet-thighs and singular pelvis, respectively. DNNs showed promising results in predicting STGPs for OA and TKA subjects based on signals from IMU sensors and overcomes the dependency on sensor locations that can hinder the design of patient monitoring systems for clinical application.

1. Introduction

Quantitative assessments of movement quality in osteoarthritic (OA) and joint reconstruction patients, specifically spatial-temporal gait parameters (STGPs), provide valuable insight into gait patterns, activity type [1], risk of falling, and disease progression [2,3]. This diagnostic information is used in a number of applications that include development of personalized treatment plans, optimized post-operative rehabilitation, monitoring changes in mobility of patients after surgery [4,5,6,7], advancement of promising new interventions, and reducing overall medical costs [2]. Conventional methods for measuring gait characteristics that include motion capture (MOCAP) systems and force plates require a laboratory environment and expensive, time-consuming, equipment [8]. On the contrary, wearable sensors, specifically inertial measurement units (IMUs), are lightweight, inexpensive, and mobile. IMU’s measurement fidelity has improved significantly in recent years and have been used in various applications including 3D character animation, robotics, automotive vehicles, drones, and human motion measurement [9].
Processing streams of IMU data to extract clinically meaningful movement characteristics, such as activity classification, spatial-temporal parameters, gait pathology, and gait phase detection is challenging [10]. Several studies calculate spatial-temporal gait parameters by reconstruction of foot trajectories through double integration of the linear accelerations measured by IMUs. Sensor fusion techniques [11,12,13,14] and extended Kalman filters [15,16,17,18] are commonly used to reduce noise and improve measurement accuracy. These methods rely on identification of the zero-velocity condition of the foot during stance for gait segmentation. However, clear zero-velocity conditions are difficult to identify for patients with pathological gait or during highly dynamic activities like free running [19].
Data-driven approaches like deep learning have shown promising results in extracting complex patterns from data in the fields of computer vision, speech-recognition, and sequence modeling. Researchers have used deep learning on IMU-based movement data to classify different activities or quantify activity-specific movements [20,21,22,23]. Hannink et al. demonstrated the ability of deep-learning algorithms to recognize the non-linear relationships between raw IMU data and stride length as well as other STGPs [24]. Using a deep convolutional neural network trained on over 1220 strides from 101 geriatric patients, the algorithm predicted stride length with a mean error of −0.15 cm, which was considerably more accurate than previous integration-based methods [12]. Moreover, stride length predictions were robust to different methods of stride segmentation, improving the clinical applicability for patients with pathologic gait [25]. Similar results have been demonstrated using neural networks for measuring stride length during free running [17], but variability in foot strike patterns (e.g., heel strike versus toe strike) reduced accuracy highlighting the importance of population-specific datasets for best results.
Single body segment mounted IMUs (e.g., wrist or pelvis) are limited in calculation of certain STGPs such as number of steps, step cadence, or step distance which may not be adequate for clinical applications [26,27]. Incorporating IMUs on additional body segments (e.g., foot, shank, thigh, pelvis, or trunk) may provide access to additional gait metrics and improve gait characteristic predictions, especially for patient populations with pathologic movement characteristics [26,28,29,30]. Carcreff et al. demonstrated that IMUs placed on the shank and thigh yielded more accurate predictions of stride time, length, and velocity compared to feet mounted IMUs for children with cerebral palsy, particularly for those patients with increased disability [29]. Patients with progressive OA typically exhibit gait adaptations including decreased joint flexibility, increased stance time on the affected side, cadence, and double support time, and an overall increase in variability of spatial temporal parameters [31,32,33,34]. It is unclear how these gait adaptations progress over time and impact the prediction of gait mechanics using inertial sensors. Additionally, systematic studies quantifying optimal sensor combinations for the best performance across various patient populations are important to this field, but are lacking.
Thus, the purpose of this study was two-fold: (1) to access the ability of multiple contemporary deep neural network architectures to predict STGPs from IMU data in the OA and joint-replacement patient populations and (2) to determine the optimal sensor combination to maximize prediction accuracy. The results of this study will help patients suffering from OA who may go on to receive a total joint replacement benefit from the accurate real-time patient monitoring of STGPs to inform their treatment, surgical planning, and rehabilitation.

2. Materials and Methods

2.1. Gait Measurements of Osteoarthritic and Total Knee-Replacement Subjects

Twenty-nine subjects, including 14 subjects with OA (Age = 67 ± 7, weight = 79 ± 12 kg, height = 168 ± 16 cm, 4 females and 10 males), 15 subjects with total knee arthroplasty (TKA) (Age = 68 ± 4, weight = 76 ± 14 kg, height = 164 ± 9 cm, 11 females and 4 males, 7 uni-lateral and 8 bi-lateral), participated in the study as part of a larger investigation. All participants signed a consent form prior to the experiment with IRB approval (# 1328728). Subjects were fitted with 71 reflective markers on anatomical landmarks and 17 IMUs on various limb segments and the trunk. For this study, only the 7 IMUs located on the feet, shanks, thighs [35,36], and pelvis [37] were used in the subsequent data analysis (Figure 1a,b). Subjects performed 15 trials of a 5-m walking task at three different speeds: self-selected, slow, and fast to cover the entire range of possible daily walking paces. During fast walking, subjects were instructed to walk at their maximum comfortable speed without running (brisk walking) typified by longer steps at a faster cadence. During slow walking, subjects were instructed to walk at their slowest speed, typified by shorter steps at a slower cadence. During the walking tests, synchronized data was collected from a 13 camera Vicon motion capture system (Centennial, CO), 4 Bertec force platforms (Columbus, OH), and IMUs (Xsens, Enschede, Netherlands) (Figure 1a). The sampling frequency of force data, MOCAP, and IMUs (free acceleration and angular velocity) were 1000 Hz, 100 Hz, and 40 Hz, respectively.

2.2. Gait Data Processing

MOCAP data were segmented into a full stride for each leg based on two successive heel strikes identified using the heel markers’ vertical position [38]. For each full stride, the heel strike and toe off times, spatial characteristics (step length, stride length, step width, and toe out angle), temporal characteristics (step time, stride time, stance time, swing time, single support time, and double support time), and general characteristics (cadence and speed) were calculated [39,40].
IMU data for each trial was up-sampled to 100 Hz and segmented into full strides for each leg based on the angular velocities of the feet sensors in the sagittal plane using the peak detection method (Figure 1c,d) [41,42]. The mean absolute error between heel strike and toe-off events identified using the IMU and MOCAP data were 0.02 ± 0.01 and 0.04 ± 0.01 s, respectively. Linear accelerations and angular velocities from the IMU-based coordinate systems for left legs were reflected about the medio-lateral axis to provide consistent anatomical directions for left and right limb segments. The IMUs’ six channels of acceleration and angular velocity data were normalized using the maximum sensor acceleration and angular velocity range. A zero-padding technique was used to ensure the IMUs’ data sets had a consistent length of 212 points prior to use in the deep-learning models [24]. The IMU data for each stride segment was labeled with the gait characteristics calculated using the MOCAP data for use in the subsequent supervised machine learning models. The gait data processing yielded 3778 segmented and labeled strides from the 29 subjects. A descriptive statistical analysis was conducted on measured spatial, temporal, and general gait parameters to characterize the dataset. This includes mean, standard deviation, coefficient of variation, and interquartile range for knee OA and TKA subject cohorts at three paces, slow, normal, and fast.

2.3. Preliminary Neural Network Architecture Benchmarking and Selection

Six contemporary multivariate time series neural network architectures were utilized to predict stride length from our subject cohort based solely on the feet IMU data (Table 1). Stride length and the feet IMUs were chosen to enable benchmarking prediction accuracy against published studies. For network training, 80% of strides from 26 of 29 subjects were randomly allocated to the training set and the remaining 20% of strides from the same subjects were allocated to a validation set. Strides from the final three subjects not included in the training set, one OA subject, one uni-lateral TKA subject, and one bilateral TKA subject, were allocated to a test set. Network prediction accuracy was assessed using 5-fold cross-validation, with training, validation, and test sets randomly reallocated for each fold of the cross-validation. Optimal architecture with the lowest errors for both validation and test sets was selected for conducting a design of experiment on prediction of STGPs with different sensor numbers and locations.
Neural networks were trained using a backpropagation and stochastic gradient descent optimization approach to minimize the loss function, mean square error (MSE), between the model-predicted and labeled stride length, using the form:
MSE = i = 1 n y ^ i y i 2 n
where y ^ i was the model predicted stride length, yi was the labeled stride length, and n was the total number of strides in the training set.
An adaptive learning rate optimization with a learning rate, beta-1, and beta-2 of 0.001, 0.9, and 0.999, respectively were used for training all networks [43] with a total epoch of 300. Once each network was trained, the predictive accuracy was quantified by calculating the mean error (ME) and the mean absolute error (MAE) between the predicted and measured stride lengths for both the validation and test sets. ME was calculated to enable comparison with previously published studies, whereas the MAE provides a better metric for true prediction accuracy. To enable an equitable comparison of the prediction accuracy across various gait characteristics with different magnitudes and units, the absolute error was divided by the mean of the labeled test data resulting in the normalized absolute percent error (NAPE).

2.4. Assessing Optimal Sensor Combinations for Each Gait Characteristic

Based on the result of the preliminary neural network architecture selection, the 1D convolution neural network (CNN) architecture proposed by Zrenner et al. was chosen for a larger design-of-experiment study on sensor combinations [19]. This network consisted of two convolutional layers followed by two max pooling layers, a flattening layer, and two fully-connected layers. Rectified linear unit (ReLu) activation functions were placed after each layer. Keras with a Tensorflow backend was used for training the architecture [44,45].
A full factorial design of experiments was implemented to analyze the prediction accuracy based on 15 unique combinations of the feet, pelvis, shank, and thigh sensors (Table 2). Leveraging the ensemble approach proposed by Hannick et al., individual CNNs were trained using the segmented and labeled stride IMU data to predict each of the 12 spatial, temporal, and general gait parameters (Figure 2) for each unique sensor combination [24]. The same training set definitions, 5-fold cross-validation, and training approaches were used as in the preliminary analysis. Likewise, the same MAE and NAPE error estimations were calculated for each gait parameter with each sensor combination.
The Friedman test, which is a non-parametric statistical test analog to a repeated measures analysis of variance (ANOVA), was used to detect statistically significant differences in prediction accuracy (NAPE) across sensor combinations. Stepwise Dunn’s post hoc tests followed by Bonferroni correction due to multiple testing was performed to establish significant differences (new p-value: 0.05/105 = 0.000476). To determine an overall optimal sensor combination, sensor combinations were ranked based on Friedman ranking and averaged across all the gait parameters for each sensor combination [46].

3. Results

3.1. Spatial-Temporal Gait Parameters (STGPs) Statistical Analysis

The OA group demonstrated larger step width (+2.8 cm) and toe out angle (+4.9 deg), as well as smaller step length (−1.9 cm), stride length (−3.9 cm), double support time (−0.1 s), and speed (−3.4 cm/s) on average for three different paces compared to the TKA group. In general, OA patients demonstrated greater variation (standard deviation (SD), coefficient of variation (CV), and range) in all but two of the STGPs measured compared to TKA patients. Increases in variability (SD) was also observed for step length, stride length, cadence, and speed for fast trials in both OA and TKA groups compared to normal and slow paces (Table 3).

3.2. Benchmarking Neural Network Architecture

MAE for stride length ranged from 2.9 ± 2.6 cm to 6.9 ± 3.2 cm for the validation set and 7.6 ± 6.1 cm to 11.9 ± 7.1 cm for the test set (Table 4). The CNN architecture proposed by Zrenner et al. yielded the lowest MAE for both the validation and test data sets, and the lowest ME, NAPE, and ME standard deviation for the test set, indicating negligible bias and low variance in the stride length predictions. Additionally, this network architecture included only 148,529 parameters which was smaller than the other networks, reducing the computational cost of training the network and preventing overfitting.

3.3. Optimal Sensor Combinations for Gait Characteristics

Across sensor combinations, network predictions for spatial gait characteristics were most accurate (lowest NAPEs) for step length (7.6 ± 6% − 9.7 ± 6.9%) and stride length (7.1 ± 5.7% − 9.6 ± 7.9%), followed by step width (34.9 ± 27.2% − 40.9 ± 32.9%) and toe-out angle (73.7 ± 50.9% − 80.6 ± 53.9%) (Figure 3). For temporal parameters, most accurate predictions were for step time (3.1 ± 2.9% − 3.5 ± 3.7%), stride time (2.1 ± 2.3% − 2.6 ± 3%), stance time (3.5 ± 3.5% − 4.8 ± 4.2%), and swing time (4.6 ± 4.1% − 5.6 ± 4.8%). Prediction errors increased for single support time (5.2 ± 4.4% − 6.6 ± 5.3%), and double support time (22.6 ± 18.1% − 28 ± 23.1%). For general parameters, cadence was predicted with the highest accuracy (3.2 ± 3.7% − 4.1 ± 4.6%) followed by speed (6.4 ± 5.2% − 9.6 ± 8.8%).
Predictive accuracy was not equivalent between the OA and TKA cohorts, with generally larger prediction errors for the OA cohort (Figure 4). The OA cohort had larger mean (19.0%) and median (6.6%) NAPE across all sensor combinations and STGPs compared to TKA (mean NAPE = 14.7%, median NAPE = 4.6%). Fast walking also resulted in lower predictive accuracy relative to normal and slow walking. The mean and median NAPEs for fast walking were 17.7% and 6.22%, for normal walking were 15.8% and 4.8%, and for slow walking were 15.8% and 5.4% (Figure 5).
None of the sensor combinations consistently yielded the highest prediction accuracy for all variables. Sensor combinations were ranked based on NAPE for each gait parameter (Table 5). Overall, the feet-thigh (F T) configuration had the best average rank (5.1), followed by the feet-shank (F S, 6.2), and shank (S, 6.3) sensor combinations. The shank sensor combination consistently yielded the highest accuracy for temporal characteristics, ranking first or second for four of six temporal parameters. By contract, the pelvis (P) and pelvis-shank-thigh (P S T) sensor combinations consistently ranked among the least accurate with average ranks of 11 and 10.9, respectively.
The Friedman test indicated statistically significant differences (p = 0.001) between sensor combinations. Multiple pairwise comparisons based on Friedman ranking are displayed as homogenous subsets in Table 6. Similar to the mean NAPE ranking, feet-thigh and feet-shank sensor combinations ranked first and second in Friedman ranking. There was not a statistically significant difference between feet-thigh and feet-shank (adjusted p-value = 0.077) while there was a statistically significant difference between feet-thigh and the rest of the sensor combinations (adjusted p-value = 0.00). The pelvis sensor had the lowest accuracy with a significant difference compared to the other homogenous subsets of sensor combinations.

4. Discussion

The primary outcome of this study was the development of a robust deep-learning framework to predict diagnostic gait parameters for subjects with OA and TKA and investigate various sensor combinations on prediction accuracy. A simple ensemble deep neural network with two layers of 1D-CNNs demonstrated robust performance in predicting each STGP compared to more complex networks. A design of experiment conducted on 15 combinations of sensors and locations for different patient populations and gait paces revealed how the prediction accuracy of STGPs can change over different conditions and identification of an optimal sensor combination might be challenging. Overall, feet sensors combined with either shank or thigh sensors produced the highest accuracy for most STGPs and the isolated pelvis sensor showed the lowest accuracy.
The CNN architecture proposed by Zrenner et al. resulted in the lowest MAE and the lowest standard deviations for both the validation and test subject datasets with errors of 2.9 ± 2.6 cm and 7.6 ± 6.1 cm, respectively [19]. The CNN and ResNet model proposed by Hannik et al. and Wang et al. had the second and third lowest mean absolute errors of 8.2 ± 6.2 cm and 9.1 ± 6.4 cm [47]. Both Hannik et al. and Zrenner et al. published the mean and standard deviation of their models’ predictive error for stride length using unique datasets, enabling a direct comparison with our results [19,24]. Hannick et al. predicted stride length based on more than 1300 strides from 101 geriatric subjects, with a mean error of −0.15 ± 6.09 cm compared to our error of −2.2 ± 9.7 cm using the same network architecture. They used a larger number of subjects (n = 99) compared to our study (n = 29) which gives more unique data points for the network to train on. We induced additional variability in our dataset by asking subjects to walk at three different paces, all determined by the subjects. However, since this additional variability was mainly within-subject and may have a large amount of replication, it resulted in a slightly larger mean error compared to Hannink et al. In general, given large standard deviations in both studies, this difference was trivial. In this context, while our dataset had considerable variability, it likely had less variability than in the running dataset employed by Zrenner et al., which reported a mean predictive error in stride length of 2.5 ± 20.1 cm and a mean absolute error of 15.3 cm. The robustness of these CNN architectures for prediction of stride length point to the validity of using deep learning for this application, but also suggests that prediction accuracy is reduced when variability in the dataset is increased.
Direct comparisons of prediction accuracy between the current study and previous studies across all the STGPs are difficult due to differences in subject characteristics, dataset size, and experimental procedures. Comparable reported results from Hannick et al. for geriatric subjects, from Zrenner et al. for runners, and from Carcreff et al. for youths with cerebral palsy for sensor combinations and gait parameters are compiled in Table 7. Specifically, when comparing spatial parameter predictions using feet sensors, our results were within the range reported by previous studies. However, our results showed a larger mean error in prediction of stride length and step width compared to Hannink et al. that could be attributed to the larger number of subjects in Hannink et al. (n = 101) compared to our study. Diseases that induce pathologic movements, like OA, inherently increase the variability in gait parameters. The accuracy in prediction of the TKA group was higher than the OA group. The NAPE for OA was 19.0% and for TKA was 14.7%. When accounting for this limitation, our errors and standard deviations were comparable to previously reported results.
The neural networks trained on all sensor combinations predicted spatial, temporal, and general parameters with varying levels of accuracy. The NAPE averaged across all sensor combinations, for step length, stride length, step width, and toe-out angle were 8.6 ± 0.7, 7.8 ± 0.7, 38.5 ± 1.8, 77 ± 2%, respectively. The increased predictive error for step width and toe-out angle was likely associated with the smaller mean movements for those parameters, reducing the signal-to-noise ratio compared to the larger sagittal plane motions. For temporal parameters, the NAPE ranged from 2.3 ± 0.1% for stride time to 24.9 ± 1.5% for double support time. For the general parameters, the NAPE was 3.5 ± 0.2 and 7.5 ± 0.8% for cadence and speed, respectively. Descriptive statistical analysis on STGPs in our dataset revealed that neural network predictions were more accurate for the parameters with a lower coefficient of variation (CV). CV was defined as the ratio of the standard deviation to the mean which is an indicator of the dispersion of a probability distribution of data [48]. This was evident in the lower prediction accuracy observed for step width, toe-out angle, double support time, and speed with larger CVs compared to other parameters (Table 3).
The differences in predicted accuracy for OA versus TKA groups was multifactorial. First, there was more variability in the gait of OA subjects due to their pathology which makes it harder to predict certain STGPs. This higher variability for OA subjects was expressed by higher standard deviations and coefficients of variation for all gait parameters except toe-out angle, stride time, and cadence (Table 3). Second, the accuracy in prediction of STGPs was slightly higher at normal and slow walking compared to fast walking with mean NAPE of 15.8% for normal and slow, and 17.6% for fast walking. This aligns with findings by Zrenner et al. that indicated increasing speed could negatively impact predictive accuracy due to higher variability at fast walking (Table 3) [19]. Stressing the OA group with higher demand walking at a fast pace resulted in even greater variability and decreased predictive accuracy.
Perhaps most important, one of the randomly selected OA test subjects (subject S21, see Appendix A Figure A1 and Figure A2), walked with the shortest step length, shortest stride length, largest step width, and slowest speed among all subjects in the study, making this subject an outlier. Since our sample size was small, the impact of a single outlier was amplified and negatively affected prediction results. Jensen–Shannon divergence, which measures the similarity between two probability distributions, showed a larger divergence for subject S21 compared to the other two subjects in the same fold (S19 and S27). The divergences of step length, stride length, and step width for subject S21 were 5.67, 7.01, and 5.96 while for subject S19 and S27 the divergences were 0.18, 0.17, 0.10 and 0.38, 0.59, 0.67, respectively (see Appendix A Table A1). The divergence of S21 from the distribution of subjects used to train the CNNs resulted in poor performance, driving up the reported error for the OA cohort. Removing subject S21 from the test set reduced the mean and median NAPE from 19.0% and 6.6% to 17.3% and 4.8% which is comparable to the TKA group. Subject S21 had severe knee OA of the right knee which caused pain during activities of daily living and manifested in a noticeable limp on the affected limb compared to the other subjects in the OA cohort. Investigation of NAPE from the validation set revealed almost equal performance on both knee groups with mean and median NAPE of 7.7% and 2.9% for OA and 7.4% and 2.8% for TKA. The impact of subject S21 in the test set is an example of how CNNs result in poor performance when faced with data that are outside the distribution of the training data, which is one of the main challenges in the use of machine-learning models for real world applications. Hence, gaining intuition on training data set completeness is important prior to interpreting prediction accuracy. Out-of-distribution detection [49,50] has recently been recognized as an important area for developing trustworthy machine learning [49,50] and will be continually addressed in this work as patient numbers increase.
Our statistical analysis indicated statistically significant differences in accuracy between various sensor combinations tested across all conditions. The F-T combination was the highest-ranked sensor combination based on a Friedman test, showing a significant improvement in accuracy with respect to every other sensor combination except the F-S. It should be noted that although statistically significant, differences between the most and least accurate sensor combinations were small. The best sensor combination based on mean NAPE was F-T-S (15.25%) while the worst sensor combination was F-P (16.65%). Similarly, the Friedman test indicated the sensor combinations of F-T, F-S, and F-P-T were the top three ranked sensor combinations TKA subjects while T, F-S, and F-T were the top three ranked sensor combinations for OA subjects (see Appendix A Table A2a,b). The F-T and F-S were the common sensor combinations suitable for both OA and TKA groups. In addition, the F-T combination was also among the top three for slow, normal, and fast walking paces (see Appendix A Table A2c,e). As noted earlier, while the F-T sensor combination proved to be statistically better than other combinations, a 2–5% improvement in overall STGP prediction accuracy may be impactful during certain clinical applications. For instance, given the small difference in stride length at the normal pace between OA and TKA groups (~3 cm) higher accuracy predictions may be necessary for diagnostic purposes. However, higher accuracy may not be important for parameters with large differences between patient groups. This is an advantage of data driven modeling compared to other algorithm-based techniques in the prediction of STGPs. If the accuracy of STGPs is not largely impacted by sensor combination, there is freedom to design patient monitoring systems for specific patient groups based on other factors, such as cost and patient compliance. Feet sensors were necessary for stride segmentation during gait which is an input for the trained models. Therefore, including feet sensors is imperative for using a data-driven approach. Testing these sensor combinations on more complex tasks such as climbing stairs, sit-to-stand, and evaluating other joint kinematic and kinetic parameters would be necessary to clarify the value of using certain sensor combinations.
There are limitations to this study that should be considered. This study focused on gait to demonstrate the ability to predict STGPs from IMU data. In the OA population, other activities of daily living that place a greater demand on the subject will likely provide additional clinical value. The methods demonstrated in this study can be extended to predict analogous spatial temporal parameters for activities that include stair ascent/descent, sit-to-stand, and other high-demand activities. This study was also limited in the number of subjects that were included. This study demonstrated acceptable accuracy with 3778 segmented and labeled strides from the 29 subjects. Increasing the number of subjects and labeled strides will improve the predictive accuracy. Like other data-driven approaches, the trained network described in this study are only suitable for the selected population. There are also practical limitations to deploying our algorithm to a large patient population outside of a laboratory environment, including variability in sensor placement, reduced signal quality from low-cost IMUs, soft-tissue artifacts for high body mass index patients, and identification of patients with gait parameters outside the training data set. In order to implement this workflow for other populations with movement impairments that would benefit from patient monitoring, such as patients with cerebral palsy or stroke, the algorithm would need to be re-trained with inclusion of data from these populations. However, with this initial model architecture defined and the trained, a transfer-learning approach could be used on other populations to drastically reduce training time and the need for high volumes of data.

5. Conclusions

This study demonstrated that a deep-learning, data-driven approach was able to predict spatial temporal gait characteristics of OA and TKA patients based on signals from IMU sensors. Using a comprehensive analysis of various sensor combinations and their sensitivity to STGPs, patient population, and walking pace, our results showed that deep learning can overcome the dependency on sensor location that hinders the design of patient monitoring systems and negatively impacts patient compliance. Additionally, we demonstrated the importance sufficient variability in training and test data as a critical factor in the performance of DL models, especially for clinically relevant data with small sample sizes. A system that is able to leverage data streams from wearable sensors to produce real-time monitoring of STGPs in OA and TKA patients has the ability to improve clinical care and patient quality of life.

Author Contributions

Conceptualization, M.S.R., C.A.M. and C.W.C.; methodology, M.S.R. and R.Z.; software, M.S.R. and R.Z.; validation, M.S.R.; formal analysis, M.S.R. and R.Z.; investigation, C.A.M. and C.W.C.; resources, C.A.M., B.S.D., and C.W.C.; data curation, M.S.R.; writing—original draft preparation, M.S.R.; writing—review and editing, C.A.M., C.W.C., M.H.M., and B.S.D.; visualization, M.S.R.; supervision, C.W.C., M.H.M., and B.S.D.; project administration, C.W.C.; funding acquisition, C.W.C., C.A.M., and M.H.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Knoebel Institute for Healthy Aging (KIHA), University of Denver.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Boxplot associated with spatiotemporal gait parameters (STGPs) of each subject grouped by their knee status (OA and TKA) for training and test set.
Figure A1. Boxplot associated with spatiotemporal gait parameters (STGPs) of each subject grouped by their knee status (OA and TKA) for training and test set.
Sensors 20 05553 g0a1
Figure A2. Boxplot associated with NAPEs of each test subjects grouped by their knee status (OA and TKA) for test set.
Figure A2. Boxplot associated with NAPEs of each test subjects grouped by their knee status (OA and TKA) for test set.
Sensors 20 05553 g0a2
Table A1. Jensen–Shannon (JS) divergence (relative entropy) between test subjects and training on fold with subjects S19, S21, and S27.
Table A1. Jensen–Shannon (JS) divergence (relative entropy) between test subjects and training on fold with subjects S19, S21, and S27.
JS(Training||S19)JS(Training||S21)JS(Training||S27)
Step Length0.184.670.38
Stride Length0.177.010.59
Step Width0.105.960.67
Toe Out Angle0.110.020.63
Step Time0.010.260.80
Stride Time0.020.311.26
Stance Time0.010.140.63
Swing Time0.020.330.55
Single Support Time0.030.830.45
Double Support Time0.080.040.29
Cadence0.210.370.52
Speed0.070.500.29
Table A2. Homogeneous subsets based on Freidman ranking and asymptotic significances (0.05) for (a) OA cohort, (b) TKA cohort, (c) slow pace, (d) normal pace, and (e) fast pace.
Table A2. Homogeneous subsets based on Freidman ranking and asymptotic significances (0.05) for (a) OA cohort, (b) TKA cohort, (c) slow pace, (d) normal pace, and (e) fast pace.
(a)OASubsets
Sensors123456
T7.333
F S 7.620
F T 7.6957.695
S 7.7767.776
F 7.7797.779
F P 7.803
F P T 7.809
F S T 7.866
F P S 7.9857.985
F P S T 8.115
P T 8.144
S T 8.164
P S 8.445
P S T 8.590
P 8.876
Test Statistic.9.0499.0529.9276.536.
Adjusted Sig. (2-sided test).0.1030.3300.0700.077.
(b)TKASubset
Sensors1234567
F T7.573
F S7.7237.723
F P T7.8077.807
F P S T 7.8227.822
F S T 7.8547.854
S T 7.8727.872
S 7.8787.8787.878
F 7.8907.8907.890
F P S 8.0298.0298.029
P 8.1338.1338.133
F P 8.1448.144
P S T 8.1978.197
P S 8.2948.294
P T 8.297
T 8.488
Test Statistic5.6595.6958.9179.5408.5628.3799.928
Adjusted Sig. (2-sided test)0.1230.2580.3430.2090.1270.2180.120
(c)SlowSubset
Sensors12345
F T7.361
F P T 7.755
S 7.7747.774
F 7.9017.901
F S 7.9227.922
F P S T 7.9347.934
F P S 7.9397.939
F S T 7.9567.956
F P 8.0068.006
S T 8.0258.025
P 8.0748.074
P S 8.2948.294
P T 8.3048.304
T 8.352
P S T 8.403
Test Statistic.16.12412.9046.7914.860
Adjusted Sig. (2-sided test).0.0670.1840.1570.530
(d)NormalSubset
Sensors123456
F S7.726
F T7.7377.737
F S T7.7767.776
F P T7.7787.778
F 7.8987.898
F P S T 7.9007.900
S T 7.9577.957
S 7.9767.976
F P S 7.9947.994
F P 8.011
T 8.133
P T 8.143
P S 8.215
P S T 8.279
P 8.476
Test Statistic3.5529.3114.66410.5463.401.
Adjusted Sig. (2-sided test)0.7570.0920.6900.0930.782.
(e)FastSubset
Sensors12345
F T7.503
F S7.7127.712
F P S T7.7627.762
S7.7727.772
S T 7.834
F S T 7.846
F P T 7.9087.908
F P S 8.0118.0118.011
P S T 8.1278.1278.127
F P 8.1938.1938.193
P T 8.2058.205
P S 8.2158.215
F 8.2388.238
T 8.3328.332
P 8.342
Test Statistic8.99412.44810.67513.32810.304
Adjusted Sig. (2-sided test)0.1060.1100.0500.0800.226

References

  1. Witjes, S.; Gouttebarge, V.; Kuijer, P.P.; van Geenen, R.C.; Poolman, R.W.; Kerkhoffs, G.M. Return to Sports and Physical Activity After Total and Unicondylar Knee Arthroplasty: A Systematic Review and Meta-Analysis. Sports Med. 2016, 46, 269–292. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Weiss, A.; Brozgol, M.; Dorfman, M.; Herman, T.; Shema, S.; Giladi, N.; Hausdorff, J.M. Does the Evaluation of Gait Quality During Daily Life Provide Insight Into Fall Risk? A Novel Approach Using 3-Day Accelerometer Recordings. Neurorehabil. Neural Repair 2013, 27, 742–752. [Google Scholar] [CrossRef] [PubMed]
  3. Hausdorff, J.M. Gait variability: Methods, modeling and meaning. J. Neuroeng. Rehabil. 2005, 2, 19. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Levinger, P.; Menz, H.B.; Morrow, A.D.; Perrott, M.A.; Bartlett, J.R.; Feller, J.A.; Bergman, N.B. Knee biomechanics early after knee replacement surgery predict abnormal gait patterns 12 months postoperatively. J. Orthop. Res. 2012, 30, 371–376. [Google Scholar] [CrossRef] [PubMed]
  5. Lee, A.; Park, J.; Lee, S. Gait analysis of elderly women after total knee arthroplasty. J. Phys. Ther. Sci. 2015, 27, 591–595. [Google Scholar] [CrossRef]
  6. Snell, D.L.; Hipango, J.; Sinnott, K.A.; Dunn, J.A.; Rothwell, A.; Hsieh, C.J.; DeJong, G.; Hooper, G. Rehabilitation after total joint replacement: A scoping study. Disabil. Rehabil. 2018, 40, 1718–1731. [Google Scholar] [CrossRef]
  7. Fransen, M.; Nairn, L.; Bridgett, L.; Crosbie, J.; March, L.; Parker, D.; Crawford, R.; Harmer, A.R. Post-Acute Rehabilitation After Total Knee Replacement: A Multicenter Randomized Clinical Trial Comparing Long-Term Outcomes. Arthr. Care Res. 2017, 69, 192–200. [Google Scholar] [CrossRef] [Green Version]
  8. Hannink, J.; Ollenschlager, M.; Kluge, F.; Roth, N.; Klucken, J.; Eskofier, B.M. Benchmarking Foot Trajectory Estimation Methods for Mobile Gait Analysis. Sensors 2017, 17, 1940. [Google Scholar] [CrossRef]
  9. Imtiaz, U.; Yamamura, K.; Kong, W.; Sessa, S.; Lin, Z.; Bartolomeo, L.; Ishii, H.; Zecca, M.; Yamada, Y.; Takanishi, A. Application of wireless inertial measurement units and EMG sensors for studying deglutition—Preliminary results. Conf. Proc. IEEE Eng. Med. Biol. Soc. 2014, 2014, 5381–5384. [Google Scholar] [CrossRef] [Green Version]
  10. Nweke, H.F.; Teh, Y.W.; Al-Garadi, M.A.; Alo, U.R. Deep learning algorithms for human activity recognition using mobile and wearable sensor networks: State of the art and research challenges. Expert Syst. Appl. 2018, 105, 233–261. [Google Scholar] [CrossRef]
  11. Trojaniello, D.; Cereatti, A.; Pelosin, E.; Avanzino, L.; Mirelman, A.; Hausdorff, J.M.; Della Croce, U. Estimation of step-by-step spatio-temporal parameters of normal and impaired gait using shank-mounted magneto-inertial sensors: Application to elderly, hemiparetic, parkinsonian and choreic gait. J. Neuroeng. Rehabil. 2014, 11, 152. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Rampp, A.; Barth, J.; Schulein, S.; Gassmann, K.G.; Klucken, J.; Eskofier, B.M. Inertial Sensor-Based Stride Parameter Calculation From Gait Sequences in Geriatric Patients. IEEE Trans. Biomed. Eng. 2015, 62, 1089–1097. [Google Scholar] [CrossRef] [PubMed]
  13. Bertoli, M.; Cereatti, A.; Trojaniello, D.; Avanzino, L.; Pelosin, E.; Del Din, S.; Rochester, L.; Ginis, P.; Bekkers, E.M.J.; Mirelman, A.; et al. Estimation of spatio-temporal parameters of gait from magneto-inertial measurement units: Multicenter validation among Parkinson, mildly cognitively impaired and healthy older adults. Biomed. Eng. Online 2018, 17, 58. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Kluge, F.; Gassner, H.; Hannink, J.; Pasluosta, C.; Klucken, J.; Eskofier, B.M. Towards Mobile Gait Analysis: Concurrent Validity and Test-Retest Reliability of an Inertial Measurement System for the Assessment of Spatio-Temporal Gait Parameters. Sensors 2017, 17, 1522. [Google Scholar] [CrossRef] [PubMed]
  15. Foxlin, E. Pedestrian Tracking with shoe-mounted inertial sensors. IEEE Comput. Graph. Appl. 2005, 25, 38–46. [Google Scholar] [CrossRef]
  16. Bailey, G.P.; Harle, R. Assessment of Foot Kinematics During Steady State Running Using a Foot-Mounted IMU. Proc. Eng. 2014, 72, 32–37. [Google Scholar] [CrossRef] [Green Version]
  17. Zizzo, G.; Ren, L. Position Tracking During Human Walking Using an Integrated Wearable Sensing System. Sensors 2017, 17, 2866. [Google Scholar] [CrossRef] [Green Version]
  18. Bakhshi, S.; Mahoor, M.H.; Davidson, B.S. Development of a body joint angle measurement system using IMU sensors. In Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Boston, MA, USA, 30 August–3 September 2011. [Google Scholar]
  19. Zrenner, M.; Gradl, S.; Jensen, U.; Ullrich, M.; Eskofier, B.M. Comparison of Different Algorithms for Calculating Velocity and Stride Length in Running Using Inertial Measurement Units. Sensors 2018, 18, 4194. [Google Scholar] [CrossRef] [Green Version]
  20. Ordonez, F.J.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [Green Version]
  21. Kautz, T.; Groh, B.H.; Hannink, J.; Jensen, U.; Strubberg, H.; Eskofier, B.M. Activity recognition in beach volleyball using a Deep Convolutional Neural Network. Data Min. Knowl. Discov. 2017, 31, 1678–1705. [Google Scholar] [CrossRef]
  22. Hu, B.; Dixon, P.C.; Jacobs, J.V.; Dennerlein, J.T.; Schiffman, J.M. Machine learning algorithms based on signals from a single wearable inertial sensor can detect surface- and age-related differences in walking. J. Biomech. 2018, 71, 37–42. [Google Scholar] [CrossRef]
  23. Zheng, X.C.; Wang, M.Q.; Ordieres-Mere, J. Comparison of Data Preprocessing Approaches for Applying Deep Learning to Human Activity Recognition in the Context of Industry 4.0. Sensors 2018, 18, 2146. [Google Scholar] [CrossRef] [Green Version]
  24. Hannink, J.; Kautz, T.; Pasluosta, C.F.; Gassmann, K.G.; Klucken, J.; Eskofier, B.M. Sensor-Based Gait Parameter Extraction With Deep Convolutional Neural Networks. IEEE J. Biomed. Health Inf. 2017, 21, 85–93. [Google Scholar] [CrossRef] [Green Version]
  25. Hannink, J.; Kautz, T.; Pasluosta, C.F.; Barth, J.; Schulein, S.; Gassmann, K.G.; Klucken, J.; Eskofier, B.M. Mobile Stride Length Estimation With Deep Convolutional Neural Networks. IEEE J. Biomed. Health Inf. 2018, 22, 354–362. [Google Scholar] [CrossRef] [Green Version]
  26. Fasel, B.; Duc, C.; Dadashi, F.; Bardyn, F.; Savary, M.; Farine, P.A.; Aminian, K. A wrist sensor and algorithm to determine instantaneous walking cadence and speed in daily life walking. Med. Biol. Eng. Comput. 2017, 55, 1773–1785. [Google Scholar] [CrossRef]
  27. Soltani, A.; Dejnabadi, H.; Savary, M.; Aminian, K. Real-world gait speed estimation using wrist sensor: A personalized approach. IEEE J. Biomed. Health Inf. 2019. [Google Scholar] [CrossRef]
  28. Zijlstra, W.; Hof, A.L. Assessment of spatio-temporal gait parameters from trunk accelerations during human walking. Gait Posture 2003, 18, 1–10. [Google Scholar] [CrossRef] [Green Version]
  29. Carcreff, L.; Gerber, C.N.; Paraschiv-Ionescu, A.; De Coulon, G.; Newman, C.J.; Armand, S.; Aminian, K. What is the Best Configuration of Wearable Sensors to Measure Spatiotemporal Gait Parameters in Children with Cerebral Palsy? Sensors 2018, 18, 394. [Google Scholar] [CrossRef] [Green Version]
  30. Atallah, L.; Wiik, A.; Jones, G.G.; Lo, B.; Cobb, J.P.; Amis, A.; Yang, G.Z. Validation of an ear-worn sensor for gait monitoring using a force-plate instrumented treadmill. Gait Posture 2012. [Google Scholar] [CrossRef] [Green Version]
  31. Bejek, Z.; Paróczai, R.; Illyés, Á.; Kiss, R.M. The influence of walking speed on gait parameters in healthy people and in patients with osteoarthritis. Knee Surg. Sports Traumatol. Arthrosc. 2006. [Google Scholar] [CrossRef]
  32. Kiss, R.M.; Bejek, Z.; Szendrői, M. Variability of gait parameters in patients with total knee arthroplasty. Knee Surg. Sports Traumatol. Arthrosc. 2012. [Google Scholar] [CrossRef]
  33. Kiss, R.M. Effect of severity of knee osteoarthritis on the variability of gait parameters. J. Electromyogr. Kinesiol. 2011. [Google Scholar] [CrossRef]
  34. Hollman, J.H.; McDade, E.M.; Petersen, R.C. Normative spatiotemporal gait parameters in older adults. Gait Posture 2011. [Google Scholar] [CrossRef] [Green Version]
  35. Patterson, M.R.; Johnston, W.; O’Mahony, N.; O’Mahony, S.; Nolan, E.; Caulfield, B. Validation of temporal gait metrics from three IMU locations to the gold standard force plate. Conf. Proc. IEEE Eng. Med. Biol. Soc. 2016, 2016, 667–671. [Google Scholar] [CrossRef] [Green Version]
  36. Vargas-Valencia, L.S.; Elias, A.; Rocon, E.; Bastos-Filho, T.; Frizera, A. An IMU-to-Body Alignment Method Applied to Human Gait Analysis. Sensors 2016, 16, 2090. [Google Scholar] [CrossRef] [Green Version]
  37. Bolink, S.A.; van Laarhoven, S.N.; Lipperts, M.; Heyligers, I.C.; Grimm, B. Inertial sensor motion analysis of gait, sit-stand transfers and step-up transfers: Differentiating knee patients from healthy controls. Physiol. Meas. 2012, 33, 1947–1958. [Google Scholar] [CrossRef]
  38. Fellin, R.E.; Rose, W.C.; Royer, T.D.; Davis, I.S. Comparison of methods for kinematic identification of footstrike and toe-off during overground and treadmill running. J. Sci. Med. Sport 2010, 13, 646–650. [Google Scholar] [CrossRef] [Green Version]
  39. Panero, E.; Digo, E.; Agostini, V.; Gastaldi, L. Comparison of Different Motion Capture Setups for Gait Analysis: Validation of spatio-temporal parameters estimation. In Proceedings of the MeMeA 2018—2018 IEEE International Symposium on Medical Measurements and Applications, Rome, Italy, 11–13 June 2018. [Google Scholar]
  40. Veilleux, L.N.; Raison, M.; Rauch, F.; Robert, M.; Ballaz, L. Agreement of spatio-temporal gait parameters between a vertical ground reaction force decomposition algorithm and a motion capture system. Gait Posture 2016, 43, 257–264. [Google Scholar] [CrossRef]
  41. Haji Ghassemi, N.; Hannink, J.; Martindale, C.F.; Gassner, H.; Muller, M.; Klucken, J.; Eskofier, B.M. Segmentation of Gait Sequences in Sensor-Based Movement Analysis: A Comparison of Methods in Parkinson’s Disease. Sensors 2018, 18, 145. [Google Scholar] [CrossRef] [Green Version]
  42. Barth, J.; Oberndorfer, C.; Pasluosta, C.; Schülein, S.; Gassner, H.; Reinfelder, S.; Kugler, P.; Schuldhaus, D.; Winkler, J.; Klucken, J.; et al. Stride segmentation during free walk movements using multi-dimensional subsequence dynamic time warping on inertial sensor data. Sensors 2015, 15, 6419–6440. [Google Scholar] [CrossRef]
  43. Kingma, D.P.; Ba, J.L. Adam: A method for stochastic optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015—Conference Track Proceedings, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  44. Keras. 2015. Available online: https://keras.io (accessed on 22 August 2018).
  45. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467. [Google Scholar]
  46. Trawinski, B.; Smetek, M.; Telec, Z.; Lasota, T. Nonparametric statistical analysis for multiple comparison of machine learning regression algorithms. Int. J. Appl. Math. Comput. Sci. 2012. [Google Scholar] [CrossRef]
  47. Wang, Z.; Yan, W.; Oates, T. Time series classification from scratch with deep neural networks: A strong baseline. In Proceedings of the International Joint Conference on Neural Networks, Anchorage, AK, USA, 14–19 May 2017. [Google Scholar]
  48. Everitt, B.S.; Skrondal, A. The Cambridge Dictionary of Statistics. J. Am. Stat. Assoc. 1999. [Google Scholar] [CrossRef]
  49. Liang, S.; Li, Y.; Srikant, R. Enhancing the reliability of out-of-distribution image detection in neural networks. In Proceedings of the 6th International Conference on Learning Representations, ICLR 2018, Vancouver, BC, Canada, 30 April—3 May 2018. [Google Scholar]
  50. DeVries, T.; Taylor, G.W. Learning Confidence for Out-of-Distribution Detection in Neural Networks. arXiv 2018, arXiv:1802.04865. [Google Scholar]
Figure 1. Subject suited up with markers and inertial measurement units (IMUs) (a) front, and (b) back view. IMUs circled in blue (feet, shanks, thighs, pelvis) were used in the supervised machine learning models. (c) IMU sensor attached on right foot with coordinate system, (d) a sample of segmented IMU data based on angular velocities of feet sensors.
Figure 1. Subject suited up with markers and inertial measurement units (IMUs) (a) front, and (b) back view. IMUs circled in blue (feet, shanks, thighs, pelvis) were used in the supervised machine learning models. (c) IMU sensor attached on right foot with coordinate system, (d) a sample of segmented IMU data based on angular velocities of feet sensors.
Sensors 20 05553 g001
Figure 2. Workflow from in-vivo to spatial temporal gait parameter prediction.
Figure 2. Workflow from in-vivo to spatial temporal gait parameter prediction.
Sensors 20 05553 g002
Figure 3. Normalized absolute percentage error (NAPE) of neural network for spatial (blue), temporal (green), and general (orange) gait parameters with various sensors configurations in the test set. Whiskers indicate 25% and 75% quartiles. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Figure 3. Normalized absolute percentage error (NAPE) of neural network for spatial (blue), temporal (green), and general (orange) gait parameters with various sensors configurations in the test set. Whiskers indicate 25% and 75% quartiles. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Sensors 20 05553 g003aSensors 20 05553 g003b
Figure 4. Normalized absolute percentage error (NAPE) of neural network predictions for all gait parameters and various sensors configurations grouped by subject cohort (OA and TKA) in the test set. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Figure 4. Normalized absolute percentage error (NAPE) of neural network predictions for all gait parameters and various sensors configurations grouped by subject cohort (OA and TKA) in the test set. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Sensors 20 05553 g004aSensors 20 05553 g004b
Figure 5. Normalized absolute percentage error (NAPE) of neural network predictions for all gait parameters and various sensors configurations grouped by gait pace (slow, self-selected, fast) in the test set. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Figure 5. Normalized absolute percentage error (NAPE) of neural network predictions for all gait parameters and various sensors configurations grouped by gait pace (slow, self-selected, fast) in the test set. For each gait parameter, sensor configurations are listed in order of increasing NAPE.
Sensors 20 05553 g005
Table 1. Contemporary multivariate time-series deep-learning models for prediction of stride length.
Table 1. Contemporary multivariate time-series deep-learning models for prediction of stride length.
ReferenceModels
Hannink 2017 [23]Convolutional Neural Network (CNN)
Zrenner 2018 [17]Convolutional Neural Network (CNN)
Wang 2017 [35]Fully Convolutional Networks (FCN)
Wang 2017 [35]Residual Network (ResNet)
Karim 2019 [36]Multivariate Long Short-Term Memory Fully Convolutional Network (MLSTM-FCN)
Karim 2019 [36]Multivariate Attention Long Short-Term Memory Fully Convolutional Network (MALSTM-FCN)
Table 2. Sensor combinations used in the design of experiment.
Table 2. Sensor combinations used in the design of experiment.
nFeetPelvisShankThighCombinations
1× Feet (F)
2 × Pelvis (P)
3×× Feet Pelvis (F P)
4 × Shank (S)
5× × Feet Shank (F S)
6 ×× Pelvis Shank (P S)
7××× Feet Pelvis Shank (F P S)
8 ×Thigh (T)
9× ×Feet Thigh (F T)
10 × ×Pelvis Thigh (P T)
11×× ×Feet Pelvis Thigh (F P T)
12 ××Shank Thigh (S T)
13× ××Feet Shank Thigh (F S T)
14 ×××Pelvis Shank Thigh (P S T)
15××××Feet Pelvis Shank Thigh (F P S T)
Table 3. Descriptive statistic of spatial, temporal, and general parameters of dataset grouped by knee and pace (SD: Standard Deviation, CV: Coefficient of Variation, IQR: Interquartile Range).
Table 3. Descriptive statistic of spatial, temporal, and general parameters of dataset grouped by knee and pace (SD: Standard Deviation, CV: Coefficient of Variation, IQR: Interquartile Range).
VariableKnee StatusPaceMeanSDCVRangeIQR
Step Length (cm) Fast66.610.415.650.09.0
OANormal56.79.116.154.69.2
Slow53.26.913.045.09.8
Fast66.09.915.146.710.8
TKANormal59.17.612.952.310.7
Slow53.76.512.137.18.4
Stride Length (cm) Fast132.920.315.391.516.3
OANormal113.017.415.488.417.0
Slow106.112.912.167.718.2
Fast132.118.914.393.417.4
TKANormal118.114.312.171.419.8
Slow107.212.311.567.214.5
Step Width (cm) Fast13.25.742.825.59.1
OANormal12.96.550.350.39.0
Slow12.45.141.324.57.5
Fast10.04.342.723.25.5
TKANormal10.04.949.327.26.5
Slow10.24.038.622.65.2
Toe out Angle (deg) Fast23.915.765.871.821.3
OANormal24.617.872.386.827.7
Slow27.416.760.972.426.1
Fast18.914.677.462.023.5
TKANormal20.815.976.5105.023.4
Slow18.413.372.176.522.0
Step Time (s) Fast0.50.114.70.30.1
OANormal0.60.111.10.50.1
Slow0.70.113.60.60.1
Fast0.50.111.20.40.1
TKANormal0.60.19.90.40.1
Slow0.70.112.40.60.1
Stride Time (s) Fast0.90.113.70.60.2
OANormal1.10.110.00.70.2
Slow1.40.212.71.00.2
Fast1.00.110.20.60.1
TKANormal1.10.18.90.60.1
Slow1.40.211.71.00.2
Stance Time (s) Fast0.50.119.90.40.1
OANormal0.60.114.70.70.1
Slow0.80.116.70.80.2
Fast0.50.112.70.40.1
TKANormal0.60.111.20.50.1
Slow0.80.113.80.70.1
Swing Time (s) Fast0.40.111.70.30.1
OANormal0.50.111.40.40.1
Slow0.60.113.50.50.1
Fast0.40.09.80.30.1
TKANormal0.50.08.50.40.1
Slow0.60.111.90.50.1
Single Support Time (s) Fast0.40.113.60.30.1
OANormal0.50.112.20.40.1
Slow0.60.114.10.50.1
Fast0.40.09.70.20.1
TKANormal0.50.08.10.20.1
Slow0.60.111.80.50.1
Double Support Time (s) Fast0.00.1270.00.50.1
OANormal0.10.160.30.50.1
Slow0.20.152.00.60.1
Fast0.10.070.30.30.1
TKANormal0.10.033.50.30.1
Slow0.30.130.00.60.1
Cadence (1/s) Fast2.20.315.31.60.5
OANormal1.80.211.01.80.3
Slow1.50.214.31.30.2
Fast2.10.312.61.90.2
TKANormal1.80.29.41.00.2
Slow1.40.212.81.20.2
Speed (cm/s) Fast146.723.015.793.630.8
OANormal99.820.220.2114.821.5
Slow80.318.222.687.123.4
Fast139.726.118.7122.038.5
TKANormal105.816.115.381.625.1
Slow77.514.218.374.717.3
Table 4. Stride length prediction errors, mean error (ME), mean absolute error (MAE), and normalized absolute percentage error (NAPE) for multiple contemporary network architectures.
Table 4. Stride length prediction errors, mean error (ME), mean absolute error (MAE), and normalized absolute percentage error (NAPE) for multiple contemporary network architectures.
ModelsME ± STD (cm)
Validation Set
ME ± STD (cm)
Test Set
MAE ± STD (cm)
Validation Set
MAE ± STD (cm)
Test Set
NAPE ± STD (%)
Validation Set
NAPE ± STD (%)
Test Set
Number Of
Parameters
CNN [23]0.5 ± 4.2−2.2 ± 9.73.4 ± 2.98.2 ± 6.23 ± 2.57.2 ± 5.52,079,921
CNN [17]0.4 ± 3.7−2.4 ± 8.72.9 ± 2.67.6 ± 6.12.5 ± 2.26.8 ± 5.5148,529
FCN [35]−2.7 ± 3.9−4.8 ± 9.18.4 ± 3.511.9 ± 7.17.3 ± 310.5 ± 6.3277,121
ResNet [35]0.5 ± 3.9−1.9 ± 9.65.1 ± 3.29.1 ± 6.44.4 ± 2.88.1 ± 5.7229,953
MLSTM-FCN [36]1.0 ± 3.6−1.2 ± 9.46.1 ± 3.19.5 ± 6.85.3 ± 2.78.3 ± 6277,801
MALSTM-FCN [36]1.0 ± 3.7−0.8 ± 9.06.9 ± 3.210.3 ± 6.55.9 ± 2.79.1 ± 5.7278,361
Hannink et al. [23]NA−0.15± 6.1NANA 2,079,921
Zrenner et al. [17]NA2.5 ± 20.1NA15.3
Table 5. Sensor combinations ranking based on mean NAPE.
Table 5. Sensor combinations ranking based on mean NAPE.
FPF PSF SP SF P STF TP TF P TS TF S TP S TF P S T
Step Length814311415912571361012
Stride Length315213714125111410896
Step Width153117145131012826194
Toe Out Angle143154621110751281139
Step Time143126758291513111014
Stride Time274112561138159131410
Stance Time714114316132151251098
Swing Time715112984143135610121
Single Support Time101415219413711356128
Double Support Time111314134615812972105
Cadence515119286141013147123
Speed815697141314253121110
Average Spatial10.08.87.88.87.89.011.36.55.57.36.39.34.010.37.8
Average Temporal6.311.29.73.75.75.75.212.34.211.39.87.58.711.27.7
Average General7.813.07.47.66.610.810.55.34.55.76.35.29.611.08.8
Average7.611.08.86.36.27.77.99.65.19.37.57.47.310.97.5
Green is the highest rank. Red is the lowest rank.
Table 6. Homogeneous subsets based on Freidman ranking and asymptotic significances (0.05).
Table 6. Homogeneous subsets based on Freidman ranking and asymptotic significances (0.05).
Subset
1234567
F T7.637
F S7.7597.759
F P T 7.792
F S T 7.8187.818
F P S T 7.8877.887
S 7.9117.911
F 7.9467.946
S T 7.9527.952
F P S 7.9877.987
F P 8.036
P T 8.181
T 8.200
P S 8.229
P S T 8.280
P 8.386
Test Statistic6.5192.0183.2285.4179.6673.830.
Sig. (2-sided test)0.0110.3650.1990.2470.0220.280.
Adjusted Sig. (2-sided test)0.0770.8960.6710.5730.0790.709.
Table 7. Deep-learning accuracy comparison with previous studies for (a) spatial parameters, (b) general, and (c) temporal parameters.
Table 7. Deep-learning accuracy comparison with previous studies for (a) spatial parameters, (b) general, and (c) temporal parameters.
(a) Spatial
ME ± STD
Step Length
(cm)
Stride Length
(cm)
Step Width
(cm)
Toe-Out
Angle
(deg)
FeetOur Results−1.7 ± 5.2−3.0 ± 8.71.1 ± 5.1−3.2 ± 15.8
HanninkNA−0.15 ± 6.09−0.09 ± 4.22NA
ZrennerNA2.5 ± 20.1NANA
CarfcreffNA2.5 ± 3.7NANA
Shank ThighOur Results −0.6 ± 5.6 0.4 ± 9.7 0.85 ± 4.6 −3.7 ± 15.2
CarfcreffNA7.5 ± 6.9NANA
Average of all sensorsOur Test−0.5 ± 0.6−1.1 ± 1.30.6 ± 0.4−3.5 ± 0.8
(b) General
ME ± STD
Cadence
(1/s)
Cadence
(1/s)
FeetOur Results0.02 ± 0.10.02 ± 0.1
HanninkNANA
ZrennerNANA
CarfcreffNANA
Shank ThighOur Results0.01 ± 0.090.01 ± 0.09
CarfcreffNANA
Average of all sensorsOur Results0.04 ± 0.00 0.04 ± 0.00
(c) Temporal
ME ± STD
Stride Time
(s)
Stance Time
(s)
Swing Time
(s)
FeetOur Results–0.01 ± 0.04–0.01 ± 0.030.01 ± 0.03
Hannink–0.00 ± 0.07–0.00 ± 0.070.00 ± 0.05
Carfcreff0.00 ± 0.02NANA
Shank Thigh Our Results0.00 ± 0.03–0.01 ± 0.040.00 ± 0.03
Carfcreff0.00 ± 0.02NANA
Average of all sensorsOur Results0.00 ± 0.000.00 ± 0.000.00 ± 0.00

Share and Cite

MDPI and ACS Style

Sharifi Renani, M.; Myers, C.A.; Zandie, R.; Mahoor, M.H.; Davidson, B.S.; Clary, C.W. Deep Learning in Gait Parameter Prediction for OA and TKA Patients Wearing IMU Sensors. Sensors 2020, 20, 5553. https://doi.org/10.3390/s20195553

AMA Style

Sharifi Renani M, Myers CA, Zandie R, Mahoor MH, Davidson BS, Clary CW. Deep Learning in Gait Parameter Prediction for OA and TKA Patients Wearing IMU Sensors. Sensors. 2020; 20(19):5553. https://doi.org/10.3390/s20195553

Chicago/Turabian Style

Sharifi Renani, Mohsen, Casey A. Myers, Rohola Zandie, Mohammad H. Mahoor, Bradley S. Davidson, and Chadd W. Clary. 2020. "Deep Learning in Gait Parameter Prediction for OA and TKA Patients Wearing IMU Sensors" Sensors 20, no. 19: 5553. https://doi.org/10.3390/s20195553

APA Style

Sharifi Renani, M., Myers, C. A., Zandie, R., Mahoor, M. H., Davidson, B. S., & Clary, C. W. (2020). Deep Learning in Gait Parameter Prediction for OA and TKA Patients Wearing IMU Sensors. Sensors, 20(19), 5553. https://doi.org/10.3390/s20195553

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop