Next Article in Journal
Spatio-Temporal Correlation Between Radon Emissions and Seismic Activity: An Example Based on the Vrancea Region (Romania)
Previous Article in Journal
Compensation Method Based on Phase Shift Between Pins of Crystal Resonator
Previous Article in Special Issue
Validity and Reliability of Kinovea® for Pelvic Kinematic Measurement in Standing Position and in Sitting Position with 45° of Hip Flexion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Impact of Running Clothes on Accuracy of Smartphone-Based 2D Joint Kinematic Assessment During Treadmill Running Using OpenPifPaf

by
Nicolas Lambricht
1,
Alexandre Englebert
2,
Anh Phong Nguyen
1,
Paul Fisette
3,
Laurent Pitance
1,4 and
Christine Detrembleur
1,*
1
Institute of Experimental and Clinical Research, UCLouvain, 1200 Brussels, Belgium
2
Institute of Information and Communication Technologies, Electronic and Applied Mathematics, UCLouvain, 1348 Louvain-la-Neuve, Belgium
3
Institute of Mechanics, Materials and Civil Engineering, UCLouvain, 1348 Louvain-la-Neuve, Belgium
4
Service de Stomatologie et de Chirurgie Maxillo-Faciale, Cliniques Universitaires Saint-Luc, 1200 Brussels, Belgium
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(3), 934; https://doi.org/10.3390/s25030934
Submission received: 23 December 2024 / Revised: 22 January 2025 / Accepted: 30 January 2025 / Published: 4 February 2025
(This article belongs to the Special Issue Wearable Systems for Monitoring Joint Kinematics)

Abstract

:
The assessment of running kinematics is essential for injury prevention and rehabilitation, including anterior cruciate ligament sprains. Recent advances in computer vision have enabled the development of tools for quantifying kinematics in research and clinical settings. This study evaluated the accuracy of an OpenPifPaf-based markerless method for assessing sagittal plane kinematics of the ankle, knee, and hip during treadmill running using smartphone video footage and examined the impact of clothing on the results. Thirty healthy participants ran at 2.5 and 3.6 m/s under two conditions: (1) wearing minimal clothing with markers to record kinematics by using both a smartphone and a marker-based system, and (2) wearing usual running clothes and recording kinematics by only using a smartphone. Joint angles, averaged over 20 cycles, were analysed using SPM1D and RMSE. The markerless method produced kinematic waveforms closely matching the marker-based results, with RMSEs of 5.6° (hip), 3.5° (ankle), and 2.9° (knee), despite some significant differences identified by SPM1D. Clothing had minimal impact, with RMSEs under 2.8° for all joints. These findings highlight the potential of the OpenPifPaf-based markerless method as an accessible, simple, and reliable tool for assessing running kinematics, even in natural attire, for research and clinical applications.

1. Introduction

Running is a popular sporting activity, primarily due to its accessibility and the associated health and longevity benefits [1]. Despite these benefits, the incidence of overuse injuries to the lower limbs remains high among novice, recreational, and competitive runners [2]. Indeed, this activity involves the high repetition of rapid unipodal movements, exposing participants to significant external forces which can be several times their body weight [3]. Running technique is considered a risk factor for running injuries [4], making the evaluation of kinematics valuable not only during training but also in the rehabilitation of injured runners and other athletes.
In the particular case of anterior cruciate ligament (ACL) rupture, a common injury in young and active individuals [5], return to running is an important milestone in rehabilitation [6,7]. A systematic review has highlighted significant differences in knee joint kinematics between ACL-injured patients and patients with uninjured legs and in control groups during running [8]. The authors noted that alterations particularly affect the sagittal plane, resulting in altered knee kinematics, and persisted for up to five years post-reconstruction. In addition, altered movement patterns have been implicated in the early onset of knee osteoarthritis [9] and in increased risks of re-injury [10]. Therefore, a comprehensive assessment of knee kinematics and a focus on restoring running biomechanics early in rehabilitation may be beneficial for patients [11].
Kinematic parameters are usually assessed using an optoelectronic system consisting of multiple infra-red cameras and reflective markers placed on the subjects. However, the high cost, need for dedicated space and skilled operators, and lack of portability make these systems difficult to use outside research settings [12]. To address these limitations, various markerless technologies have been investigated. Inertial measurement units (IMUs) [13], depth cameras (e.g., the Kinect) [14], and single or networked RGB cameras have been extensively studied [15,16]. For the third approach, the development of deep learning techniques for human pose detection has enabled markerless computation of 2D and 3D motion kinematics from single or multiple 2D video sources [17,18]. In particular, the kinematics in the sagittal plane derived from various pose estimation methods have been validated against traditional 2D tools, such as Kinovea, as well as reference 3D motion analysis systems during walking, running, and tasks like squatting or jumping [19,20,21]. When analysing running kinematics in the sagittal plane, two approaches can be employed: measuring 3D kinematics and extracting sagittal plane data, or directly measuring only the sagittal plane in 2D. For the first approach, multiple cameras are required, which must be synchronized and calibrated. Traditionally, the 2D coordinates of anatomical key points are extracted from each video using neural networks, such as OpenPose [22]. Subsequently, 3D coordinates are calculated by triangulating the measurements from the different cameras, and kinematics are derived using inverse kinematics [15,23]. The second approach, while less robust and versatile, is more accessible and easier to implement in clinical practice, as it requires only a single camera, such as a smartphone camera. Researchers have evaluated the performance of OpenPose-based systems for kinematics evaluation [20], as well as other solutions requiring lower computing costs, such as BlazePose [16]. OpenPifPaf [24,25] is a convolutional neural network designed for human pose detection that shows good performance, outperforming OpenPose in challenging conditions, such as low-resolution and crowded, cluttered, or occluded scenes [26]. A recent study compared sagittal plane kinematics during jumps measured based on OpenPifPaf to those obtained using a marker-based system [27]. The results highlighted its potential clinical value due to its performance and ease of use. However, its application for running analysis has not yet been explored. Unlike marker-based systems, markerless motion capture does not require attaching markers to body segments. Instead, it relies on automatic landmark estimation, which can be influenced by factors such as clothing [28]. Recent studies indicate that clothing has minimal impact on 3D markerless walking and running kinematics measured with multiple cameras [29,30,31]. Investigating whether this holds for single-camera 2D sagittal plane studies would be valuable.
The primary aim of this study was to evaluate the performance of an OpenPifPaf-based markerless method for measuring running kinematics. This involved comparing its results for hip, knee, and ankle joints in the sagittal plane with those obtained using a 2D marker-based approach. The secondary aim was to examine the effect of subjects wearing their usual running clothes on the markerless method results.

2. Materials and Methods

2.1. Participants

The participants were 30 healthy men and women (15/15; age: 23.1 ± 4.0 years old; height: 1.77 ± 0.1 m; mass: 66.6 ± 9.1 kg) who had not suffered any lower extremity musculoskeletal injuries for at least 6 months and were able to run at 3.6 ms−1 for at least 3 min. Most of the participants were young, Caucasian, and in good physical condition.
This study was approved by the ethics committee of our university (B403201523492). The purpose of the study was explained to the participants and they all provided their written informed consent to the use of their anonymized data before the experiments.

2.2. Procedure

Marker-based motion capture was conducted at 100 Hz using a height-camera optoelectronic system (Vicon V5 Motion Systems, Oxford Metrics Ltd., Oxford, UK) positioned around the treadmill. For markerless analysis, a smartphone (iPhone12; frequency: 60 Hz; resolution: 1280 × 720 pixels) was mounted on a tripod with a fixed height of 1 m and placed 3 m laterally perpendicular to the movement plane (Figure 1A).
During the experiment, movement was recorded for 30 s as participants ran at two fixed speeds, 2.5 m.s−1 and 3.6 m.s−1, under two clothing conditions. The first condition (‘Lab clothing’) required participants to wear close-fitting underwear, their running shoes, and eighteen infra-red reflective markers. Sixteen markers were used to comply with the Nexus ‘Plug-in Gait lower body’ model, with an additional marker placed on the middle of the lateral edge of each acromion. In the second condition (‘Running clothes’), participants wore their usual running clothes without any specific guidance, typically shorts or leggings for the bottoms and loose-fitting t-shirts for the tops.
Participants were given a 5-min familiarization period to adjust to treadmill running and select a comfortable speed. During this time, they were instructed to test and briefly run at the two experimental speeds to ensure familiarity with these conditions. After the familiarization period, participants were equipped with the reflective markers and ran on the treadmill at the lower speed. After 60 s, they were asked whether they felt comfortable at this speed. Once participants indicated they were comfortable, data collection was initiated, capturing 30 s of motion. The same process was then repeated for the higher speed after a 2-min pause between the two speeds. Between the two clothing conditions, a longer pause was required to allow for the removal of motion capture markers and for participants to dress. Although the exact duration of this pause was not recorded, care was taken to ensure participants were ready before resuming the protocol. In both clothing conditions, participants always started with the lower speed.
Under the Lab clothing condition, a smartphone video was recorded simultaneously with the static calibration of the optoelectronic system prior to capturing running motion. Motion under this condition was captured using both the optoelectronic system and the smartphone. To synchronize the recordings and allow for the selection of the same running cycles during analysis, participants performed a ‘butt-kick’ movement approximately one second after the start of motion capture recording. Under the Running clothes condition, movements were recorded exclusively via smartphone video, and no ‘butt-kick’ trigger movement was required.

2.3. Data Processing

Spyder scientific environment (Python 3.12) was used for data processing. Since the smartphone was positioned to the left of the participants, only the joint kinematics on that side were assessed. For the marker-based reference system, kinematics were not computed using the Plug-in-Gait model. Instead, the optoelectronic system was used to extract accurate spatial coordinates of the markers placed on anatomical landmarks, similar to the key points identified by OpenPifPaf, for joint kinematics computation. The raw 3D marker trajectory data from the optoelectronic system were exported and filtered using a low-pass, zero-lag, fourth-order Butterworth filter with a cut-off frequency of 6 Hz. The approximate position of the left hip joint was estimated using the Newington–Gage model [32]. This estimation involved utilizing data from the four superior iliac spine markers as well as the participant’s leg length. The 2D coordinates of the four markers on the participant’s left side, along with the hip position estimated by the Newington–Gage model (represented by triangles and a star in Figure 1), were used for kinematics computation. For the markerless pose detection, we used the WholeBody extension of OpenPifPaf with the shufflenetv2k30-wholebody checkpoint. The video was processed by the neural network after rescaling by a factor 0.5 to 640 × 360 pixels. This allowed us to obtain the coordinates of five key points of interest for joint kinematics computation (represented by black dots in Figure 1). Key point trajectory gaps were filled using cubic spline interpolation, and the data were filtered using the same low-pass Butterworth filter as applied to the optoelectronic system data. Joint kinematics were calculated similarly for both methods using trigonometry and the dot product. For the ankle, the foot was considered a rigid body, and joint angles were corrected by accounting for the average angles measured in the neutral position during static calibration. The markerless method used key points of the knee, ankle, and big toe, while the marker-based method used markers on the lateral epicondyle, lateral malleolus, and second metatarsal. For the knee, the markerless method used key points of the hip, knee, and ankle, while the marker-based method used markers on the lateral malleolus and lateral epicondyle, and the estimated hip position. For the hip, the markerless method used key points of the shoulder, hip, and knee, while the marker-based method used markers on the acromion and lateral epicondyle, and the estimated hip location.
Figure 2 provides an overview of the procedure used to obtain the average joint kinematic curves over 20 cycles for the ankle, knee, and hip of a participant at a given speed and under a specific condition. These average curves were calculated for each of the 30 participants at the two speeds and across the three conditions: marker-based, markerless in lab clothing, and markerless in running clothes. For the Lab clothing condition, to ensure consistency in comparing the results obtained with the marker-based system and without markers, we utilized the ‘butt-kick’ movement as a trigger to select corresponding cycles. Indeed, during this trigger movement, the left ankle reached an abnormally high vertical position compared to the rest of the recording (see Trigger in Figure 2A). Then, we defined the start and end indexes of each running cycle based on the horizontal position of the left ankle. Given the treadmill setting, we inferred that the foot’s ground contact aligned with the ankle reaching its maximum anterior position. For subsequent analysis, we selected the indexes of the 20 cycles immediately following the second cycle after the trigger movement (see Figure 2A). The 2D joint kinematics for the ankle, knee, and hip joints based on both the markers and key point coordinates were computed (see Figure 2B). Then, the joint angles were segmented into running cycles using the previously determined indexes. These 20 cycles were resampled using linear interpolation to represent joint angles as a percentage of the cycle. Finally, the average angle for each joint was calculated at each percentage point across the 20 cycles (see Figure 2C). These averaged angles were subsequently used for further analysis.
A similar procedure was used to obtain average angle curves when running with usual clothes. Since the movement was recorded using only a smartphone, no trigger movement was required, and the twenty running cycles following the second cycle were used for analysis.

2.4. Statistical Analysis

To evaluate the performance of the markerless method, the averaged angle curves for the ankle, knee, and hip joints obtained without markers were compared to those obtained using markers at both running speeds. To assess how the differences between the angles measured by the two systems evolved throughout the running cycle, the absolute error was calculated at each percentage of the cycle for each participant and averaged across all 30 participants.
The Root Mean Square Error (RMSE) was calculated for the three joints for each participant. This metric provides an assessment of the overall accuracy of the markerless system in estimating lower limb kinematics and facilitates comparison with recent studies.
To analyse differences in joint kinematics between the two systems, time-series data for all participants were subjected to a one-dimensional Statistical Parametric Mapping two-way repeated-measures ANOVA using the SPM1D package (v0.4.22) [33]. The normality of the data was assessed prior to the analysis, and the alpha level of 0.05 was adjusted using Sidak correction to account for multiple comparisons, resulting in an effective alpha level of 0.017 for the three joints. The analysis included two factors: the system used (marker-based versus markerless) and the running speed. Although the ANOVA allows for examining three effects—the main effect of the system, the main effect of speed, and the interaction between system and speed—we focused exclusively on the main effect of the system and the interaction between system and speed, as these were most relevant to our research objectives.
To investigate the potential effect of clothing, the same methods were applied to compare joint kinematics obtained with the markerless system under Lab clothing and Running clothes conditions.

3. Results

Figure 3 presents the comparison between the markerless system results and those of the marker-based reference system throughout the running cycle. It includes the joint kinematic curves for the ankle, knee, and hip joints at 2.5 m.s−1 and 3.6 m.s−1, the results of the SPM1D analyses for system comparison and speed influence, and the mean absolute differences between the two systems. The SPM1D analysis revealed significant differences between the marker-based and markerless systems, with minimal influence of speed on these differences. The biggest discrepancy was observed at the hip joint, particularly between 20% and 60% of the cycle. Although we did not explicitly delimit the stance and swing phases during running in our study, this region approximately corresponds to the transition from the end of the stance phase to the beginning of the swing phase. Indeed, maximum hip extension typically occurs at the moment of toe-off during running [34]. During this period, the markerless system reported up to 9° greater hip extension compared to the marker-based system. Another notable difference occurred at the knee joint during the final 10% of the cycle, just before foot–ground contact, where the markerless system reported approximately 8° greater knee flexion.
Figure 4 compares the markerless kinematic results across the two clothing conditions. The joint kinematic curves appear very similar, and the SPM1D analysis comparing Lab clothing and Running clothes conditions reveals no significant differences. Additionally, the effect of speed on these results remains minimal. For all three joints, the maximum mean absolute error does not exceed 4°.
Figure 5 illustrates the distributions of the RMSEs for the three joints across the 30 participants at 2.5 m.s−1 and 3.6 m.s−1 for both systems. When comparing marker-based and markerless kinematics, the mean RMSE values range from 2.9° to 5.6°, depending on the joint and the speed. In contrast, the comparison of kinematic curves obtained by the markerless system under different clothing conditions yields mean RMSE values between 2.1° and 2.8°. Detailed RMSE values for all joints and conditions are provided in Table 1.

4. Discussion

In this study, we first compared the joint kinematics of the ankle, knee, and hip during running at 2.5 m.s−1 and 3.6 m.s−1, obtained using a 2D marker-based approach and an OpenPifPaf-based markerless 2D method. Overall, the results were consistent across both speeds. Despite several significant differences identified by the SPM1D analysis, the kinematic waveforms obtained without markers for the three joints compared reasonably well with those obtained from the marker-based system (see Figure 3). The averaged RMSE was highest for the hip joint at 5.6°, with lower values for the ankle and knee joints at 3.5° and 2.9°, respectively. Next, to assess the feasibility of testing participants in their usual running clothes instead of their underwear, we also investigated the influence of wearing running clothes on the markerless joint kinematic results. The curves obtained under both clothing conditions were highly similar, with no significant differences observed (see Figure 4), and the mean RMSE values were less than 2.8° for all joints. These findings are consistent with recent studies suggesting that clothing has minimal effect on 3D markerless walking and running kinematics measured using multiple cameras [29,30,31]. Furthermore, this indicates that the markerless method demonstrates acceptable repeatability [35] and robustness in consistently providing joint kinematics, even when participants wear their usual running clothes.
The accuracy of joint angle measurements during running cycles using OpenPifPaf is consistent with other studies that have compared pose detection methods to marker-based reference motion capture systems during gait and running. Focusing on gait, one study [36] reported mean absolute errors in the sagittal plane of 4 ± 2.1°, 5.6 ± 2.9°, and 7.4 ± 4.8° for the hip, knee, and ankle joints, respectively, using OpenPose for human pose detection. Another recent study [37] found RMSEs of 5.4 ± 2.1°, 5.7 ± 2°, and 4.6 ± 1.7° for the hip, knee, and ankle joint in the sagittal plane under normal walking conditions using OpenCap, a web-based service that utilizes videos from at least two smartphones to compute 3D movement dynamics [23]. Similarly, during running, a recent study [20] demonstrated mean RMSE values of 4.9 ± 2.2° for the hip, 6.5 ± 2.4° for the knee, and 5.7 ± 1.4° for the ankle when comparing OpenPose to an optoelectronic system, but only after temporal alignment and offset removal of their data. They also trained and tested a DeepLabCut model [38], which yields slightly higher RMSE values. This model was trained using video footage with reflective markers and was unable to detect the ankle and foot key points when videos of patients without reflective markers were tested. When comparing these results with other studies, it is important to consider that the method used to locate joint centres and the approach used to calculate joint kinematics can influence the outcomes. Additionally, the reference against which the results are compared also plays a significant role. For example, a study comparing two widely used musculoskeletal models, OpenSim and Plug-in-Gait, during running reported average RMSEs of 5.2°, 8.3°, and 5.5° for the hip, knee, and ankle at a running speed of 3.3 m.s−1 [39]. Therefore, a comparison of accuracy with other studies should be interpreted as an overview of their order of magnitude rather than as a basis for exact comparison. The lower RMSEs observed in our study are likely due to the fact that the comparison of the markerless method was made with a 2D method using markers fixed to the subjects, rather than with a 3D biomechanical model. This configuration is probably closer to the annotated pictures that the OpenPifPaf neural network was trained with. Furthermore, the same method was used to compute joint kinematics from the coordinates of the key points and markers for both the markerless and marker-based methods, ensuring a consistent approach to kinematic calculation.
From a clinical point of view, non-laboratory approaches for gait and running analysis, such as those relying on pose detection, should ideally capture joint kinematics in the sagittal plane with a maximum acceptable error of less than 5° [40,41]. The mean RMSE values for the knee and ankle angles fall below this threshold, indicating satisfactory performance. However, this is not the case for the hip, where the markerless method measures greater hip extension at toe-off, leading to a higher mean RMSE. This discrepancy might be explained by a mismatch between the spatial localisation of the markers on the acromion and the shoulder key point detected by OpenPifPaf. Additionally, at toe-off, the shoulders and pelvis are at their maximal rotation in opposite directions. With the limitation of 2D, both markers and key points are particularly susceptible to cross-talk artefacts [42], causing additional differences between the two systems. On the contrary, the RMSE for the hip is similar to that of the knee and ankle joints when comparing the markerless system with and without running clothes, highlighting the consistency of OpenPifPaf in detecting key points under both conditions. In the particular case of ACL injury, a systematic review reported alterations in knee flexion motion for the involved limb during the stance phase of running. These alterations were observed when compared with both the contralateral limb and control subjects, from 3 months to 5 years post-ACL reconstruction [8]. The authors emphasized the clinical significance of 2D video analysis in detecting and monitoring knee flexion excursion alterations during running. While traditional 2D motion analysis systems based on video offer an accessible tool to clinicians [43], pose estimation brings additional advantages. Firstly, it eliminates the tedious and time-consuming need to manually identify anatomical landmarks. Secondly, the pose estimation method provides the temporal evolution of the movement throughout its entirety, rather than being limited to discrete points of interest such as initial contact, toe-off, or peak knee flexion angle. These continuous curves enable not only the measurement and comparison of specific point values but also the evaluation of entire movement patterns, potentially allowing for the computation of more complex parameters involving temporal variations. Thirdly, using human pose detection for motion analysis when running on a treadmill allows for easy averaging over a large number of cycles, making the results more stable [44]. The use of a treadmill has also been shown to improve between-day repeatability of kinematic running assessment [45].
Our results should be interpreted within the context of our study’s limitations. First, we compared joint kinematics obtained from a markerless system with those obtained using markers placed on anatomical landmarks. This approach aligns with typical 2D motion analysis systems, such as Kinovea, but differs from the standard 3D motion analysis reference. In particular, the hip joint angle measured was between the thigh segment and the upper body segment, not between the pelvis and femur. Additionally, despite markers being placed by the same experienced operator, marker-based kinematics can be affected by palpation errors [46] and skin movement relative to the anatomical landmark [47]. Moreover, our method is limited to two-dimensional analysis exclusively in the sagittal plane, requiring perpendicular alignment of video recording to the treadmill. Deviations from this alignment could compromise measurement accuracy and relevance. Finally, our participants were healthy and predominantly Caucasian. Future studies should include a larger and more diverse sample, including actual patients, to broaden the applicability of our findings. Testing the method in real clinical conditions, outside the controlled laboratory setting, would be valuable. This would account for less standardized factors such as lighting and available space. Using a low-powered pose estimation framework [16] could enhance accessibility by reducing the need for specialized equipment and shortening computation time.
In summary, analysing treadmill running kinematics in the sagittal plane using OpenPifPaf for human pose detection yields compelling results compared to those obtained using markers placed on anatomical landmarks to simulate typical video-based 2D motion analysis systems. This method provides valuable insights into joint kinematics throughout the entire running cycle. The use of a treadmill and averaging results over a large number of cycles increase the method’s robustness. Additionally, wearing running clothes does not appear to significantly affect the method’s outcomes. Markerless motion capture shows great promise for both research and clinical practice due to its accessibility and simplicity, making it well suited for integration into clinical settings.

Author Contributions

Conceptualization: N.L., P.F. and C.D.; data curation: N.L.; formal analysis: N.L. and A.E.; funding acquisition: L.P. and C.D.; investigation: N.L.; methodology: N.L., A.P.N. and C.D.; project administration: P.F. and C.D.; resources: C.D.; software: N.L. and A.E.; supervision: P.F.; validation: N.L., P.F. and C.D.; writing—original draft: N.L.; writing—review and editing: A.E., A.P.N., P.F., L.P. and C.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study was conducted in accordance with the Declaration of Helsinki and approved by the Ethics Committee of UCLouvain (Comité d’Ethique Hospitalo-Facultaire, B403201523492, obtained 17 April 2023).

Informed Consent Statement

Informed consent was obtained from all subjects involved in this study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors would like to thank Louise Gaspard and Emeline Cleymans for their help in collecting the data.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lee, D.; Brellenthin, A.G.; Thompson, P.D.; Sui, X.; Lee, I.-M.; Lavie, C.J. Running as a Key Lifestyle Medicine for Longevity. Prog. Cardiovasc. Dis. 2017, 60, 45–55. [Google Scholar] [CrossRef] [PubMed]
  2. Fredette, A.; Roy, J.-S.; Perreault, K.; Dupuis, F.; Napier, C.; Esculier, J.-F. The Association Between Running Injuries and Training Parameters: A Systematic Review. J. Athl. Train. 2022, 57, 650–671. [Google Scholar] [CrossRef] [PubMed]
  3. Cavanagh, P.R.; Lafortune, M.A. Ground Reaction Forces in Distance Running. J. Biomech. 1980, 13, 397–406. [Google Scholar] [CrossRef] [PubMed]
  4. Van Hooren, B.; Goudsmit, J.; Restrepo, J.; Vos, S. Real-Time Feedback by Wearables in Running: Current Approaches, Challenges and Suggestions for Improvements. J. Sports Sci. 2020, 38, 214–230. [Google Scholar] [CrossRef] [PubMed]
  5. Sanders, T.L.; Maradit Kremers, H.; Bryan, A.J.; Larson, D.R.; Dahm, D.L.; Levy, B.A.; Stuart, M.J.; Krych, A.J. Incidence of Anterior Cruciate Ligament Tears and Reconstruction: A 21-Year Population-Based Study. Am. J. Sports Med. 2016, 44, 1502–1507. [Google Scholar] [CrossRef]
  6. Kotsifaki, R.; Korakakis, V.; King, E.; Barbosa, O.; Maree, D.; Pantouveris, M.; Bjerregaard, A.; Luomajoki, J.; Wilhelmsen, J.; Whiteley, R. Aspetar Clinical Practice Guideline on Rehabilitation after Anterior Cruciate Ligament Reconstruction. Br. J. Sports Med. 2023, 57, 500–514. [Google Scholar] [CrossRef] [PubMed]
  7. Rambaud, A.J.M.; Ardern, C.L.; Thoreux, P.; Regnaux, J.-P.; Edouard, P. Criteria for Return to Running after Anterior Cruciate Ligament Reconstruction: A Scoping Review. Br. J. Sports Med. 2018, 52, 1437–1444. [Google Scholar] [CrossRef]
  8. Pairot-de-Fontenay, B.; Willy, R.W.; Elias, A.R.C.; Mizner, R.L.; Dubé, M.-O.; Roy, J.-S. Running Biomechanics in Individuals with Anterior Cruciate Ligament Reconstruction: A Systematic Review. Sports Med. 2019, 49, 1411–1424. [Google Scholar] [CrossRef] [PubMed]
  9. Andriacchi, T.P.; Mündermann, A. The Role of Ambulatory Mechanics in the Initiation and Progression of Knee Osteoarthritis. Curr. Opin. Rheumatol. 2006, 18, 514–518. [Google Scholar] [CrossRef]
  10. Paterno, M.V.; Schmitt, L.C.; Ford, K.R.; Rauh, M.J.; Myer, G.D.; Huang, B.; Hewett, T.E. Biomechanical Measures during Landing and Postural Stability Predict Second Anterior Cruciate Ligament Injury after Anterior Cruciate Ligament Reconstruction and Return to Sport. Am. J. Sports Med. 2010, 38, 1968–1978. [Google Scholar] [CrossRef] [PubMed]
  11. Sigward, S.M.; Lin, P.; Pratt, K. Knee Loading Asymmetries during Gait and Running in Early Rehabilitation Following Anterior Cruciate Ligament Reconstruction: A Longitudinal Study. Clin. Biomech. 2016, 32, 249–254. [Google Scholar] [CrossRef] [PubMed]
  12. Ancillao, A.; Tedesco, S.; Barton, J.; O’Flynn, B. Indirect Measurement of Ground Reaction Forces and Moments by Means of Wearable Inertial Sensors: A Systematic Review. Sensors 2018, 18, 2564. [Google Scholar] [CrossRef]
  13. Benson, L.C.; Räisänen, A.M.; Clermont, C.A.; Ferber, R. Is This the Real Life, or Is This Just Laboratory? A Scoping Review of IMU-Based Running Gait Analysis. Sensors 2022, 22, 1722. [Google Scholar] [CrossRef] [PubMed]
  14. Guo, J.; Zhang, Q.; Chai, H.; Li, Y. Obtaining Lower-Body Euler Angle Time Series in an Accurate Way Using Depth Camera Relying on Optimized Kinect CNN. Measurement 2022, 188, 110461. [Google Scholar] [CrossRef]
  15. Pagnon, D.; Domalain, M.; Reveret, L. Pose2Sim: An End-to-End Workflow for 3D Markerless Sports Kinematics—Part 2: Accuracy. Sensors 2022, 22, 2712. [Google Scholar] [CrossRef] [PubMed]
  16. Young, F.; Mason, R.; Morris, R.; Stuart, S.; Godfrey, A. Internet-of-Things-Enabled Markerless Running Gait Assessment from a Single Smartphone Camera. Sensors 2023, 23, 696. [Google Scholar] [CrossRef] [PubMed]
  17. Cronin, N.J. Using Deep Neural Networks for Kinematic Analysis: Challenges and Opportunities. J. Biomech. 2021, 123, 110460. [Google Scholar] [CrossRef] [PubMed]
  18. Scataglini, S.; Abts, E.; Van Bocxlaer, C.; Van Den Bussche, M.; Meletani, S.; Truijen, S. Accuracy, Validity, and Reliability of Markerless Camera-Based 3D Motion Capture Systems versus Marker-Based 3D Motion Capture Systems in Gait Analysis: A Systematic Review and Meta-Analysis. Sensors 2024, 24, 3686. [Google Scholar] [CrossRef] [PubMed]
  19. D’Haene, M.; Chorin, F.; Colson, S.S.; Guérin, O.; Zory, R.; Piche, E. Validation of a 3D Markerless Motion Capture Tool Using Multiple Pose and Depth Estimations for Quantitative Gait Analysis. Sensors 2024, 24, 7105. [Google Scholar] [CrossRef]
  20. Van Hooren, B.; Pecasse, N.; Meijer, K.; Essers, J.M.N. The Accuracy of Markerless Motion Capture Combined with Computer Vision Techniques for Measuring Running Kinematics. Scand. J. Med. Sci. Sports 2023, 33, 966–978. [Google Scholar] [CrossRef] [PubMed]
  21. Haberkamp, L.D.; Garcia, M.C.; Bazett-Jones, D.M. Validity of an Artificial Intelligence, Human Pose Estimation Model for Measuring Single-Leg Squat Kinematics. J. Biomech. 2022, 144, 111333. [Google Scholar] [CrossRef]
  22. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.-E.; Sheikh, Y. OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. arXiv 2019. [Google Scholar] [CrossRef] [PubMed]
  23. Uhlrich, S.D.; Falisse, A.; Kidziński, Ł.; Muccini, J.; Ko, M.; Chaudhari, A.S.; Hicks, J.L.; Delp, S.L. OpenCap: Human Movement Dynamics from Smartphone Videos. PLoS Comput. Biol. 2023, 19, e1011462. [Google Scholar] [CrossRef]
  24. Kreiss, S.; Bertoni, L.; Alahi, A. OpenPifPaf: Composite Fields for Semantic Keypoint Detection and Spatio-Temporal Association. arXiv 2021. [Google Scholar] [CrossRef]
  25. Kreiss, S.; Bertoni, L.; Alahi, A. PifPaf: Composite Fields for Human Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 11969–11978. [Google Scholar]
  26. Zheng, C.; Wu, W.; Chen, C.; Yang, T.; Zhu, S.; Shen, J.; Kehtarnavaz, N.; Shah, M. Deep Learning-Based Human Pose Estimation: A Survey. ACM Comput. Surv. 2024, 56, 1–37. [Google Scholar] [CrossRef]
  27. Lambricht, N.; Englebert, A.; Pitance, L.; Fisette, P.; Detrembleur, C. Quantifying Performance and Joint Kinematics in Functional Tasks Crucial for Anterior Cruciate Ligament Rehabilitation Using Smartphone Video and Pose Detection. Knee 2025, 52, 171–178. [Google Scholar] [CrossRef] [PubMed]
  28. Sarafianos, N.; Boteanu, B.; Ionescu, B.; Kakadiaris, I.A. 3D Human Pose Estimation: A Review of the Literature and Analysis of Covariates. Comput. Vis. Image Underst. 2016, 152, 1–20. [Google Scholar] [CrossRef]
  29. Horsak, B. Repeatability and Minimal Detectable Change Including Clothing Effects for Smartphone-Based 3D Markerless Motion Capture. J. Biomech. 2024, 175, 112281. [Google Scholar] [CrossRef] [PubMed]
  30. Kanko, R.M.; Outerleys, J.B.; Laende, E.K.; Selbie, W.S.; Deluzio, K.J. Comparison of Concurrent and Asynchronous Running Kinematics and Kinetics From Marker-Based and Markerless Motion Capture Under Varying Clothing Conditions. J. Appl. Biomech. 2024, 40, 129–137. [Google Scholar] [CrossRef]
  31. Keller, V.T.; Outerleys, J.B.; Kanko, R.M.; Laende, E.K.; Deluzio, K.J. Clothing Condition Does Not Affect Meaningful Clinical Interpretation in Markerless Motion Capture. J. Biomech. 2022, 141, 111182. [Google Scholar] [CrossRef] [PubMed]
  32. Davis, R.B.; Õunpuu, S.; Tyburski, D.; Gage, J.R. A Gait Analysis Data Collection and Reduction Technique. Hum. Mov. Sci. 1991, 10, 575–587. [Google Scholar] [CrossRef]
  33. Pataky, T.C. One-Dimensional Statistical Parametric Mapping in Python. Comput. Methods Biomech. Biomed. Eng. 2012, 15, 295–301. [Google Scholar] [CrossRef]
  34. Novacheck, T.F. The Biomechanics of Running. Gait Posture 1998, 7, 77–95. [Google Scholar] [CrossRef] [PubMed]
  35. McGinley, J.L.; Baker, R.; Wolfe, R.; Morris, M.E. The Reliability of Three-Dimensional Kinematic Gait Measurements: A Systematic Review. Gait Posture 2009, 29, 360–369. [Google Scholar] [CrossRef]
  36. Stenum, J.; Rossi, C.; Roemmich, R.T. Two-Dimensional Video-Based Analysis of Human Gait Using Pose Estimation. PLoS Comput. Biol. 2021, 17, e1008935. [Google Scholar] [CrossRef]
  37. Horsak, B.; Eichmann, A.; Lauer, K.; Prock, K.; Krondorfer, P.; Siragy, T.; Dumphart, B. Concurrent Validity of Smartphone-Based Markerless Motion Capturing to Quantify Lower-Limb Joint Kinematics in Healthy and Pathological Gait. J. Biomech. 2023, 159, 111801. [Google Scholar] [CrossRef] [PubMed]
  38. Mathis, A.; Mamidanna, P.; Cury, K.M.; Abe, T.; Murthy, V.N.; Mathis, M.W.; Bethge, M. DeepLabCut: Markerless Pose Estimation of User-Defined Body Parts with Deep Learning. Nat. Neurosci. 2018, 21, 1281–1289. [Google Scholar] [CrossRef] [PubMed]
  39. Wouda, F.J.; Giuberti, M.; Bellusci, G.; Maartens, E.; Reenalda, J.; Van Beijnum, B.-J.F.; Veltink, P.H. On the Validity of Different Motion Capture Technologies for the Analysis of Running. In Proceedings of the 2018 7th IEEE International Conference on Biomedical Robotics and Biomechatronics (Biorob), Enschede, The Netherlands, 26–29 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1175–1180. [Google Scholar]
  40. Okahisa, T.; Matsuura, T.; Tomonari, K.; Komatsu, K.; Yokoyama, K.; Iwase, J.; Yamada, M.; Sairyo, K. Between-Day Reliability and Minimum Detectable Change of the Conventional Gait Model 2 and Plug-in Gait Model during Running. Gait Posture 2023, 100, 171–178. [Google Scholar] [CrossRef] [PubMed]
  41. Wilken, J.M.; Rodriguez, K.M.; Brawner, M.; Darter, B.J. Reliability and Minimal Detectible Change Values for Gait Kinematics and Kinetics in Healthy Adults. Gait Posture 2012, 35, 301–307. [Google Scholar] [CrossRef] [PubMed]
  42. Kadaba, M.P.; Ramakrishnan, H.K.; Wootten, M.E. Measurement of Lower Extremity Kinematics during Level Walking. J. Orthop. Res. 1990, 8, 383–392. [Google Scholar] [CrossRef]
  43. Mousavi, S.H.; Hijmans, J.M.; Moeini, F.; Rajabi, R.; Ferber, R.; van der Worp, H.; Zwerver, J. Validity and Reliability of a Smartphone Motion Analysis App for Lower Limb Kinematics during Treadmill Running. Phys. Ther. Sport 2020, 43, 27–35. [Google Scholar] [CrossRef] [PubMed]
  44. Dingenen, B.; Barton, C.; Janssen, T.; Benoit, A.; Malliaras, P. Test-Retest Reliability of Two-Dimensional Video Analysis during Running. Phys. Ther. Sport 2018, 33, 40–47. [Google Scholar] [CrossRef]
  45. Bramah, C.; Preece, S.J.; Gill, N.; Herrington, L. The Between-Day Repeatability, Standard Error of Measurement and Minimal Detectable Change for Discrete Kinematic Parameters during Treadmill Running. Gait Posture 2021, 85, 211–216. [Google Scholar] [CrossRef] [PubMed]
  46. Della Croce, U.; Leardini, A.; Chiari, L.; Cappozzo, A. Human Movement Analysis Using Stereophotogrammetry. Part 4: Assessment of anatomical landmark misplacement and its effects on joint kinematics. Gait Posture 2005, 21, 226–237. [Google Scholar] [CrossRef] [PubMed]
  47. Leardini, A.; Chiari, L.; Croce, U.D.; Cappozzo, A. Human Movement Analysis Using Stereophotogrammetry. Part 3. Soft tissue artifact assessment and compensation. Gait Posture 2005, 21, 212–225. [Google Scholar] [CrossRef] [PubMed]
Figure 1. (A) A schematic of the experimental setup where motion is simultaneously recorded by an optoelectronic system and a smartphone camera. The images show a participant in the Lab clothing condition performing the experiment, both before (B) and after (C) OpenPifPaf processing.
Figure 1. (A) A schematic of the experimental setup where motion is simultaneously recorded by an optoelectronic system and a smartphone camera. The images show a participant in the Lab clothing condition performing the experiment, both before (B) and after (C) OpenPifPaf processing.
Sensors 25 00934 g001
Figure 2. Examples of curves obtained for a participant at one speed. (A) The horizontal and vertical coordinates of the left ankle marker or key point, highlighting the synchronization trigger and the start and end indices of the cycles to be analysed. (B) The evolution of joint angles for the ankle, knee, and hip during the recording. (C) Curves showing joint angles at each cycle percentage for the 20 analysed cycles (grey) and their mean values (black).
Figure 2. Examples of curves obtained for a participant at one speed. (A) The horizontal and vertical coordinates of the left ankle marker or key point, highlighting the synchronization trigger and the start and end indices of the cycles to be analysed. (B) The evolution of joint angles for the ankle, knee, and hip during the recording. (C) Curves showing joint angles at each cycle percentage for the 20 analysed cycles (grey) and their mean values (black).
Sensors 25 00934 g002
Figure 3. Top: Marker-based and markerless motion capture sagittal plane joint angles at 2.5 m.s−1 and 3.6 m.s−1 for the ankle (left column), knee (middle column), and hip (right column) across the 30 participants. The solid lines represent the mean of the averaged angles of the 30 participants, and the shaded areas represent the standard deviation. Middle: F-statistics from SPM1D (SPMF) for the measurement method (solid line) and the interaction of speed and measurement method (dotted line), with dashed lines indicating the critical thresholds and regions of significant differences highlighted in grey. Bottom: The mean absolute error between marker-based and markerless motion capture across the 30 participants at 2.5 m.s−1 (black) and 3.6 m.s−1 (grey).
Figure 3. Top: Marker-based and markerless motion capture sagittal plane joint angles at 2.5 m.s−1 and 3.6 m.s−1 for the ankle (left column), knee (middle column), and hip (right column) across the 30 participants. The solid lines represent the mean of the averaged angles of the 30 participants, and the shaded areas represent the standard deviation. Middle: F-statistics from SPM1D (SPMF) for the measurement method (solid line) and the interaction of speed and measurement method (dotted line), with dashed lines indicating the critical thresholds and regions of significant differences highlighted in grey. Bottom: The mean absolute error between marker-based and markerless motion capture across the 30 participants at 2.5 m.s−1 (black) and 3.6 m.s−1 (grey).
Sensors 25 00934 g003
Figure 4. Top: Markerless motion capture sagittal plane joint angles at 2.5 m.s−1 and 3.6 m.s−1 for the ankle (left column), knee (middle column), and hip (right column) across the 30 participants for both clothing conditions. The solid lines represent the mean of the averaged angles of the 30 participants, and the shaded areas represent the standard deviation. Middle: F-statistics from SPM1D (SPMF) for the clothing condition (solid line) and the interaction of speed and clothing condition (dotted line), with dashed lines indicating the critical thresholds. Bottom: The mean absolute error between marker-based and markerless motion capture across the 30 participants at 2.5 m.s−1 (black) and 3.6 m.s−1 (grey).
Figure 4. Top: Markerless motion capture sagittal plane joint angles at 2.5 m.s−1 and 3.6 m.s−1 for the ankle (left column), knee (middle column), and hip (right column) across the 30 participants for both clothing conditions. The solid lines represent the mean of the averaged angles of the 30 participants, and the shaded areas represent the standard deviation. Middle: F-statistics from SPM1D (SPMF) for the clothing condition (solid line) and the interaction of speed and clothing condition (dotted line), with dashed lines indicating the critical thresholds. Bottom: The mean absolute error between marker-based and markerless motion capture across the 30 participants at 2.5 m.s−1 (black) and 3.6 m.s−1 (grey).
Sensors 25 00934 g004
Figure 5. A violin plot showing the root mean square differences between the marker-based and markerless systems for the 30 participants at 2.5 m.s−1 (left) and 3.6 m.s−1 (right) for the ankle, knee, and hip joints. The mean values and associated standard deviations for each distribution are displayed below the plots.
Figure 5. A violin plot showing the root mean square differences between the marker-based and markerless systems for the 30 participants at 2.5 m.s−1 (left) and 3.6 m.s−1 (right) for the ankle, knee, and hip joints. The mean values and associated standard deviations for each distribution are displayed below the plots.
Sensors 25 00934 g005
Table 1. Mean RMSE ± standard deviation [°] per joint averaged over all percentage points of running cycle and participants at 2.5 m.s−1 and 3.6 m.s−1.
Table 1. Mean RMSE ± standard deviation [°] per joint averaged over all percentage points of running cycle and participants at 2.5 m.s−1 and 3.6 m.s−1.
Speed2.5 m.s−13.6 m.s−1
Marker-based vs. markerless in Lab clothing condition
Ankle3.36 ± 1.24°3.59 ± 1.56°
Knee2.85 ± 0.96°2.87 ± 0.80°
Hip5.53 ± 1.45°5.61 ± 1.41°
Markerless in Lab clothing vs. Running clothes conditions
Ankle2.06 ± 0.88°2.72 ± 1.81°
Knee2.80 ± 1.75°2.58 ± 1.42°
Hip2.66 ± 1.35°2.68 ± 1.27°
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lambricht, N.; Englebert, A.; Nguyen, A.P.; Fisette, P.; Pitance, L.; Detrembleur, C. Impact of Running Clothes on Accuracy of Smartphone-Based 2D Joint Kinematic Assessment During Treadmill Running Using OpenPifPaf. Sensors 2025, 25, 934. https://doi.org/10.3390/s25030934

AMA Style

Lambricht N, Englebert A, Nguyen AP, Fisette P, Pitance L, Detrembleur C. Impact of Running Clothes on Accuracy of Smartphone-Based 2D Joint Kinematic Assessment During Treadmill Running Using OpenPifPaf. Sensors. 2025; 25(3):934. https://doi.org/10.3390/s25030934

Chicago/Turabian Style

Lambricht, Nicolas, Alexandre Englebert, Anh Phong Nguyen, Paul Fisette, Laurent Pitance, and Christine Detrembleur. 2025. "Impact of Running Clothes on Accuracy of Smartphone-Based 2D Joint Kinematic Assessment During Treadmill Running Using OpenPifPaf" Sensors 25, no. 3: 934. https://doi.org/10.3390/s25030934

APA Style

Lambricht, N., Englebert, A., Nguyen, A. P., Fisette, P., Pitance, L., & Detrembleur, C. (2025). Impact of Running Clothes on Accuracy of Smartphone-Based 2D Joint Kinematic Assessment During Treadmill Running Using OpenPifPaf. Sensors, 25(3), 934. https://doi.org/10.3390/s25030934

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop