Next Article in Journal
Transversely Excited Multipass Photoacoustic Cell Using Electromechanical Film as Microphone
Previous Article in Journal
Basic Test Framework for the Evaluation of Text Line Segmentation and Text Parameter Extraction
Previous Article in Special Issue
Non-Linearity Analysis of Depth and Angular Indexes for Optimal Stereo SLAM
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Bayesian Framework for Human Body Pose Tracking from Depth Image Sequences

Honda Research Institute USA, 800 California Street, Mountain View, CA 94041-2810, USA
*
Author to whom correspondence should be addressed.
Sensors 2010, 10(5), 5280-5293; https://doi.org/10.3390/s100505280
Submission received: 7 April 2010 / Revised: 19 May 2010 / Accepted: 20 May 2010 / Published: 25 May 2010
(This article belongs to the Special Issue Motion Detectors)

Abstract

:
This paper addresses the problem of accurate and robust tracking of 3D human body pose from depth image sequences. Recovering the large number of degrees of freedom in human body movements from a depth image sequence is challenging due to the need to resolve the depth ambiguity caused by self-occlusions and the difficulty to recover from tracking failure. Human body poses could be estimated through model fitting using dense correspondences between depth data and an articulated human model (local optimization method). Although it usually achieves a high accuracy due to dense correspondences, it may fail to recover from tracking failure. Alternately, human pose may be reconstructed by detecting and tracking human body anatomical landmarks (key-points) based on low-level depth image analysis. While this method (key-point based method) is robust and recovers from tracking failure, its pose estimation accuracy depends solely on image-based localization accuracy of key-points. To address these limitations, we present a flexible Bayesian framework for integrating pose estimation results obtained by methods based on key-points and local optimization. Experimental results are shown and performance comparison is presented to demonstrate the effectiveness of the proposed approach.

1. Introduction

For the past decades, human body pose tracking from video inputs has been an active research field motivated by various applications including human computer interaction, motion capture systems, and gesture recognition. The major challenges of recovering the large number of degrees of freedom in human body movements are the difficulties to resolve various ambiguities in the projection of human motion onto the image plane and the diversity of visual appearance caused by clothing and varying illumination.
Existing approaches for human pose tracking include methods based on single cameras, multiple cameras, and sensors beyond visible spectrum. Time-of-flight (TOF) based imaging devices have attracted researchers’ attention due to the potential to resolve depth ambiguity [14]. Robust pose tracking in 3D usually is difficult by using a single optical camera alone. In particular, methods based on silhouette information often fail to track 3D poses where there are self-occlusions. Although non-silhouette based methods [5, 6] have been proposed to track poses with self-occluded limbs, their robustness depends much on illumination conditions, body texture, and perhaps extensive training in case of learning based methods. Depth data, as in Figure 1, provides a valuable cue in resolving the depth ambiguity problem. Other advantages of TOF cameras include their portability, relatively good depth resolution compared with stereo cameras.
Most existing approaches to track human body pose from depth sequences [14] are related to the Iterative Closest Point (ICP) approach [7]. These approaches are able to track the human body pose with a high accuracy due to dense correspondences. However, these approaches based on local optimization are vulnerable to tracking failure when body parts get close to each other and often fail to recover from tracking failure afterwards. Knoop et al. [2] show that they can achieve more accurate pose tracking by integrating hand/face tracking. However, it becomes a challenging task to have a 2D hand/face tracker that works well for various complicated motion, and they do not elaborate on how the robustness of a 2D feature tracker could affect their 3D pose estimation. Zhu et al. [4] use coarse body identification to reduce the ambiguity during dense correspondence search. However, it has difficulties to detect arms when they re-appear.
Recovering from pose tracking failure is indeed an important component for a robust pose tracking algorithm. Considering example postures shown in Figure 1, on one hand, a visible arm could get so close to the torso that depth resolution is not high enough to detect the arm. Also, it is possible that a visible limb could be occluded temporarily by another limb. On the other hand, a missing limb can reappear later. A robust tracking algorithm must deal with intermittent occlusions to prevent tracking failures.
For many existing pose tracking methods, tracking long sequences will result in tracking failure which cannot be easily recovered. This paper presents a key-point based method to reconstruct poses from anatomical landmarks detected and tracked from depth image analysis. The key-point based method is robust and can recover from tracking failure when a body part is re-detected and tracked. However, its pose estimation accuracy depends solely on the image-based localization accuracy of key-points. To address these limitations, we present a Bayesian framework to integrate pose estimation results from methods using local optimization and key-point detection. Our contribution of the work is to integrate pose estimation results from multiple methods. In particular, we use results obtained by using key-points and local optimization and show that accuracy is improved compared with either method alone.
The rest of the paper is organized as follows. Section 2 introduces the human model used in this paper, and the background on pose estimation with constrained inverse kinematics. Our Bayesian method for accurate and robust pose tracking is presented in Section 3. Methods using key-points and local optimization are described in Subsections 3.1 and 3.2, respectively. Experimental results are shown in Section 4. Section 5 concludes the paper.

2. Human Body Model and Pose Estimation with Constraint Inverse Kinematics

The human body model is represented as a hierarchy of joint link models with a skin mesh attached to it as in Lewis et al. [8]. The human model in Figure 2(a) includes 28 dofs for whole body, and 20 dofs for upper body. During pose estimation, one natural constraint is to enforce joint limits. For example, by enforcing elbow joint limits, we could avoid generating the backward bending arms as in the Figure 2(b).
Let q0 be the initial model pose, V be the set of model marker points, and P be the set of observed points from the sensor. Let = ConstraintIK(q0, V, P) denote the constrained inverse kinematics as:
q ^ = q 0 + s J * ( P V )
J * = W 1 1 J T ( J W 1 1 J T + W 2 ) 1
J = [ J 1 T J i T J N T ] T
where J is the augmented Jacobian matrix, Ji is Jacobian for ith model vertex, s is a scalar to adjust the step size of inverse kinematics, W1 and W2 are defined for singularity avoidance and joint limit avoidance. This type of formulation using inverse kinematics is often used to derive manipulators orientation at each joint, when given a desired position of the end-effector. See Zhu et al. [9] for more details.
Our model marker points (for key-point detection) include the set of model vertices as shown in Figure 3. In Figure 3(a), model marker points are located at the human anatomical landmarks, and observed points are detected through low-level depth image analysis as described in Subsection 3.1. On the contrary, for model fitting, model marker points are sampled randomly from the model vertices (Figure 3(b)), and observed points are found during the ICP correspondence search as described in Subsection 3.2.

3. Robust 3D Pose Tracking with Bayesian Method

The main idea of tracking is illustrated in Figure 4. Two contrasting methods are used independently to give estimates of a human pose. The righthand side of Figure 4 represents a sparse method (based on tracking several anatomical features of the body), while the left-hand side of Figure 4 represents a dense method (tracking based on a mesh representation of the body). Each method generates hypothesis of the current pose (where output formats are a mesh and a set of anatomical landmark points, respectively) using the time sequence obtained so far. The results are integrated to produce the best estimate.
Let qt be the model pose parameters, including all degrees of freedom of the human model at time t, and p(qt|I1, I2, · · · , It) be the probability distribution of pose parameters given all observed images { I1, I2, · · ·, It}, then Bayesian tracking is formulated as:
p ( q t | I 1 , I 2 , , I t ) p ( I t | q t ) p ( q t | I 1 , I 2 , , I t 1 ) = p ( I t | q t ) q t 1 p ( q t | q t 1 ) p ( q t 1 | I 1 , I 2 , , I t 1 ) d q t 1
Let us assume that we can approximate the observation distribution as mixture of Gaussian:
p ( I t | q t ) = k = 1 K w k t N ( q t ; μ k t , Λ k t )
where N ( q t ; μ k t , Λ k t ) denotes that qt has a Gaussian distribution with mean qt and covariance Λ k t.
Let human dynamics have Gaussian noise N(0, W), the temporal propagation is given by:
p ( q t | I 1 , I 2 , , I t 1 ) = j = 1 M π j t 1 N ( q t ; f ( μ j t 1 ) , Λ j t 1 + W )
where f ( μ j t 1 ) is any appropriate pose dynamic process and π’s are weights.
Using the above Bayesian tracking equation, we can represent the posterior probability distribution as:
p ( q t | I 1 , I 2 , , I t ) = k = 1 K w k t N ( q t ; μ k t , Λ k t ) j = 1 M π j t 1 N ( q t ; f ( μ j t 1 ) , Λ j t 1 + W )
which will be, in general, a mixture of K × M Gaussian components. As we can see, this will result in an exponential increase of Gaussian components for the posterior probability distribution along the updating of time. To prevent this exponential increase in Gaussian components, we approximate it with M component Gaussian-mixture distribution:
p ( q t | I 1 , I 2 , , I t ) j = 1 M π j t N ( q t ; μ ^ j t , Λ ^ j t )
Such an approximation is reasonable in our pose estimation method as we integrate the data-driven estimation as described in Subsections 3.1. and 3.2. into Bayesian update as described in Subsection 3.4., and these data-driven estimation is very effective to allow us only maintaining a small number of pose hypotheses during the tracking.
Since we represent the posterior probability distribution as a sum of Gaussian, there are available methods to perform density approximation. One simple way is to keep the dominant modes in the posterior probability distribution. Researchers [5, 10] also suggest to pick modes from a likelihood function and combine them with compatible ones from the predicted prior probabilities. Some authors [11] also pick the modes from a likelihood function and re-weight with predicted prior probability.
The detailed illustration of this Bayesian inference method to pose tracking is shown in Figure 4, where we are able to integrate three sources of information: key-point detection from low-level image analysis, local pose optimization with ICP, and temporal prediction information if that is available. We describe these components in the following subsections.

3.1. Key-Point Detection from Depth Image Sequence for Pose Tracking

3.1.1. Body part detection

In order to have a robust pose tracker, one of the crucial processing steps is to localize each visible limb. We present a method to detect, label and track body parts using depth images as shown in Figure 5. To detect major body parts such as the head, torso, and waist, we make use of a deformable template referred to as the HNT template which consists of a head, neck, and trunk. The trunk is further decomposed into a torso and waist. They are represented by a circle, trapezoid, rectangle, and another trapezoid, respectively as in Figures 5 and 6. To localize the HNT template, our algorithm takes a background-subtracted depth image I as input and deforms the HNT template to produce the optimal template configuration by minimizing the discrepancy between the deformed HNT template and the background-subtracted depth image. See Zhu et al. [9] for more details about HNT template and its detection algorithm.
Once the head, neck, and trunk are detected, limbs (two arms and two legs) are to be detected as shown in Figure 6. For example, we can detect a upper body limb that is open, or that forms a loop, or that is in front of torso based on depth image analysis. We can detect lower limbs by finding all pixels that are lower than the waist.

3.1.2. Labeling

After the limbs are detected, we perform a labeling step in order to differentiate the left and right limbs as well as to determine the limb occlusion status. We use the following steps to label detected arms (same steps applied to leg labeling) based on the arm occlusion status at the last frame. For image frames where both arms are visible (in previous frame), let us define HLA and HRA to be the histograms of depth values for the left and right arms respectively, and we assign each pixel x in detected limb a label Lx (either Left or Right) based on its geometric and appearance distance to the tracked arms. Likelihood of x being Left Arm (LA) or Right Arm (RA) is computed by using the following formula:
P ( L x t = LA | X LA t , H LA t , X RA t , H RA t ) = e γ d LA ( x ) H LA ( I x ) e γ d LA ( x ) H LA ( I x ) + e γ d RA ( x ) H RA ( I x )
where Xt represents configuration at time t and dLA(x) is the distance from pixel x to the left arm:
d LA ( x ) = { 0 if  x is inside left arm d ( x , LA ) otherwise
where d(x, LA) is the minimal distance from x to edges of the left arm. dRA(x) is defined similarly. In short, a pixel x has a high probability of belonging to LA, if x is sufficiently close to where LA was in the previous frame. While two arms are overlapping in the image, x has a high probability of belonging to LA if it has a depth value that is close to depth values represented by the left arm in the previous frame.
When only one arm is visible from the last frame, we compute the geometric distance from the detected arm pixels to the tracked arm, and decide the label based on the maximal arm movement distance between successive frames. When both arms are not visible from the last frame, we label the detected arm based on its spatial distribution relative to the torso center line, where the left arm is located to the left of torso center line.
Finally, when the observed number of pixels for a limb is less than the threshold, we declare that the limb is occluded. For each visible limb, we preform a local optimization to align the 2-D scaled prismatic model [12] to the detected limbs.

3.1.3. Pose hypotheses from features

Key-points corresponding to the human anatomical landmarks as in Figure 3(a) are extracted from the deformed HNT template and the aligned 2-D scaled prismatic model. Due to self-occlusions, we might only be able to detect a subset of landmarks at any frame. In our Bayesian framework, we use these bottom-up depth image analysis results to improve the robustness of pose estimation and recover from tracking failure.
Referring to Figure 4, let Ps denote the extracted Key-points, and let Ms denote the corresponding subset of human anatomical landmarks. We then generate 3D pose hypotheses based on constrained inverse kinematics (defined at Equation 1). Without loss of generality, let us denote it as:
q ^ = Constraint I K ( q 0 , M s , P s )
For certain poses (e.g., straight arm), we can only obtain approximate elbow positions. Also, the estimated pose based on constrained inverse kinematics depends on starting pose values q0. Let t–1 be the optimal pose estimation from the last frame and let q t 1 0 be the human resting pose. We use the constrained inverse kinematics to generate three sets of pose hypotheses (L1 = 3 as in Figure 4):
  • H1: q t 1 is the pose generated based on both the optimal estimation t–1 and all feature points.
  • H2: q t 2 is the pose generated based on the resting pose q t 1 0 and all feature points. This hypothesis is useful to prevent the possibly erroneous estimation from the last frame.
  • H3: q t 3 is the pose generated based the optimal estimation t–1 without using the extracted elbow feature points. This hypothesis is useful to prevent the large error in elbow detection and extraction.

3.2. Temporal Prediction, Density Sampling and Dense Correspondence Searching for Pose Tracking

Since the motion to be tracked in this study is general and has high uncertainty, a common approach is to model the human pose temporal dynamics as zero velocity with a Gaussian noise N(0, W). Therefore, we can approximate the temporal prediction prior probability distribution as:
p ( q t | I 1 , I 2 , , I t 1 ) = j = 1 M π j t 1 N ( q t ; μ j t 1 , Λ j t 1 + W )
Density sampling can be performed based on this temporal prediction prior probability distribution as this is a standard Gaussian mixture distribution.
Let q t 1 i be one of samples from density sampling, V s denote a set of sampled model vertices that is visible from camera, C s denote the set of 3D depth points that is closest to V s (as shown in Figure 3(b)), and q t i denote the pose from local pose optimization:
q t i = Constraint I K ( q t 1 i , V s , C s )
We obtain visible model vertices V s from the depth buffer technique of OpenGL rendering. Closest point set C s is obtained through its grid acceleration data structure.

3.3. Tracking Error Evaluation

To evaluate tracking quality, we use a tracking error measurement function that is based on the sum of the distances from sampled depth points to their corresponding closest model vertices. Without loss of generality, let us use P s to denote the set of sampled depth points and V s the set of visible model vertices that are closest to the P s. Then, our tracking error measurement function is defined as:
d 2 ( P s , V s ( q t ) ) = j P s j V s j ( q t ) 2
With this tracking error measurement function, we can approximate the observation distribution as:
p ( I t | q t ) exp { d 2 ( P s , V s ( q t ) ) }
We can further approximate the observation distribution by keeping only a few modes from the local optimization and constrained inverse kinematics on key-points. Let { μ k t, k = 1, · · ·, k = K} denote the set of modes, we can approximate the observation distribution as:
p ( I t | q t ) k = 1 K w k t N ( q t ; μ k t , Λ k t )
where, w k t can be estimated as:
w ˜ k t exp { d 2 ( P s , V s ( μ k t ) ) }
w k t = w ˜ k t k = 1 K w ˜ k t
Λ k t can be estimated as:
Λ k t ( J V s T J V s ) 1

3.4. Bayesian Updating and MAP Selection

Given observation distribution p(It|qt) as Equation 16, and temporal prediction prior probability distribution p(qt|I1, I2, · · ·, It–1) as Equation 12, we obtain the posterior probability distribution as:
p ( q t | I 1 , I 2 , , I t ) = k = 1 k w k t N ( q t ; μ k t , Λ k t ) j = 1 M π j t 1 N ( q t ; μ j t 1 , Λ j t 1 + W )
In order to avoid the exponential increase of Gaussian components, without loss of generality, we first approximate it by the first M dominant observation modes as:
p ( q t | I 1 , I 2 , , I t ) k = 1 M w ^ k t N ( q t ; μ k t , Λ k t ) j = 1 M π j t 1 N ( q t ; μ j t 1 , Λ j t 1 + W )
and then re-weight them with temporal prior probability:
p ( q t | I 1 , I 2 , I t ) j = 1 M π j t N ( q t ; μ j t , Λ j t )
where weights π j t can be estimated as:
w ˜ j t = w ^ j t k = 1 M π k t 1 N ( μ j t ; μ k t 1 , Λ k t 1 + W )
π j t = w ˜ j t j = 1 M w ˜ j t
At any frame, the optimal pose estimation is exported as the mode in the posterior probability distribution p(qt|I1, I2, · · ·, It).

4. Experiments

The Bayesian pose tracking algorithm is implemented and tested on a set of upper and whole body sequences captured from a single time-of-flight (TOF) range sensor [13] at 16 frame per second. Upper body data sequences are captured with a distance between 1.5 m and 2 m, and whole body data sequences are captured with a distance around 3 m. Each sequence has a duration about between 10 to 15 s. Through experiments, our major goal is to show that
  • The proposed Bayesian framework is able to track robustly and recover from tracking failure by integrating low-level key-point detection from depth image analysis;
  • The proposed Bayesian framework is able to achieve a higher accuracy by taking advantage of the ICP to refine the alignment between 3D model and point clouds;
To that end, the captured data sequences include the type of human motion where it has self-occlusions between body parts. As a result, these captured data sequences are complicated, and previous methods based on local optimization fail to track them because of self-occlusion. Local optimization methods have especially poor performance to track such scenario where limb disappears and reappears again during the motion.
Our current implementation works well for body twists up to 40 degree rotation on either side of a front facing posture. Large twists and severe interaction between upper and lower body limbs remain as a challenge in the current implementation. Example upper-body and whole-body tracking results are shown in Figures 710. In all of our experiments, we use K = L1 + L2 = 3 + 3 = 6, M = 3 (refer to Equation 7). Currently these values are selected based on the empirical method. Firstly, we select L1 = 3 as explained in Subsection 3.1. Secondly, L2 = 3 and M = 3 are empirically selected based on the performance of the tracker on the example motion sequences in our database. Increasing L2 and M could improve accuracy further, but could also slow down the tracking dramatically. In contrast with our current brute-force implementation, we are seeking other implementation methods such as parallel programming techniques to take advantage of inherent parallelism between the hypothesis computations.
We summarize and compare its performance with the ICP method and key-point based method as in Table 1. The ICP method utilizes general correspondences to estimate the pose, which does not require tracking of key-points. Nevertheless, the ICP method could result in tracking failure for transient occlusions, and is difficult to recover from it. Furthermore, the ICP method could not be integrated with other information flexibly. The key-point based method is able to track through transient occlusion, and recover from tracking failures when the body parts are detected again. However, it is not able to take advantage of other information. As seen, the Bayesian-based framework is able to take advantage of both ICP and key-point based methods. It is able to track through transient occlusions, recover from tracking failure whenever body parts are detected again, and update the pose by performing local optimization without key-points. The Bayesian-based framework has the potential to make use of other information flexibly whenever available, for example, pose prediction from machine learning approaches. Furthermore, the Bayesian-based framework could achieve a higher accuracy for joint trajectories than key-point based methods because it could take advantage of ICP to refine the alignment between 3D model and point clouds, as shown in Table 2.

5. Conclusions and Future work

We have presented a Bayesian framework for human pose tracking from depth image sequences. Human pose tracking remains as a challenging problem, primarily because pose is difficult to track due to occlusion, fast movements, and ambiguity. Generating multiple hypotheses for human pose for one image is at times necessary to arrive at a correct solution. A method has been proposed to demonstrate a potential to integrate pose estimation results from different modalities to improve the robustness and accuracy. We believe the parallel nature of the hypothesis evaluation permits us to achieve a faster implementation with latest parallel programming techniques.

References

  1. Grest, D.; Woetzel, J.; Koch, R. Nonlinear body pose estimation from depth images. Proceedings of the 27th Annual Meeting of the German Association for Pattern Recognition, Vienna, Austria, August 30–September 2, 2005.
  2. Knoop, S.; Vacek, S.; Dillmann, R. Sensor fusion for 3d human body tracking with an articulated 3d body model. Proceedings of the International Conference on Robotics and Automation, Orlando, FL, USA, May 15–19, 2006; pp. 1686–1691.
  3. Ziegler, J.; Nickel, K.; Stiefelhagen, R. Tracking of the articulated upper body on multi-view stereo image sequences. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, New York, NY, USA, June 17–22, 2006; pp. 774–781.
  4. Zhu, Y.; Fujimura, K. Constrained optimization for human pose estimation from depth sequences. Proceedings of IEEE Asian Conference on Computer Vision, Tokyo, Japan, November 2007.
  5. Sminchisescu, C.; Triggs, B. Kinematic jump processes for monocular 3d human tracking. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, June 16–22, 2003; pp. 18–20.
  6. Lee, M.W.; Cohen, I. Proposal maps driven mcmc for estimating human body pose in static images. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, June 27–July 2, 2004; pp. 334–341.
  7. Besl, P.; McKay, N. A method for registration of 3-d shapes. IEEE Trans. Pattern Anal. Mach. Intell 1992, 14, 239–256. [Google Scholar]
  8. Lewis, J.P.; Cordner, M.; Fong, N. Pose space deformations: A unified approach to shape interpolation and skeleton-driven deformation. Proceedings of SIGGRAPH 2000, New Orleans, LA, USA, July 23–28, 2000; pp. 165–172.
  9. Zhu, Y.; Dariush, B.; Fujimura, K. Controlled human pose estimation from depth image streams. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition Time-of-Flight Workshop (CVPRW’08), Anchorage, AK, USA, June 23–28, 2008.
  10. Cham, T.J.; Rehg, J. A multiple hypothesis approach to figure tracking. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR’99), Ft. Collins, CO, USA, June 23–25, 1999; pp. 239–245.
  11. Demirdjian, D.; Taycher, L.; Shakhnarovich, G.; Grauman, K.; Darrell, T. Avoiding the streetlight effect: Tracking by exploring likelihood modes. Proceedings of IEEE International Conference on Computer Vision (ICCV’05), Beijing, China, October 17–20; 2005; pp. 357–364. [Google Scholar]
  12. Morris, D.; Rehg, J. Singularity analysis for articulated object tracking. Proceedings of IEEE Conference on Computer Vision and Pattern Recognition (CVPR’98), Santa Barbara, CA, USA, June 23–25, 1998; pp. 189–196.
  13. SwissRanger. Online Time-of-Flight Camera Information. Available online: http://www.mesa-imaging.ch/prodviews.php (accessed on 5 March 2010).
Figure 1. Depth data (a) Example upper body postures; (b) Example whole body postures.
Figure 1. Depth data (a) Example upper body postures; (b) Example whole body postures.
Sensors 10 05280f1
Figure 2. Human body model (a) Hierarchical joint link model with 28 dofs; (b) Elbow joint limit constraints for natural pose tracking.
Figure 2. Human body model (a) Hierarchical joint link model with 28 dofs; (b) Elbow joint limit constraints for natural pose tracking.
Sensors 10 05280f2
Figure 3. Model marker points (a) from key-point detection; (b) from dense ICP correspondences (each yellow vector represents a correspondence pair).
Figure 3. Model marker points (a) from key-point detection; (b) from dense ICP correspondences (each yellow vector represents a correspondence pair).
Sensors 10 05280f3
Figure 4. Robust pose estimation with Bayesian tracking framework.
Figure 4. Robust pose estimation with Bayesian tracking framework.
Sensors 10 05280f4
Figure 5. Body part detection, labeling and tracking.
Figure 5. Body part detection, labeling and tracking.
Sensors 10 05280f5
Figure 6. HNT template localization (shown in red) and limb detection: (a) Open arm detection; (b) Looped arm detection; (c) Arm detection that is in front of torso; (d) Lower limb detection.
Figure 6. HNT template localization (shown in red) and limb detection: (a) Open arm detection; (b) Looped arm detection; (c) Arm detection that is in front of torso; (d) Lower limb detection.
Sensors 10 05280f6
Figure 7. Upper body pose tracking for violin playing motion. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Figure 7. Upper body pose tracking for violin playing motion. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Sensors 10 05280f7
Figure 8. Upper body pose tracking for frisbee throwing motion. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Figure 8. Upper body pose tracking for frisbee throwing motion. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Sensors 10 05280f8
Figure 9. Whole body pose tracking with self occlusions during leg crossing. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Figure 9. Whole body pose tracking with self occlusions during leg crossing. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Sensors 10 05280f9
Figure 10. Whole body pose tracking during a dancing sequence. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Figure 10. Whole body pose tracking during a dancing sequence. Rows 1 and 3: depth image sequence with the detected body parts. Rows 2 and 4: corresponding reconstructed pose.
Sensors 10 05280f10
Table 1. Comparison between various human pose tracking approaches.
Table 1. Comparison between various human pose tracking approaches.
MethodsTracking through occlusionError-recoveryTracking with missing key-pointsIntegration with other informationSpeed
ICP based methodNoNoYesNo5∼9 Hz
Key-point based methodYesYesNoNo3∼6 Hz
Bayesian-based methodYesYesYesYes0.1 Hz
Table 2. A comparison of overall trajectory accuracy between key-point based method and Bayesian-based method.
Table 2. A comparison of overall trajectory accuracy between key-point based method and Bayesian-based method.
MethodsX trajectory accuracyY trajectory accuracyZ trajectory accuracy
Key-point based method80 mm84 mm93 mm
Bayesian-based method73 mm78 mm87 mm

Share and Cite

MDPI and ACS Style

Zhu, Y.; Fujimura, K. A Bayesian Framework for Human Body Pose Tracking from Depth Image Sequences. Sensors 2010, 10, 5280-5293. https://doi.org/10.3390/s100505280

AMA Style

Zhu Y, Fujimura K. A Bayesian Framework for Human Body Pose Tracking from Depth Image Sequences. Sensors. 2010; 10(5):5280-5293. https://doi.org/10.3390/s100505280

Chicago/Turabian Style

Zhu, Youding, and Kikuo Fujimura. 2010. "A Bayesian Framework for Human Body Pose Tracking from Depth Image Sequences" Sensors 10, no. 5: 5280-5293. https://doi.org/10.3390/s100505280

Article Metrics

Back to TopTop