Next Article in Journal
Visual-Aided Obstacle Climbing by Modular Snake Robot
Previous Article in Journal
Automatic Monitoring Methods for Greenhouse and Hazardous Gases Emitted from Ruminant Production Systems: A Review
Previous Article in Special Issue
A Multi-Scale Natural Scene Text Detection Method Based on Attention Feature Extraction and Cascade Feature Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Temporal–Spatial Contextual Adaptation for Three-Dimensional Human Pose Estimation

College of Information Engineering, Capital Normal University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(13), 4422; https://doi.org/10.3390/s24134422
Submission received: 3 April 2024 / Revised: 20 June 2024 / Accepted: 4 July 2024 / Published: 8 July 2024
(This article belongs to the Special Issue Computer Vision and Virtual Reality: Technologies and Applications)

Abstract

:
Three-dimensional human pose estimation focuses on generating 3D pose sequences from 2D videos. It has enormous potential in the fields of human–robot interaction, remote sensing, virtual reality, and computer vision. Existing excellent methods primarily focus on exploring spatial or temporal encoding to achieve 3D pose inference. However, various architectures exploit the independent effects of spatial and temporal cues on 3D pose estimation, while neglecting the spatial–temporal synergistic influence. To address this issue, this paper proposes a novel 3D pose estimation method with a dual-adaptive spatial–temporal former (DASTFormer) and additional supervised training. The DASTFormer contains attention-adaptive (AtA) and pure-adaptive (PuA) modes, which will enhance pose inference from 2D to 3D by adaptively learning spatial–temporal effects, considering both their cooperative and independent influences. In addition, an additional supervised training with batch variance loss is proposed in this work. Different from common training strategy, a two-round parameter update is conducted on the same batch data. Not only can it better explore the potential relationship between spatial–temporal encoding and 3D poses, but it can also alleviate the batch size limitations imposed by graphics cards on transformer-based frameworks. Extensive experimental results show that the proposed method significantly outperforms most state-of-the-art approaches on Human3.6 and HumanEVA datasets.

1. Introduction

Three-dimensional human pose estimation is a rapidly developing task at the intersection of computer vision [1,2,3], human–robot interaction [4,5,6], and sensors [7,8]. It enables reconstruction of the 3D structure of the human body from 2D images or videos, providing a more comprehensive and detailed understanding of human movement and behavior [9,10]. The advancements in 3D human pose estimation are driven by the remarkable progress in deep learning techniques and computer vision methods. The combination of these technologies has enabled more accurate segmentation and localization of body parts [11,12,13], as well as the reconstruction of their 3D spatial configuration [14,15].
The initial 3D human pose estimation based on deep learning takes single 2D image/pose as input [16,17,18]. Due to the lack of depth information, the estimated pose may appear accurate on the imaging plane, but there is a significant deviation in the world coordinates. Taking Figure 1 as an example, the second line displays three failed results from videoposed3D [16]. videoposed3D is a representative work for 3D pose estimation using a single image as input. However, when facing problems such as occlusion, complex actions, and cluttered backgrounds, accurate pose estimation is still not possible. The main reason behind this is the lack of depth clues. It is difficult to obtain accurate depth information from monitoring data or remote sources in real scenes. The posture changes expressed by continuous 2D sequences can alleviate this problem. Therefore, video-based 3D pose estimation has emerged and gradually become a research hotspot.
Related algorithms generally aim to encode the spatial locations and temporal variations of the pose sequence [20,21]. In some early works, convolution was often used in the temporal or spatial encoder [16,22]. On this basis, graph convolution is introduced to utilize the spatial correlation between different keypoints [10,23]. Recently, in response to the difficulty of inferring invisible joints, Yang et al. [24] proposed a weak supervision framework for 3D human pose estimation from a single image, which can be applied not only to 2D-to-3D pose pairs, but also to 2D individual annotations. Moreover, some methods [25,26] combine global adaptation and local generalization, which is a simple and effective unsupervised domain adaptation framework for 3D human pose estimation. With the success of transformer in other vision tasks, the transformer block has become a core component of current network architecture design [27,28,29]. For instance, MAED [30] designed a spatial–temporal encoder based on a transformer block to independently extract temporal and spatial features for 3D pose estimation. On this basis, MotionBert [31] proposed a two-branch DSTformer. It contains temporal-to-spatial and spatial-to-temporal models, which can realize pose inference from 2D to 3D through direct parallel fusion. Although these methods have promoted the development of 3D pose estimation, they have neglected the spatial–temporal synergistic effects. This leads to the estimated 3D pose sequence being prone to missing details or lagging changes.
To compensate for this deficiency, we propose in this paper a dual-adaptive spatial–temporal former (DASTFormer). It achieves spatial–temporal collaborative encoding through attention- adaptive ( A t A ) and pure adaptive ( P u A ) modes. The network’s three-branch design serves two main functions: it reduces the impact of the spatial transformer block (STB) on the temporal transformer block (TTB) when they are arranged sequentially, and it addresses the issue of overlooking the overall spatial and temporal consistency of joints that can occur with a parallel arrangement. Furthermore, regarding the feature fusion module, P u A ensures global stability, while A t A ensures local accuracy. This design can enhance the different effects of temporal and spatial encoding on each keypoint, allowing a more accurate estimation of the 3D pose. Furthermore, a novel one-more supervised training strategy with batch variance loss (BVLoss) is designed to seek the global optimum. Two-round parameter updates are performed on the same batch. And the second forward propagation result is required to be better than the first one. Through this strategy, the proposed method will deeply explore the potential association between spatial–temporal encoding and 3D pose estimation. Meanwhile, it can also alleviate the batch size limitation of graphics cards on DASTFormer. To verify the effectiveness of our proposed approach, we conduct on a series of experiments on popular Human3.6M and HumanEVA datasets. The results demonstrate that our proposed approach achieves outstanding performance on MPJPE (mean per joint position error) and P-MPJPE (MPJPE after rigid alignment by pose-processing).
Our contributions are summarized as follows:
  • Designed a dual-adaptive spatial–temporal former encoder, which contains a pure adaptive mode and an attention adaptive mode. It can enhance the different effects of temporal and spatial encoding on each keypoint to improve 3D pose estimation.
  • Designed a one-more supervised training strategy with batch variance loss, which conducts a two-round parameter update for the same batch data. It can effectively explore potential 3D location cues from 2D inputs and alleviate the batch-size limitation of the GPU (graphics processing unit).
  • State-of-the-art performance was achieved on Human3.6M and HumanEVA datasets. Apart from that, the recovery effect is better in some complex postures and in wild images.

2. Related Work

Three-dimensional human pose estimation is a task that aims to estimate the spatial location of human body key points from images or videos. This task has numerous applications in areas such as animation, remote sensing, virtual reality, and healthcare. Over the years, numerous algorithms and techniques have been proposed to address this problem. In this section, we discuss and analyze existing work from three aspects: input form, spatial–temporal cue, and training strategy.

2.1. Image and Video Inputs

In the initial studies, images served as the primary input format for 3D human pose estimation [32,33]. Zhou et al. [34] employed a weakly supervised approach, incorporating a mix of 2D and 3D labels for network training. However, this approach did not exploit sequential frames to address challenges such as occlusions, leading to unreliable performance in complex scenarios, particularly with occlusions. Building on the insights of Pavllo et al. [16], subsequent research began to consider video input by integrating temporal convolutional networks (TCN). The capacity of TCN to handle sequential data through temporal convolution proved to be advantageous. However, this method has not thoroughly explored the integration of temporal and spatial information, nor has it separately modeled spatiotemporal information. Additionally, it has not considered the positional relationships between human joints, leading to an excessive reliance on the accuracy of 2D keypoint detection in the network. Hossain and Little [35] took a step further by leveraging temporal information from a sequence of 2D joint positions to estimate a series of 3D poses. This work relied on a purpose-designed sequence-to-sequence network, incorporating long short-term memory (LSTM) units with normalized layers and temporal smoothness constraints for training [36,37]. Animepose [38] utilized scene LSTM to predict one-step actions of a person. Specifically, the initial step involves forecasting the action in the preceding frame, followed by the estimation of joint positions in subsequent frames based on the key point sequence. Lee et al. [39] introduced a novel architecture built on LSTM, with the aim of learning intrinsic representations to reconstruct 3D depth from centroids to key points. Due to the absence of depth information in 2D images, directly obtaining 3D pose from 2D images is highly unreliable. Hence, this paper focuses on video-based 3D human pose estimation. Utilizing video input offers continuous frames, allowing for improved joint localization through inter-frame complementarity. However, most methods, such as those that employ LSTM to extract temporal cues [38,39], often neglect the importance of spatial information, potentially leading to a lack of details in the estimated 3D sequences.

2.2. Spatial and Temporal Clues

In addition to network structures such as CNN and RNN, the adaptation of transformers, which has seen tremendous success in the NLP domain, has also produced significant advancements in computer vision. ViT [40] was the pioneer in applying Transformers for classification in computer vision. Yang et al. [41] introduced a network that leverages transformers to extract 2D poses, showcasing the versatility and effectiveness of this approach in the field of pose estimation. Zheng et al. [20] developed a transformer network based on ViT for 3D human pose estimation.
In the past two years, an increasing number of methods have been based on spatial–temporal transformers for this task, highlighting their growing popularity and effectiveness in the field of 3D human pose estimation [20,21,28,42,43]. Shen et al. [44] adopted a mask pose and shape estimation strategy to introduce a global transformer for long-term modeling. This strategy randomly masks features of several frames to stimulate the global transformer to learn more interframe dependencies. The local transformer is responsible for utilizing local details on the human mesh and interacting with the global transformer through the use of cross-attention. However, Zheng et al. [20] requires a fixed order of spatial and temporal encoders and only reconstructs the central frame of a video. In addition, recent work has focused on optimizing the transformer structure, given the substantial computational demands and complexity involved. Einfalt et al. [45] proposes a transformer-based pose-boosting scheme that can operate on time-sparse 2D pose sequences, but still produces time-intensive 3D pose estimation. It also demonstrates how mask labeling modeling can be used for temporal upsampling within transformer blocks, greatly reducing the overall computational complexity. Li et al. [46] proposed an improved architecture based on transformers, which simply promotes long-sequence 2D joint positions to a single 3D pose, effectively aggregating remote information into a single vector representation in a hierarchical global and local manner, significantly reducing the computational cost. Even if these methods all reduce the amount of computation, due to insufficient utilization of global information and using only the consistency of adjacent frames to solve, performance may be affected.
Overall, many of the above methods use temporal and spatial information, but they do not effectively integrate the temporal and spatial information, which can lead to the network leaning toward one side and failing to achieve the role of global perception, leading to inefficient capture of global context by the network. This results in limited performance in tasks that demand a comprehensive understanding of spatial–temporal relationships.

2.3. Diverse Training Strategies

Three-dimensional human pose estimation includes end-to-end recovery from the original image [10,32,47,48] and using 2D joint points extracted from the original image, recovering the 3D pose through joint point mapping between 2D and 3D [22,49,50,51,52,53]. However, recovering 3D human pose directly from the original image requires a higher computational cost. And moreover, the background and noise present in the original image often have a counterproductive effect on recovering 3D poses, significantly enhancing both the complexity and computational load of 3D human body recovery. Recently, methods such as CPN [13], AlphaPose [54], and HRNet [55] have become commonly used 2D joint detection algorithms. For example, HRNet [55] gradually increases the number of stages by maintaining high-resolution throughout the process, progressively adding high-to-low-resolution subnetwork structures. Then, it parallelly connects multi-resolution subnetworks and estimates keypoints on the high-resolution feature map outputted by the network. These methods utilize deep learning techniques to accurately detect the location of human joints in images, providing important information for subsequent 3D human pose recovery and providing the detected 2D joint points as input to a two-stage approach, the accuracy of 2D joint detection greatly benefits the two-stage method, and the network parameters are also lower compared to directly recovering end-to-end from the original image. Martinez et al. [22] constructed a method to recover 3D human pose from 2D joint points using simple linear layers and ReLU activation functions, utilizing high-dimensional modeling based on 2D keypoints to recover 3D human body poses, achieving promising results given the hardware limitations at the time. However, due to the simplicity of the network architecture, it could not fully capture the intricacies of the keypoints. Pavllo et al. [16] improved performance by using 2D joint point sequences as input and utilizing backprojection as a semisupervised method when labeled data are scarce. While this method provides a solution for datasets lacking 3D information annotations, the lack of precision in the recovery results can lead to cumulative errors, ultimately compromising the network’s accuracy.
Although many current methods have made various improvements to the input during network training, employing both original image inputs and 2D keypoints coordinates, and some have adopted semi-supervised approaches to address the challenge of acquiring 3D annotation data, there still lacks a method to fully utilize existing recovery data, leading to underutilization of data resources. Furthermore, most of these approaches feature simple structures and overlook the relationships between human body keypoints, resulting in poses that may not conform to conventional standards.

3. Methods

To extract spatial–temporal features and learn effective spatial–temporal correlations, a novel 3D pose estimation architecture with DASTFormer and BVLoss is proposed in this paper. Its outline is shown in Figure 2. Given an original video, the 2D pose extractor is first utilized to provide the 2D pose P 2 D as input. Then they perform linear embedding, temporal embedding, and spatial embedding through LTS embedding module. The output is marked as L T S ( P 2 D ) and becomes the input of DASTFormer. During the DASTFormer encoding process, the relationship between spatial–temporal features and the 3D pose will be enhanced to improve prediction accuracy. Unlike conventional training routines, after completing a parameter update, BVloss will be introduced to conduct a second update based on data from the same batch. This will be beneficial for finding the global optimum of the proposed algorithm. More details are illustrated in the following subsection.

3.1. Dual Adaptive Spatial–Temporal Former

The DASTFormer architecture features a well-crafted three-branch system, integrating the spatial transformer block (STB) and temporal transformer block (TTB). Within this framework, the STB and TTB model different attributes through various dimensions. The STB calculates self-attention between joints based on spatial position embedding, learning the relationships between body joints. Meanwhile, the TTB calculates self-attention between joint frames based on temporal position embedding, learning the motion trajectories of the joints. Furthermore, DASTFormer includes modeling of the overall spatiotemporal dimension, accomplishing the modeling of spatiotemporal collaborative information through the mutual influence of the three branches. As illustrated in Figure 3, each branch seamlessly incorporates residual connections tailored for spatial or temporal transformer encoders, ensuring the efficacy of information propagation across the network. The thoughtful integration of these spatial–temporal branches serves to augment the model’s capacity to capture intricate spatial and temporal dependencies, thereby enhancing its overall performance. This multibranch design facilitates comprehensive feature extraction and context-aware learning, contributing to the robustness of the proposed DASTFormer architecture.
To fortify the connectivity between these blocks, two distinct modes, namely A t A and P u A , are employed, referencing the formulation in Equation (1).  P u A uses the features of different branches predicted by the network to fit the weights of different branches, respectively. This method is stable globally, but directly assigning weights to branches can lead to ignoring the internal situation of the branches. The  A t A is intended for remodeling the internal situation of the branches, which is achieved by calculating the internal attention, thereby obtaining an internal attention weight matrix. However, although  A t A can distribute weights internally, it is susceptible to the influence of local errors. If  A t A is fully adopted, it will lead to model instability. Considering comprehensively, the attention mechanism of the low-level features in the first few layers of the model has a relatively small impact. It is better at capturing local features and short-distance dependency relationships in the input sequence. Therefore, the  A t A feature fusion method is adopted to better integrate local features. In the higher-level features of the model, as the number of layers increases, the attention mechanism can learn higher-level abstract representations. It understands the input data from a global perspective, and the P u A fusion method is adopted to make the model more stable globally.
I i = L T S ( P 2 D ) , i = 1 A t A ( F S i 1 , F T i 1 , F S T i 1 ) , 1 < i 3 P u A ( F S i 1 , F T i 1 , F S T i 1 ) , 3 < i N
In Equation (1), I i represents the input of the i-th block. F S i , F T i , F S T i , respectively, represent the spatial, temporal, and spatial–temporal features of the i-th block. A t A ( a , b , c ) and P u A ( a , b , c ) indicate attention-adaptive calculation and pure-adaptive calculation, respectively. Their forward propagation rules are shown in Figure 3, and their formula is defined in Equations (2)–(5).
According to Figure 3, the first two blocks adopt the A t A mode for spatial–temporal encoding, and the last three blocks use the P u A mode to perform three-branch fusion. The calculation of A t A is demonstrated in Equation (2),
A t A ( F S i , F T i , F S T i ) = a t M a p i · a p M a p i
where a t M a p i represents the self-attention map with F S i as query, F T i as key, and  F S T i as value. a p M a p i denotes the pure-adaptive output and is also seen as the output of P u A ( F S i , F T i , F S T i ) .
a t M a p i = S F ( F S i ( F T i ) T d T ) F S T i
a p M a p i = A d d ( α S i F S i , α S T i F S T i , α T i F T i )
where S F represents the Softmax operation, and scaling factor d represents the dimension of F T i . α denotes the adaptive weight from Equation (5). A d d signifies the element-wise addition.
[ α S i , α S T i , α T i ] = S F ( ω [ C a t ( F S i , F S T i , F T i ) ] )
where C a t represents the concatenation of the values from the three branches along the last dimension, and  ω denotes the dimension reduction of the last dimension to three dimensions.
Although A t A is meaningful, for efficiency, it is used only in the first two blocks. The last three blocks adopt the P u A mode. To verify its ability, the adaptive weights in the third block are visualized in Figure 4.
Upon scrutinizing the actual weights in the initial row, it becomes apparent that different keypoints exert diverse influences on 3D pose estimation, contingent on whether they are subjected to spatial, temporal, or spatial–temporal encoding. The postnormalization of these weights reveals a notable emphasis on the F S T branch within DASTFormer.
This observation suggests that the P u A mode adeptly assigns weights, discerning the spatial–temporal collaborative and independent impacts on 3D pose estimation. The deliberate design of the P u A mode plays a pivotal role in automatically adjusting these weights, underscoring its efficacy in uncovering intricate spatial–temporal dependencies. This nuanced approach enhances the model’s ability to discern the nuanced interplay between spatial and temporal features, contributing to the nuanced and context-aware 3D pose estimation facilitated by DASTFormer.

3.2. Supervised Training with Batch Variance

To refine the model’s spatial–temporal understanding, a two-round supervised training regimen is implemented on the same batch, as depicted in Figure 2. Following the initial training phase, where the proposed network derives the 3D pose P F i r s t 3 D , the LTS module and DASTFormer undergo updates based on the joint information from P F i r s t 3 D and P 2 D . This dual-input strategy leverages the spatial and temporal features encoded in P 2 D and the refined 3D pose P 3 D F i r s t to iteratively enhance the model’s understanding of spatial–temporal relationships. The relevant calculations are summarized in Equation (6),
L F S T = L 3 D + λ T L T + λ 2 d L 2 d ϕ , ϕ x y , x z , y z
where L 3 D is the distance error between ground truth and predicted pose P 3 D F i r s t . L T represents the velocity loss, as described in previous work [16]. And L 2 d ϕ denotes the keypoint projection loss between P 3 D F i r s t and ground truth on xy, xz, and yz planes.
L 2 d ϕ = t = 1 T j = 1 J | | P t , j ϕ ^ P t , j ϕ | | 2 , ϕ x y , x z , y z
where T represents the frame length of P 2 D , J represents the keypoint number, and P t , j x y represents the projected position of j-th keypoint on the x y plane at frame t-th frame. Similarly, xz and yz planes.
After the first parameter update, P 2 D is once again employed to derive the 3D pose for the second round of supervised training, denoted as P 3 D S e c o n d . Subsequently, the batch variance loss (BVLoss) is applied for the second update on LTS and DASTFormer. During the secondary training process, we perform two updates for the same batch of data in each epoch operation. This means that during the second update, it is not only influenced by the original batch data but also by the first network prediction results. Since the overall trend of the network during training is to fit in a better direction, the prediction results after each parameter update are often better than before the update. Therefore, after sending the same batch through the network twice, the first prediction can be used as a negative sample. This means that the process of updating network parameters is influenced by the data equivalent to two batches, which appropriately alleviates the limitation of the graphics card batch size with the graphics memory size remaining unchanged. This iterative training approach enhances the model’s capacity to capture intricate spatial–temporal dependencies, leading to improved overall performance. The formulation of BVLoss is provided in Equation (8).
L B V = M a x ( L F S T C u r L F S T O l d + m , 0 )
where L F S T O l d represents the loss between P 2 D and P 3 D F i r s t , and L F S T C u r indicates the loss between P 2 D and P 3 D S e c o n d , m is a hyperparameter obtained from experience. So far, the final objective function can be summarized as Equation (9),
L O b = γ L B V + L F S T C u r
where the γ is the weight of BVLoss. We detail an example of the codes in Algorithm 1.
Algorithm 1: Pseudo-code of training process
  • Input: I img
  • Parameter: 2D pose extractor T 2 D , DASTFormer T DAST , T s , T t and T s t represents the spatial, temporal, spatial-temporal block, respectively
  • Initialization T 2 D from posetrack
  • while  in train stage  do (
  •   Extract P 2 D = T 2 D ( I img )
  •   for  t in time  do (
  •     Achieve LTS Embedding of 2D Keypoints F 2 D according to Equation (1)
  •     for  i in depth  do
  •       Extract F S = T s ( F 2 D ) , F T = T t ( F 2 D ) , F S T = T s t ( F 2 D )
  •       Calc. different branches of weights α S , α T , α S T according to Equation (5)
  •       Calc. the output of PuA and is also seen as the a p m a p i according to Equation (4)
  •       if  i in AtAList  then
  •         Calc. the self-attention map a t m a p i according to Equation (3)
  •         Calc. the output result of AtA according to Equation (2)
  •       end
  •       Achieve final output according to Equation (1) (
  •     end
  •     if  t in first update  then
  •       Optimize T DAST according to Equations (6) and (7) // ( first update
  •     else
  •       Optimize T DAST according to Equations (6)–(8) // ( second update
  •     end
  •   end
  • end

4. Experiments

4.1. Datasets and Implementation Details

We trained and tested on two mainstream datasets (human3.6M [56], humaneva [57]), and the evaluation indicators are MPJPE and P-MPJPE, which are transformed by rotation and alignment before MPJPE. They are used to measure the average error between the estimated 3D joint point positions and the true joint point positions. Specifically, for each joint point, the Euclidean distance between the estimated joint point and the true joint point is computed, and then the distance is averaged over all joint points. P-MPJPE is often used to evaluate the performance of attitude estimation algorithms in a more comprehensive way. P-MPJPE inherits the concepts of MPJPE while introducing the idea of Procrustes analysis to more robustly measure the error in attitude estimation results. For frame f and skeleton S, the detailed calculation formula is as follows:
E M P J P E ( f , S ) = 1 N S i = 1 N s m f , S ( f ) ( i ) m gt , S ( f ) ( i ) ) 2
where N S represents the number of joints contained in the skeleton S, and taking the average value of MPJPE for the sequence. m f , S ( f ) ( i ) is the pose estimator f that returns the coordinates of the i-th joint point of skeleton S at frame f.
Human3.6M is a large-scale public dataset for 3D human pose estimation research, which is currently the most important dataset based on 3D human pose research. The dataset consists of 3.6 million 3D human poses and corresponding images, captured by four calibrated high-resolution 50 HZ cameras and precise 3D joint positions and angles from high-speed motion capture systems. Each actor was also subjected to 3D laser scanning to ensure accurate capture. The BMI index range of these action actors is between 17 and 29, ensuring a moderate range of body shape variability and different activity levels. The subjects wore their own daily clothing, rather than special motion capture suits, to maintain a sense of realism as much as possible. Data from seven subjects were used for training and validation, while data from the other four subjects were used for testing. The data were organized into 15 training actions, including various asymmetric walking poses (such as walking with hands in pockets, walking with a shoulder bag), sitting poses, lying poses, various waiting poses, and other types of poses. The actors were given detailed tasks with examples to help them plan a set of stable poses between repetitions to create training, validation, and testing sets. Then, during the execution of these tasks, the actors had considerable freedom to go beyond a strict interpretation of the task. We adopted a 17-joint 3D skeleton, following previous works [9,49,58]. Training used S1, S5, S6, S7, S8, with evaluation on S9 and S11.
HumanEva includes six actions by four actors, e.g., Walking, Jogging, Boxing, and Greeting. It established the quantitative evaluation of human pose estimation using well-defined metrics in 2D and 3D. We used actions by S1, S2, S3 for training and reserved the remaining actions (Walking, Jogging) for testing.
In the experiments, we configured the video frames to a length of 243 for the Human3.6 dataset. The network was trained for a total of 80 epochs, using a batch size of 10. The model depth was fixed at 5, and we employed eight multihead self-attention mechanisms. Both the first and second updates were executed with a learning rate of 0.0002. The specific parameters are shown in Table 1.

4.2. Comparison with the State-of-the-Art

Results on Human3.6M. In our experiments, we generated two-dimensional joint data using the method proposed in [16]. Table 2 displays the results, comparing MPJPE and P-MPJPE for 15 actions using our approach versus other methods. To ensure a fair comparison, we maintained consistent input sequence lengths (in this case, 243 frames) and kept other modules unchanged. Noticeably, categories such as phone, pose, smoke, and walk exhibit significant improvements when using the BVLoss model compared to the model without it. This suggests that through iterative parameter updates on the same dataset, the network can effectively learn the positional variations of global information, enabling secondary fusion and ultimately enhancing the model’s performance.
The model without BVLoss outperforms most state-of-the-art methods in various categories. This effectively validates our approach of incorporating both temporal and spatial positional information into DASTFormer. The designed attention-adaptive and pure-adaptive modes successfully facilitate global positional feature fusion. Our method achieved MPJPE of 39.6 mm for Protocol 1 and P-MPJPE of 33.4 mm for Protocol 2, surpassing P-STMO [21] by 3.2 mm in terms of MPJPE (7.5%).
Furthermore, we conducted a comprehensive comparison of our method with those utilizing ground truth data, as detailed in Table 3. The results underscore the significant superiority of our approach, surpassing all other methods and achieving a notable 2.2 mm improvement in terms of MPJPE (11.6%) over Diffpose [59].
Table 2. Pose estimation results under Protocol 1 and Protocol 2 on the Human3.6M Dataset. (Top table) Results for MPJPE under Protocol 1. (Bottom table) Results for P-MPJPE under Protocol 2; T denotes the number of input frames estimated by the respective approaches, (*) indicates the transformer-based methods. The best and second-best results are highlighted in bold and underlined formats, respectively. A lower A v g metric is preferable.
Table 2. Pose estimation results under Protocol 1 and Protocol 2 on the Human3.6M Dataset. (Top table) Results for MPJPE under Protocol 1. (Bottom table) Results for P-MPJPE under Protocol 2; T denotes the number of input frames estimated by the respective approaches, (*) indicates the transformer-based methods. The best and second-best results are highlighted in bold and underlined formats, respectively. A lower A v g metric is preferable.
Protocol #1PublicationTDir1.Disc.EatGreetPhonePhotoPosePur.SitSitD.SmokeWaitWalkD.WalkWalkT.Avg
LCN [60]ICCV 2019146.852.344.750.452.968.949.646.460.278.951.250.054.840.443.352.7
Xu et al. [61]CVPR 2021145.249.947.550.954.966.148.546.359.771.551.448.653.939.944.151.9
Liu et al. [58]CVPR 202024341.844.841.144.947.454.143.442.256.263.645.343.545.331.332.245.1
Chen et al. [9]TCSVT 20218142.143.841.043.846.153.542.443.153.960.545.742.146.232.233.844.6
Wehrbein et al. [62]ICCV 202120038.542.539.941.746.551.639.940.849.556.845.346.446.837.840.444.3
MHFormer [43](*)CVPR 202235139.243.140.140.944.951.240.641.353.560.343.741.143.829.830.643.0
P-STMO [21](*)ECCV 202224338.942.740.441.145.649.740.939.955.559.444.942.242.729.429.442.8
STCFormer-L [29](*)CVPR 202324338.441.236.838.042.750.538.738.252.556.841.838.440.226.227.740.5
Ours (wo BVLoss,*) 24336.840.239.434.442.250.737.836.851.960.042.138.537.926.026.540.0
Ours (w BVLoss,*) 24336.839.739.334.340.950.636.836.750.959.041.438.437.925.325.839.6
ine Protocol #2PublicationTDir1.Disc.EatGreetPhonePhotoPosePur.SitSitD.SmokeWaitWalkD.WalkWalkT.Avg
Liu et al. [58]CVPR 202024332.335.233.335.835.941.533.232.744.650.937.032.437.025.227.235.6
PoseFormer [20](*)ICCV 20218134.136.134.437.236.442.234.433.645.052.537.433.837.825.627.336.5
Chen et al. [9]TCSVT 20218133.135.333.435.936.141.732.833.342.649.437.032.736.525.527.935.6
MHFormer [43](*)CVPR 202235131.534.932.833.635.339.632.032.243.548.736.432.634.323.925.134.4
P-STMO [21](*)ECCV 202224331.335.232.933.935.439.332.531.544.648.236.332.934.423.823.934.4
GLA-GCN [63](*)ICCV 202324332.435.332.634.235.042.132.131.945.549.536.132.435.623.524.734.8
Ours (wo BVLoss,*) 24331.333.833.929.634.639.731.132.244.451.136.531.432.922.323.033.8
Ours (w BVLoss,*) 24331.133.733.829.434.039.630.331.443.549.736.031.332.822.022.633.4
Results on HumanEVA. To assess the generalizability of our model, we evaluated our approach on the HumanEVA dataset. Following the methodology introduced in [28], we used the MixSTE method for data preprocessing. Utilizing 43 frames of 2D pose sequences as input to the model, we adapted the sequence length due to the dataset’s limited samples and shorter sequences compared to the Human3.6M dataset. Furthermore, we used a smaller data sample stride (interval = 1). As depicted in Table 4, our method consistently achieved the best performance in terms of MPJPE. Furthermore, comparing models with and without BVLoss demonstrated a significant improvement in all categories, confirming the improved generalizability of our model. We achieved an MPJPE of 9.6mm with ground truth data, showcasing the superior pose accuracy of our approach.

4.3. Ablation Study

To assess the impact of each component and design in the proposed model, we conducted ablation experiments on the Human3.6M dataset under Protocol 1 with MPJPE evaluation.
Impact of Temporal–Spatial Branch. For video-based human pose estimation tasks, achieving accurate results heavily relies on the effective interaction of temporal–spatial features. As illustrated in Table 5, we initially present the results of the structural design of the temporal-spatial blocks within DASTFormer. In this context, S T and T S correspond to the sequential configurations of the temporal–spatial blocks, with S T exhibiting superior performance in sequential mode. This highlights the significance of the sequential structure in capturing temporal dependencies and enhancing the overall accuracy of the model. S + T represents a design of parallel structures, and considering the superiority of S T in the sequential structure, we introduced a three-branch structure S + S T + T . It is evident that, when the model employs the same pure adaptive mode, the three-branch network structure exhibits superior performance. This underscores the effectiveness of our method in capturing nuanced inter-frame global temporal–spatial dependencies. The intricate design of the three-branch structure allows for enhanced adaptability and improved representation of complex temporal–spatial relationships in video-based human pose estimation tasks. From the results of DASTFormer with a batch size of 24, it can be observed that there is a minor improvement in the outcomes with the increase in batch size. Furthermore, even when the batch size is the same at 24, our final structure still outperforms the best-performing structure among the aforementioned four, which is the S + S T + T configuration. Under an input sequence length of 243 frames, we achieved a 1.3 mm improvement compared to S + T , indicating the effectiveness of the proposed method in capturing larger inter-frame global temporal–spatial dependencies.
Impact of BVLoss. At the bottom of Table 5, we validated the results of adding BVLoss to DASTFormer. The experiments indicate that the inclusion of BVLoss slightly improves its performance during one more training. In addition, we investigated the impact of different batch sizes on the results. Our observations indicate that the inclusion of BVLoss in the S + S T + T branch, along with a batch size set to 24, resulted in performance inferior to the accuracy achieved by our DASTFormer with a batch size of 10. This comparison underscores the efficiency of our model in achieving high accuracy with a reduced number of data. The effectiveness of our designed BVLoss is evident in its ability to enhance learning from the available data, thereby contributing to improved model performance. This outcome validates the rationale behind incorporating BVLoss and highlights its role in reinforcing the model’s capacity to learn and adapt to diverse training scenarios.

4.4. Visualization and Analysis

For visual analysis, we conducted a visualization on the Human3.6M dataset, as depicted in Figure 5, comparing the pose estimation results to the ground truth 3D poses. The visual representation highlights the superior accuracy of our method over PoseFormer. Furthermore, we emphasize the robust performance of our approach across diverse scenarios, reinforcing its effectiveness in practical applications.
In addition, we also visualize on videos outside the dataset, as shown in Figure 6. The green arrows in groups a and b indicate accurate pose estimation, while the red arrows in group a signify deviations in the estimated pose. This illustrates the strong generalization capability of our model, demonstrating excellent performance. Even in challenging scenarios such as joint overlap or occlusion, the estimated poses maintain a high level of consistency with the actual human body poses.

5. Conclusions and Discussion

In summary, 3D human pose estimation holds paramount importance in numerous domains, serving as a foundational technology for advancements in areas such as virtual and augmented reality, rescue based on remote sensing, and sports science. Due to the lack of depth information in 2D inputs, spatial and temporal cues play an important role in inferring 3D poses. Therefore, this work proposed a novel architecture with DASTFormer and one-more supervised training. The DASTFormer can consider the spatial–temporal cooperative and independent effects on 3D pose inference by two adaptive learning. The one-more supervised training with batch variance loss is different from the common training strategy. It can conduct a two-round parameter update on the same batch data, which will not only better explore the potential relationship between spatial–temporal encoding and 3D poses, but also alleviate the batch-size limitation of graphics cards on transformer-based frameworks. To validate the effectiveness of the proposed method, a large number of experiments were tested on Human3.6 and HumanEVA datasets. The experimental results demonstrate that DASTFormer and one-more supervised training with BVLoss can significantly improve the MPJPE and P-MPJPE.
Of course, it still needs more researchers to work together to facilitate the development of 3D human pose estimation in actual application scenarios. In real scenes, human actions are often affected by occlusion and changes in perspective, which lead to challenges to action recognition. However, video-based 3D pose estimation can obtain pose information from multiple perspectives, allowing for a more comprehensive understanding of human movements and mitigating the effects of occlusion and perspective changes. In addition, the results of 3D pose estimation can also serve as auxiliary input for action recognition tasks. By combining them with pixel-level clues such as parsing and segmentation, more representative and discriminative features can be extracted to improve action recognition. Using the results of 3D pose estimation, human motions can be edited and optimized more accurately. By mapping the estimated joint position to the 3D human model, the pose and motion will be adjusted. To further improve the quality of motions, it can also be combined with physical simulation technology to make the action of the generated model more in line with the laws and constraints of the real world. In addition, it can also be used in combination with motion capture technology when obtaining posture from real human body data and applying it to generate motions, so as to achieve highly realistic animation effects.
We envision that continued research in this direction will not only foster advancements in 3D human pose estimation but also contribute to broader fields such as action recognition, animation, human–computer interaction, and virtual reality. In the future, exploring the intersection of the aforementioned directions will ultimately enrich various applications and societal benefits.

Author Contributions

Conceptualization, N.J.; methodology, H.W. and W.Q.; validation, R.Z; formal analysis, M.Z.; investigation, H.W., W.Q., and N.J.; resources, R.Z.; writing—original draft preparation, N.J. and M.Z.; writing—review and editing, N.J.; visualization, H.W. and W.Q.; supervision, M.Z.; project administration, N.J.; funding acquisition, N.J. All authors have read and agreed to the published version of the manuscript.

Funding

The research was supported by the 2023 Innovation Fund of Engineering Research Center of Integration and Application of Digital Learning Technology, Ministry of Education (1311021) and the National Natural Science Foundation of China 62201365.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Liu, W.; Bao, Q.; Sun, Y.; Mei, T. Recent advances of monocular 2D and 3D human pose estimation: A deep learning perspective. ACM Comput. Surv. 2022, 55, 1–41. [Google Scholar] [CrossRef]
  2. Höll, M.; Oberweger, M.; Arth, C.; Lepetit, V. Efficient physics-based implementation for realistic hand-object interaction in virtual reality. In Proceedings of the 2018 IEEE Conference on Virtual Reality and 3D User Interfaces (VR), Tuebingen/Reutlingen, Germany, 18–22 March 2018; pp. 175–182. [Google Scholar]
  3. Ying, J.; Zhao, X. RGB-D fusion for point-cloud-based 3D human pose estimation. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 3108–3112. [Google Scholar]
  4. Gong, J.; Fan, Z.; Ke, Q.; Rahmani, H.; Liu, J. Meta agent teaming active learning for pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleands, LA, USA, 19–24 June 2022; pp. 11079–11089. [Google Scholar]
  5. Svenstrup, M.; Tranberg, S.; Andersen, H.J.; Bak, T. Pose estimation and adaptive robot behaviour for human-robot interaction. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3571–3576. [Google Scholar]
  6. Ye, M.; Li, H.; Du, B.; Shen, J.; Shao, L.; Hoi, S.C. Collaborative refining for person re-identification with label noise. IEEE Trans. Image Process. 2021, 31, 379–391. [Google Scholar] [CrossRef]
  7. Liu, J.; Gao, Y. 3D pose estimation for object detection in remote sensing images. Sensors 2020, 20, 1240. [Google Scholar] [CrossRef] [PubMed]
  8. Yang, C.; Wang, L.; Wang, X.; Mao, S. Environment Adaptive RFID-Based 3D Human Pose Tracking With a Meta-Learning Approach. IEEE J. Radio Freq. Identif. 2022, 6, 413–425. [Google Scholar] [CrossRef]
  9. Chen, T.; Fang, C.; Shen, X.; Zhu, Y.; Chen, Z.; Luo, J. Anatomy-aware 3D human pose estimation with bone-based pose decomposition. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 198–209. [Google Scholar] [CrossRef]
  10. Cai, Y.; Ge, L.; Liu, J.; Cai, J.; Cham, T.J.; Yuan, J.; Thalmann, N.M. Exploiting spatial-temporal relationships for 3D pose estimation via graph convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2272–2281. [Google Scholar]
  11. Sun, X.; Xiao, B.; Wei, F.; Liang, S.; Wei, Y. Integral human pose regression. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 529–545. [Google Scholar]
  12. Kocabas, M.; Karagoz, S.; Akbas, E. Multiposenet: Fast multi-person pose estimation using pose residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 417–433. [Google Scholar]
  13. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded pyramid network for multi-person pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7103–7112. [Google Scholar]
  14. Saito, S.; Simon, T.; Saragih, J.; Joo, H. Pifuhd: Multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 84–93. [Google Scholar]
  15. Zheng, Z.; Yu, T.; Wei, Y.; Dai, Q.; Liu, Y. Deephuman: 3D human reconstruction from a single image. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7739–7749. [Google Scholar]
  16. Pavllo, D.; Feichtenhofer, C.; Grangier, D.; Auli, M. 3D human pose estimation in video with temporal convolutions and semi-supervised training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 7753–7762. [Google Scholar]
  17. Zhou, K.; Han, X.; Jiang, N.; Jia, K.; Lu, J. Hemlets posh: Learning part-centric heatmap triplets for 3d human pose and shape estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 3000–3014. [Google Scholar] [CrossRef] [PubMed]
  18. Li, H.; Shi, B.; Dai, W.; Zheng, H.; Wang, B.; Sun, Y.; Guo, M.; Li, C.; Zou, J.; Xiong, H. Pose-oriented transformer with uncertainty-guided refinement for 2D-to-3D human pose estimation. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington DC, USA, 7–14 February 2023; Volume 37, pp. 1296–1304. [Google Scholar]
  19. Von Marcard, T.; Henschel, R.; Black, M.J.; Rosenhahn, B.; Pons-Moll, G. Recovering accurate 3D human pose in the wild using imus and a moving camera. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 601–617. [Google Scholar]
  20. Zheng, C.; Zhu, S.; Mendieta, M.; Yang, T.; Chen, C.; Ding, Z. 3D human pose estimation with spatial and temporal transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 11656–11665. [Google Scholar]
  21. Shan, W.; Liu, Z.; Zhang, X.; Wang, S.; Ma, S.; Gao, W. P-stmo: Pre-trained spatial temporal many-to-one model for 3D human pose estimation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 461–478. [Google Scholar]
  22. Martinez, J.; Hossain, R.; Romero, J.; Little, J.J. A simple yet effective baseline for 3D human pose estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2640–2649. [Google Scholar]
  23. Xu, T.; Takano, W. Graph stacked hourglass networks for 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021; pp. 16105–16114. [Google Scholar]
  24. Yang, C.Y.; Luo, J.; Xia, L.; Sun, Y.; Qiao, N.; Zhang, K.; Jiang, Z.; Hwang, J.N.; Kuo, C.H. CameraPose: Weakly-Supervised Monocular 3D Human Pose Estimation by Leveraging In-the-wild 2D Annotations. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2924–2933. [Google Scholar]
  25. Chai, W.; Jiang, Z.; Hwang, J.N.; Wang, G. Global Adaptation meets Local Generalization: Unsupervised Domain Adaptation for 3D Human Pose Estimation. arXiv 2023, arXiv:2303.16456. [Google Scholar]
  26. Tu, Z.; Zhang, J.; Li, H.; Chen, Y.; Yuan, J. Joint-bone fusion graph convolutional network for semi-supervised skeleton action recognition. IEEE Trans. Multimed. 2022, 25, 1819–1831. [Google Scholar] [CrossRef]
  27. Zhao, Q.; Zheng, C.; Liu, M.; Wang, P.; Chen, C. PoseFormerV2: Exploring Frequency Domain for Efficient and Robust 3D Human Pose Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 8877–8886. [Google Scholar]
  28. Zhang, J.; Tu, Z.; Yang, J.; Chen, Y.; Yuan, J. Mixste: Seq2seq mixed spatio-temporal encoder for 3d human pose estimation in video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 13232–13242. [Google Scholar]
  29. Tang, Z.; Qiu, Z.; Hao, Y.; Hong, R.; Yao, T. 3D Human Pose Estimation With Spatio-Temporal Criss-Cross Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 4790–4799. [Google Scholar]
  30. Wan, Z.; Li, Z.; Tian, M.; Liu, J.; Yi, S.; Li, H. Encoder-decoder with multi-level attention for 3D human shape and pose estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 13033–13042. [Google Scholar]
  31. Zhu, W.; Ma, X.; Liu, Z.; Liu, L.; Wu, W.; Wang, Y. MotionBERT: A Unified Perspective on Learning Human Motion Representations. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023. [Google Scholar]
  32. Dabral, R.; Mundhada, A.; Kusupati, U.; Afaque, S.; Sharma, A.; Jain, A. Learning 3D human pose from structure and motion. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 668–683. [Google Scholar]
  33. Pavlakos, G.; Zhou, X.; Daniilidis, K. Ordinal depth supervision for 3D human pose estimation. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7307–7316. [Google Scholar]
  34. Zhou, X.; Huang, Q.; Sun, X.; Xue, X.; Wei, Y. Towards 3D human pose estimation in the wild: A weakly-supervised approach. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 398–407. [Google Scholar]
  35. Hossain, M.R.I.; Little, J.J. Exploiting temporal information for 3D human pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 68–84. [Google Scholar]
  36. Si, C.; Jing, Y.; Wang, W.; Wang, L.; Tan, T. Skeleton-based action recognition with spatial reasoning and temporal stack learning. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 103–118. [Google Scholar]
  37. Li, C.; Wang, P.; Wang, S.; Hou, Y.; Li, W. Skeleton-based action recognition using LSTM and CNN. In Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China, 10–14 July 2017; pp. 585–590. [Google Scholar]
  38. Kumarapu, L.; Mukherjee, P. Animepose: Multi-person 3D pose estimation and animation. Pattern Recognit. Lett. 2021, 147, 16–24. [Google Scholar] [CrossRef]
  39. Lee, K.; Lee, I.; Lee, S. Propagating lstm: 3d pose estimation based on joint interdependency. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 119–135. [Google Scholar]
  40. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  41. Yang, S.; Quan, Z.; Nie, M.; Yang, W. Transpose: Towards explainable human pose estimation by transformer. arXiv 2020, arXiv:2012.14214. [Google Scholar]
  42. Lin, K.; Wang, L.; Liu, Z. End-to-end human pose and mesh reconstruction with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 1954–1963. [Google Scholar]
  43. Li, W.; Liu, H.; Tang, H.; Wang, P.; Van Gool, L. Mhformer: Multi-hypothesis transformer for 3D human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–24 June 2022; pp. 13147–13156. [Google Scholar]
  44. Shen, X.; Yang, Z.; Wang, X.; Ma, J.; Zhou, C.; Yang, Y. Global-to-Local Modeling for Video-based 3D Human Pose and Shape Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 8887–8896. [Google Scholar]
  45. Einfalt, M.; Ludwig, K.; Lienhart, R. Uplift and upsample: Efficient 3D human pose estimation with uplifting transformers. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2903–2913. [Google Scholar]
  46. Li, W.; Liu, H.; Ding, R.; Liu, M.; Wang, P.; Yang, W. Exploiting temporal contexts with strided transformer for 3d human pose estimation. IEEE Trans. Multimed. 2022, 25, 1282–1293. [Google Scholar] [CrossRef]
  47. Zhang, T.; Huang, B.; Wang, Y. Object-occluded human shape and pose estimation from a single color image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 13–19 June 2020; pp. 7376–7385. [Google Scholar]
  48. Luvizon, D.C.; Picard, D.; Tabia, H. Multi-task deep learning for real-time 3D human pose estimation and action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2752–2764. [Google Scholar] [CrossRef] [PubMed]
  49. Pavlakos, G.; Zhou, X.; Derpanis, K.G.; Daniilidis, K. Coarse-to-fine volumetric prediction for single-image 3D human pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7025–7034. [Google Scholar]
  50. Tekin, B.; Márquez-Neila, P.; Salzmann, M.; Fua, P. Learning to fuse 2D and 3D image cues for monocular body pose estimation. In Proceedings of the IEEE international Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3941–3950. [Google Scholar]
  51. Rayat Imtiaz Hossain, M.; Little, J.J. Exploiting temporal information for 3D pose estimation. arXiv 2017, arXiv:1711. 08585. [Google Scholar]
  52. Šajina, R.; Ivašić-Kos, M. 3D Pose Estimation and Tracking in Handball Actions Using a Monocular Camera. J. Imaging 2022, 8, 308. [Google Scholar] [CrossRef] [PubMed]
  53. Liu, J.; Rojas, J.; Li, Y.; Liang, Z.; Guan, Y.; Xi, N.; Zhu, H. A graph attention spatio-temporal convolutional network for 3D human pose estimation in video. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 3374–3380. [Google Scholar]
  54. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. Rmpe: Regional multi-person pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2334–2343. [Google Scholar]
  55. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 5693–5703. [Google Scholar]
  56. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  57. Sigal, L.; Balan, A.O.; Black, M.J. Humaneva: Synchronized video and motion capture dataset and baseline algorithm for evaluation of articulated human motion. Int. J. Comput. Vis. 2010, 87, 4–27. [Google Scholar] [CrossRef]
  58. Liu, R.; Shen, J.; Wang, H.; Chen, C.; Cheung, S.c.; Asari, V. Attention mechanism exploits temporal contexts: Real-time 3D human pose reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 13–19 June 2020; pp. 5064–5073. [Google Scholar]
  59. Gong, J.; Foo, L.G.; Fan, Z.; Ke, Q.; Rahmani, H.; Liu, J. Diffpose: Toward more reliable 3D pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 13041–13051. [Google Scholar]
  60. Ci, H.; Wang, C.; Ma, X.; Wang, Y. Optimizing network structure for 3d human pose estimation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2262–2271. [Google Scholar]
  61. Xu, Y.; Zhu, S.C.; Tung, T. Denserac: Joint 3d pose and shape estimation by dense render-and-compare. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7760–7770. [Google Scholar]
  62. Wehrbein, T.; Rudolph, M.; Rosenhahn, B.; Wandt, B. Probabilistic monocular 3D human pose estimation with normalizing flows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 11199–11208. [Google Scholar]
  63. Yu, B.X.; Zhang, Z.; Liu, Y.; Zhong, S.h.; Liu, Y.; Chen, C.W. GLA-GCN: Global-local Adaptive Graph Convolutional Network for 3D Human Pose Estimation from Monocular Video. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 1–12. [Google Scholar]
Figure 1. Some failed visualize examples of 3D pose estimation using a single image from the wild dataset 3DPW [19] as input. The first line refers to the raw inputs.The second line shows the 3D pose estimated by videoposed3D [16]. (a) When the body is obstructed, the estimated pose of the right arm deviates. (b) When the body is in a complex posture, there is unexpected overlap in the 3D pose of upper body. (c) When the background is cluttered, there is an incorrect association between the left and right legs.
Figure 1. Some failed visualize examples of 3D pose estimation using a single image from the wild dataset 3DPW [19] as input. The first line refers to the raw inputs.The second line shows the 3D pose estimated by videoposed3D [16]. (a) When the body is obstructed, the estimated pose of the right arm deviates. (b) When the body is in a complex posture, there is unexpected overlap in the 3D pose of upper body. (c) When the background is cluttered, there is an incorrect association between the left and right legs.
Sensors 24 04422 g001
Figure 2. Outline of the proposed method. LTS and DASTFormer are responsible for feature encoding. BVLoss is only applied for the second supervised training and guides the 3D Pose_After results surpass the 3D Pose_Before. Best viewed in color.
Figure 2. Outline of the proposed method. LTS and DASTFormer are responsible for feature encoding. BVLoss is only applied for the second supervised training and guides the 3D Pose_After results surpass the 3D Pose_Before. Best viewed in color.
Sensors 24 04422 g002
Figure 3. DASTFormer. DASTFormer consists of N spatial–temporal blocks (in grey) with two adaptive modes. The green subgraph on the left represents the attention-adaptive mode ( A t A ), while the blue part on the right shows the pure-adaptive mode ( P u A ).
Figure 3. DASTFormer. DASTFormer consists of N spatial–temporal blocks (in grey) with two adaptive modes. The green subgraph on the left represents the attention-adaptive mode ( A t A ), while the blue part on the right shows the pure-adaptive mode ( P u A ).
Sensors 24 04422 g003
Figure 4. Visualizations of PuA in Block 3. The first row presents the real weights, while the second row depicts the normalized weights. Each column represents the attention weights α S , α T , and α S T , respectively. The x-axis and y-axis represent frame number and keypoint id.
Figure 4. Visualizations of PuA in Block 3. The first row presents the real weights, while the second row depicts the normalized weights. Each column represents the attention weights α S , α T , and α S T , respectively. The x-axis and y-axis represent frame number and keypoint id.
Sensors 24 04422 g004
Figure 5. Qualitative comparison with PoseFormer [20] and GT. Our method is qualitatively compared with PoseFormer [20] on some actions in Human3.6M. The blue circles highlight positions where our method achieves superior results.
Figure 5. Qualitative comparison with PoseFormer [20] and GT. Our method is qualitatively compared with PoseFormer [20] on some actions in Human3.6M. The blue circles highlight positions where our method achieves superior results.
Sensors 24 04422 g005
Figure 6. Visualization under challenging in real-world videos. The green arrows indicate accurate pose estimation, while the red arrows signify deviations in the estimated pose. The labels (a) and (b) represent two different videos.
Figure 6. Visualization under challenging in real-world videos. The green arrows indicate accurate pose estimation, while the red arrows signify deviations in the estimated pose. The labels (a) and (b) represent two different videos.
Sensors 24 04422 g006
Table 1. Human Posture Estimation with Parameter Settings for Different Datasets in Training Phase.
Table 1. Human Posture Estimation with Parameter Settings for Different Datasets in Training Phase.
configHuman3.6MHumanEVA
optimizerlearning_rate0.00002
weight_decay0.01
lr_decay0.99
modelmaxlen24343
dim_feat256
mlp_ratio4
depth5
dim_rep512
num_heads8
datadata_stride8122
num_joints1715
Batchsize1064
Table 3. Pose estimation results of MPJPE on Human3.6M under Protocol 1 using 2D ground truth keypoints as input. The best results are highlighted in bold.
Table 3. Pose estimation results of MPJPE on Human3.6M under Protocol 1 using 2D ground truth keypoints as input. The best results are highlighted in bold.
Protocol #1PublicationTDir1.Disc.EatGreetPhonePhotoPosePur.SitSitD.SmokeWaitWalkD.WalkWalkT.Avg
Liu et al. [58]CVPR 202024334.537.133.634.232.937.139.635.840.741.433.033.833.026.626.934.7
PoseFormer [20]ICCV 20218130.033.629.931.030.233.334.831.437.838.631.731.529.023.323.131.3
MHFormer [43]CVPR 202235127.732.129.128.930.033.933.031.237.039.333.031.029.422.223.030.5
P-STMO [21]ECCV 202224328.530.128.627.929.833.231.327.836.037.429.729.528.121.021.029.3
Diffpose [59]CVPR 202324318.619.318.018.418.321.521.519.123.622.318.618.818.312.813.918.9
Ours (w BVLoss) 24316.817.816.515.716.717.818.418.920.921.017.315.315.510.210.816.7
Table 4. The MPJPE on HumanEva testset under Protocol 1. The best result is highlighted in bold.
Table 4. The MPJPE on HumanEva testset under Protocol 1. The best result is highlighted in bold.
Protocol #1PublicationT Walk Jog Avg
Pavllo et al. [16]CVPR20198114.012.527.120.317.917.518.2
zheng et al. [20]ICCV20214314.410.246.622.713.413.420.1
MixSTE [28]CVPR20224312.710.917.622.615.817.016.1
Ours (w/o BVLoss) 4310.36.814.414.68.58.910.6
Ours (w BVLoss) 439.65.712.313.87.98.19.6
Table 5. An ablation analysis of individual components within our methodology, evaluated using MPJPE in millimeters on the Human3.6M dataset.
Table 5. An ablation analysis of individual components within our methodology, evaluated using MPJPE in millimeters on the Human3.6M dataset.
ComponentBatchsizePuAAtABVLossMPJPE
T S 10--41.1
S T 10--40.9
S + T 1041.7
S + S T + T 1040.4
DASTFormer1040.0
DASTFormer1039.6
S + S T + T 2440.0
DASTFormer2439.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Quan, W.; Zhao, R.; Zhang, M.; Jiang, N. Learning Temporal–Spatial Contextual Adaptation for Three-Dimensional Human Pose Estimation. Sensors 2024, 24, 4422. https://doi.org/10.3390/s24134422

AMA Style

Wang H, Quan W, Zhao R, Zhang M, Jiang N. Learning Temporal–Spatial Contextual Adaptation for Three-Dimensional Human Pose Estimation. Sensors. 2024; 24(13):4422. https://doi.org/10.3390/s24134422

Chicago/Turabian Style

Wang, Hexin, Wei Quan, Runjing Zhao, Miaomiao Zhang, and Na Jiang. 2024. "Learning Temporal–Spatial Contextual Adaptation for Three-Dimensional Human Pose Estimation" Sensors 24, no. 13: 4422. https://doi.org/10.3390/s24134422

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop