Next Article in Journal
Accuracy Improvement of Energy Expenditure Estimation Through Neural Networks: A Pilot Study
Previous Article in Journal
Explainable Machine Learning in Critical Decision Systems: Ensuring Safe Application and Correctness
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Differencing-Based Hybrid Network for Improved 3D Skeleton-Based Motion Prediction

1
School of Electronic Engineering and Computer Science, Queen Mary University of London, London E1 4NS, UK
2
College of Electronics and Information Engineering, Shenzhen University, Shenzhen 518060, China
*
Author to whom correspondence should be addressed.
AI 2024, 5(4), 2897-2913; https://doi.org/10.3390/ai5040139
Submission received: 11 September 2024 / Revised: 5 November 2024 / Accepted: 3 December 2024 / Published: 11 December 2024

Abstract

:
Background: Three-dimensional skeleton-based human motion prediction is an essential and challenging task for human–machine interactions, aiming to forecast future poses given a history of previous motions. However, existing methods often fail to effectively model dynamic changes and optimize spatial–temporal features. Methods: In this paper, we introduce Dynamic Differencing-based Hybrid Networks (2DHnet), which addresses these issues with two innovations: the Dynamic Differential Dependencies Extractor (2D-DE) for capturing dynamic features like velocity and acceleration, and the Attention-based Spatial–Temporal Dependencies Extractor (AST-DE) for enhancing spatial–temporal correlations. The 2DHnet combines these into a dual-branch network, offering a comprehensive motion representation. Results: Experiments on the Human3.6M and 3DPW datasets show that 2DHnet significantly outperforms existing methods, with average improvements of 4.7% and 26.6% in MPJPE, respectively.

1. Introduction

Human motion prediction is a key task in human–computer interaction and computer vision, with vital applications in areas such as security surveillance [1], human–robot interaction [2], and virtual gaming [3]. The aim is to predict future human poses from past movements with accuracy. This task is challenging because of the high dimensionality, complexity, and dynamic nature of human motion.
In the past, human motion prediction models have been built using Gaussian process latent variable models [4] and hidden Markov models [5]. However, there has been a recent shift towards deep learning approaches, leveraging a variety of neural networks, including recurrent neural networks (RNNs) [6,7,8,9], graph convolutional networks (GCNs) [10,11,12,13,14,15], Transformers [16], and multi-layer perceptrons (MLPs) [17,18,19].
Although promising results have been achieved in previous works, existing approaches do not efficiently address two main challenging factors. Firstly, existing methods often struggle to effectively model rapid changes or abrupt dynamics in actions, ignoring the effectiveness of the dynamic characteristics for motion prediction; e.g., they capture the acceleration when jumping from a standing position or deceleration when coming to a sudden stop. Acceleration information helps the model to understand abrupt changes or dynamic patterns in motion. Secondly, previous methods typically focus on optimizing a single type of feature or refining specific network architectures to enhance the extraction of spatial or temporal features.
To address the first issue, we propose the Dynamic Differential Dependencies Extractor (2D-DE), which is inspired by the observation that differential operation can capture dynamic features for human motion modeling. For example, first-order temporal differencing calculates the positional changes in skeletal key points between consecutive frames, effectively capturing the velocity of each key point. Velocity provides insights into the speed and dynamic variations in the motion, such as the swinging speed of an arm or the leg movement speed while walking, which is valuable for understanding and predicting the progression of the action. Additionally, second-order temporal differencing further computes the acceleration of skeletal key points, which reveals the acceleration or deceleration patterns in the motion. For example, it captures the acceleration when jumping from a standing position or deceleration when coming to a sudden stop. Acceleration information helps the model to understand abrupt changes or dynamic patterns in motion.
For the second issue, we develop an Attention-based Spatial–Temporal Dependencies Extractor (AST-DE) to capture compact and effective motion dependencies at the spatial and temporal levels. In AST-DE, we propose Temporal-wise Attention and Spatial-wise Feature Extractor to fully exploit the spatial–temporal kinematic correlations, respectively.
Building on these two modules, we propose Dynamic Differencing-based Hybrid Networks (2DHnet), a novel model that uses a dual-branch network to learn motion representations for improved prediction (see Figure 1). 2DHnet aims to combine spatiotemporal dependency features with dynamic differential features, which provides a more holistic representation of human motion, encompassing both static skeletal structure and dynamic motion changes. Extensive experiments on two large-scale datasets, Human3.6M and 3DPW, demonstrate that our model outperforms most state-of-the-art methods in both short-term and long-term predictions regarding effectiveness and efficiency. The main contributions of this paper are as follows:
  • We propose Dynamic Differencing-based Hybrid Networks (2DHnet), providing a more holistic representation of human motion, encompassing both static skeletal structure and dynamic motion changes, and enabling effective human motion prediction.
  • In 2DHnet, we develop the Attention-based Spatial–Temporal Dependencies Extractor (AST-DE), which includes two main components: (1) Temporal-wise Attention, to capture features from historical information in the time dimension; and (2) Spatial-wise Feature Extractor, to capture spatial features between human joints.
  • In 2DHnet, we develop the Dynamic Differential Dependencies Extractor (2D-DE), which leverages multiple differential operations on 3D skeleton data to extract rich dynamic features, enabling 2DHnet to more accurately capture dynamic information such as velocity and acceleration, which enhances prediction accuracy.
  • We conduct experiments to quantitatively and qualitatively verify that our proposed 2DHnet consistently outperforms existing methods, by 4.7 % , 26.6 % of MPJPE on average on the Human3.6M and 3DPW datasets, respectively.
The rest of this paper is organized as follows: We review the related work on 3D human motion prediction most relevant to our approach in Section 2. Section 3 mainly introduces the overall framework and method. Section 4 describes the details of the experimental implementation and test results on the two datasets to demonstrate the effectiveness of our model. Section 5 concludes the work.

2. Related Work

2.1. Three-Dimensional Skeleton-Based Human Motion Prediction

The extensive research into neural networks has catalyzed significant progress in the domain of motion forecasting based on 3D skeletal data. Among various predictive models, recurrent neural networks (RNNs) stand out as a prevalent choice for tackling tasks involving the prediction of sequences, such as anticipating the trajectory of human body movements. These networks are particularly adept at discerning patterns within temporal data, which is crucial for accurately simulating the future positions of skeletal joints. Li et al. [20] introduced a human motion prediction approach grounded in a convolutional sequence-to-sequence model, leveraging RNNs to capture the temporal characteristics of human movement sequences. Li et al. [21] proposed an independent recurrent neural network architecture designed to construct longer and deeper RNN models, thereby enhancing the performance of human motion prediction. Liu et al. [22] presented an LSTM network based on global context awareness from skeletal data. Tang et al. [23] combined RNNs with an attention mechanism for human pose prediction. This approach takes into account the spatial relationships between different joints and their temporal correlations, utilizing the query, key, and value of the attention mechanism to select the necessary information for subsequent predictions.
RNNs have traditionally shown their strength in predicting the temporal aspects of motion, leveraging their ability to process sequences and recognize patterns over time. In contrast, graph convolutional networks (GCNs) have been particularly adept at analyzing spatial relationships within data, effectively handling the structural aspects of motion by operating on graphs that represent the spatial configuration of entities. Zhong et al. [24] introduced a spatio-temporal gated adjacency GCN, which captured complex spatio-temporal dependencies by balancing the weighting of spatial and temporal modeling and integrating decoupled spatio-temporal features. He et al. [25] proposed a two-stage model for human motion prediction that builds on enhanced GCNs. This model first produces preliminary predictions through spatial attention graph convolutional layers, then refines these predictions with causal temporal graph convolutional layers to improve accuracy. Fu et al. [26] developed a model that applies GCNs on spatiotemporal graphs to capture dynamic spatiotemporal dependencies among human joints, thereby predicting motion trajectories. To achieve a more detailed and enriched representation of joint connections, Gu et al. [14] introduced a dual-path GCN framework that learns the GCN adjacency matrices of two paths interactively, capturing the correlations between joint positions and velocities.
Transform-based models like Transformers offer parallel processing and enhanced spatio-temporal feature capture. Martínez-González et al. [27] proposed Pose Transformer (POTR), a non-autoregressive Transformer architecture capable of decoding elements in the query sequence in parallel, thereby reducing computational complexity and potentially preventing error propagation in long-term predictions. Mi et al. [28] presented a Transformer network that integrates spatial and positional encoding for human motion prediction from skeletal data. This approach enhances GCNs through the incorporation of a deep, multilayered residual graph framework, effectively reducing over-smoothing effects and more precisely modeling spatial relationships. Zhao et al. [29] introduced a Bidirectional Transformer GAN (BiTGAN), utilizing bidirectional Transformer encoders and decoders to capture historical motion information and employing Soft-DTW loss to maintain similarity between predictions and actual motions.

2.2. MLP-Based Feature Modeling

The employment of MLP-based methods for feature extraction has demonstrated their prowess in unearthing intricate data patterns through the intricate interplay of neurons within their layered frameworks.
MLP has been widely applied in the field of object detection. Menese et al. [30] proposed SmartSORT for real-time multi-object tracking. The method leverages the powerful feature extraction capabilities of MLPs to achieve fast and accurate tracking of multiple targets. Cao et al. [31] employed a Query-Independent Category Supervision (QICS) method for modeling category information and introduced a deep MLP to capture both long-range and short-range information, which integrated a deep MLP into a detection framework based on Transformers. Chen et al. [32] proposed CycleMLP based on MLP for dense visual prediction tasks such as object detection, and enhanced the transmission and fusion of features through recurrent connections, thereby improving detection performance.
MLP has also been applied to facial recognition. Boughrara et al. [33] proposed a constructive training algorithm, which incrementally adds hidden neurons to the network in a manner that enhances its ability to learn and generalize from facial data. Shahreza et al. [34] aimed to address privacy protection in facial recognition and proposed MLP-Hash, a method that transforms facial templates into irreversible and unlinkable forms by passing them through a randomized MLP and then hashing the output. This approach ensures that sensitive biometric data remain secure while still allowing for recognition tasks.
In addition, MLP has been proven effective in the field of 3D skeleton-based motion analysis. Guo et al. [19] proposed a network only using MLP to extract the spatial and temporal features, respectively, and proved that MLP can efficiently extract patterns from human historical pose sequences, greatly reducing the complexity and parameter count of the model. As a classic approach to feature extraction, the sustained prominence of MLPs in the field of computer vision indicates a continuous evolution, characterized by ongoing enhancements that tailor MLPs to meet the shifting complexities of advanced predictive analytics. Bouazizi et al. [35] introduced Mixer, using a spatial MLP to extract fine-grained spatial dependencies of body joints and a temporal MLP to model their interactions over time, ultimately aggregating these spatial–temporal mixed features to predict future human poses.

3. Method

3.1. Overall Architecture

Our task is to forecast the future sequence of human poses based on the given past poses, using multiple frames of 3D coordinates to represent an action. Specifically, let s t R 3 × K represent the pose matrix at the t-th frame, where K denotes the number of body joints. S 1 : T = [ s 1 , s 2 , . . . , s T ] denotes T consecutive frames of historical human poses and the goal is to predict the future N frames of the actions, which are denoted as S T + 1 : T + N = [ s T + 1 , s T + 2 , . . . , s T + N ] .
Figure 1 indicates the whole structure of our network, which consists of two key components: Attention-based Spatial–Temporal Dependencies Extractor (AST-DE) to capture compact and effective motion dependencies at spatial and temporal levels, and the Dynamic Differential Dependencies Extractor (2D-DE) to capture the dynamic characteristics of movements. In AST-DE, the model learns the temporal and spatial features of action sequences. Inspired by Mao et al. [13], we propose Temporal-wise Attention for comparing similarities in historical sequences and extracting temporal information. In the Spatial-wise Feature Extractor, we use a GCN to extract the position information and correlation degree of joint points in human motion. For 2D-DE, we devised a technique to compute the differences within a given sequence. We prepend one frame to the beginning of the resulting sequence to maintain consistent data dimensions. This process is repeated P times, yielding P distinct series that capture various dynamics of human motion, such as velocity, acceleration, and finer motion variations. Finally, we merge the information extracted from the two branches to obtain a more complete and representative feature representation. Throughout the process, we utilize Discrete Cosine Transform (DCT) [12] and Inverse Discrete Cosine Transform (IDCT) transformations, transforming the input sequence from the time domain to the frequency domain, representing the original sequence with DCT coefficients to improve the robustness and generalization ability of the model, and finally using IDCT to transform the sequence back to the time domain as the output.

3.2. Attention-Based Spatial–Temporal Dependencies Extractor (AST-DE)

This section of the network separately extracts spatial and temporal features from the sequence in Euclidean space and then fuses them. The model takes an input sequence of T frames representing an action and outputs a predicted sequence of N frames. It comprises two main components: (1) Temporal-wise Attention, which takes advantage of Attention mechanism and extracts features from historical information along the time dimension; and (2) Spatial-wise Feature Extractor, which captures positional relationships such as connectivity and other spatial characteristics between human joints.

3.2.1. Temporal-Wise Attention

To recognize similarities and repetitions in human motion sequences, we apply a Temporal-wise Attention. This technique determines the attention score by matching the current frame with historical motion frames through a q u e r y and various k e y v a l u e pairs, allowing us to identify similar frames. The input historical sequence is divided into three parts: q u e r y , k e y , and v a l u e . For a sequence, there is only one q u e r y , the k e y and v a l u e slide on the sequence, allowing us to count the correlation for each segment. The q u e r y is the last M frames of the input, used to calculate the similarity with the sequence before it. In order to determine attention weights with the q u e r y , the k e y is also set to M consecutive frames, and it has T M N + 1 in the entire sequence. The q u e r y and k e y are, respectively, represented as S T M + 1 : T and { S i : i + M 1 } i = 1 T N M + 1 , mapped to the unified dimension using their respective mapping functions H Q (·) and H K (·), that is,
Q = H Q ( S T M + 1 : T )
K i = H K { S i : i + M + N 1 } i = 1 T N M + 1
The attention score between each frame of these two sequences is formulated as
a i = Q × K i i = 1 T N M + 1 Q × K i
where a i denotes the attention score. We utilize sum normalization and the ReLU [36] activation function to constrain the weight values within the range of 0 and 1.
For each k e y , there is a corresponding v a l u e . In this work, v a l u e contains the k e y and the subsequent N frames following it, with the length of N + M , denoted as { S i : i + M + N 1 } i = 1 T N M + 1 . The same as for the k e y , there is also T M N + 1 values in the entire sequence. Before weighting, we first apply DCT to transform it from the time domain to the frequency domain, and use the frequency-domain information V i to encode the features of the sequence. Then, the model yields the sequence S t e m by performing the dot product between V i and a i , that is,
S t e m = i = 1 T N M + 1 a i V i

3.2.2. Spatial-Wise Feature Extractor

GCN demonstrates exceptional adaptability to various poses and deformations in human motion due to its flexibility in handing graph-structured data. It effectively captures spatial relationships and connectivity patterns between human joints, which are essential for understanding the structure and semantics of human motion. Therefore, we utilize stacked GCN blocks to extract features only in the spatial dimension. Before this, we pad the q u e r y sequence S T M + 1 : T with N frames of the last frame and convert it to the frequency domain using DCT. Mathematically, A ( j ) R K × K represents the trainable graph adjacency matrix, where j denotes the number of the layer. At the j-th layer, the graph convolution is
G ( j + 1 ) = σ ( A ( j ) × G ( j ) × W ( j ) )
where σ (·) denotes the activation function t a n h (·), and W ( j ) denotes the trainable weights of the j-th layer, which effectively enforces limitations on feature values to maintain consistent predictions.
We subsequently concatenate the outputs S t e m and S s p a , then utilize fully connected layers to blend the spatial and temporal features, with the goal of enhancing the representation of human joint coordinates. After IDCT, the sequence returns to the time-domain space, and we take the last N + M frames S ^ as the output of the network.

3.3. Principle of SkDiff

Paralleling with the AST-DE method, the Dynamic Differential Dependencies Extractor empowers the effectiveness of the prediction. As for the building method SkDiff, we take frames with a length of T as our input. As the series is discontinuous, the N-th initial difference block D 0 N can simply be given by
D 0 N = X 2 : T N X 1 : T 1 N
However, the length of the generated series is T 1 , which does not correspond to the required input series length of our model. Thus, we developed two methods for difference series compensation: (a) Repeat the first frame at the beginning of the series, and (b) repeat the last frame at the end of the series. The former method focuses on the vitality of the recent motion, while the latter concentrates on stimulating the model capacity of eliminating inaccurate prediction. Based on the two methods, we can refine the series for difference feature extraction as D N :
D N = { [ x 1 , x 1 , x 2 , x 3 , . . . , x T 2 , x T 1 ] ( a ) [ x 1 , x 2 , x 3 , . . . , x T 2 , x T 1 , x T 1 ] ( b )
where x i represents the i-th frame in the original difference block D 0 N . Therefore, we can ensure that the length of the input D N is T, which means our difference extractor can grasp the feature from the difference block. Specifically, our difference extractor is named as SkDiff Extractor, whose structure is demonstrated in Figure 2. For corresponding to the data feature, our SkDiff Extractor can be flexibly alternated from the given models.
SkDiff can generate blocks of various difference depths, which facilitates the advantage that it can handle a range of circumstances varying in the extent of motion change. In Figure 1, we demonstrate that the 1-SkDiff block is generated by SkDiff performed on the original motion series, while the 2-SkDiff block is yielded by SkDiff performed on the 1-SkDiff block. The depths of SkDiff generation can be adjusted according to the motion characteristics in reality.

3.4. Loss

The model yields a predict sequence with a total length of M + N . For accuracy assessment, we measure the deviation between the 3D spatial coordinates of each human body joint as predicted by the model and the actual coordinates. To achieve this, we adopt the mean per-joint position error (MPJPE), which was initially suggested in [12] for the estimation of 3D joint positions. This metric quantifies the mean squared Euclidean distance for each joint’s predicted location in comparison with the ground truth location, and is expressed by the following equation:
L = 1 K k = 1 K | | J ^ k J k | | 2
where J ^ k represents the 3D coordinates of the k-th joint of the human pose, and J k denotes the ground truth.

4. Experimental Results and Datasets

4.1. Datasets

In order to evaluate the effectiveness of the proposed 2DHnet in the human pose estimation task, we utilize publicly available datasets as benchmarks, specifically Human3.6M (H3.6M) [37] and 3D Pose in the Wild (3DPW) [38].

4.1.1. Human3.6M

Human3.6M [37] stands as a pivotal and extensively utilized resource in the field of 3D human pose estimation and human motion prediction. Encompassing over 3.6 million images, this dataset was meticulously captured using the advanced Vicon motion capture system. It features seven actors performing a diverse array of 15 distinct activities, ranging from routine actions such as walking and smoking to more complex interactions like greeting and discussing.
Each pose within the dataset is anatomically defined by 32 joints, which are meticulously pre-processed into 3D coordinates. To facilitate a more consistent temporal resolution, the sequences have been uniformly down-sampled to 25 frames per second. This down-sampling aids in streamlining the data for computational processes while preserving the essential dynamics of human motion.
Following established research routines, we make use of subjects 1 (S1), S6, and S7–S9 for training, S11 for validation, and S5 for testing. This methodology aligns with the conventions observed in previous studies, thereby promoting standardization in evaluation metrics across the research community.

4.1.2. 3DPW

The 3D Pose in the Wild (3DPW) dataset [38] stands as a significant resource in the field of 3D human pose estimation, particularly known for its comprehensive capture of natural human movements in diverse and complex environments. It is one of the first datasets to offer precise 3D poses in outdoor scenarios, making it an essential benchmark for evaluating models in realistic settings. The dataset includes over 51,000 frames across 60 video sequences. Each human pose in 3DPW is articulated through 24 to 26 joints, with variations depending on specific evaluation protocols. The dataset has been divided into training, validation, and test sets. The high frame rate of 60 Hz further adds to its suitability for dynamic and real-time pose estimation tasks. Following a common practice adopted by most studies, we train our models over 50 epochs on the training set before testing on the designated test set.

4.2. Experimental Metric

To evaluate the performance of our model, which outputs 3D positions, we adopted the mean per-joint position error (MPJPE) metric on the 3D joints. The MPJPE is the most widely used metric for assessing errors in 3D pose estimation, as it measures the discrepancies in 3D coordinate space. For an integrated comparison with other methods, we adhered to the evaluation standard outlined in [12,13,39,40], applying a sampling rate of 25 frames per second across both datasets. The L2 distance between the predicted and ground truth positions of each joint was calculated and then averaged to quantify the error.

4.3. Experimental Settings

In order to implement the experiment, we made a series of parameter settings to enhance the model effectiveness. Accordingly, we performed experiments and obtained the most effective parameter configuration for our model.
Within the processing of both datasets, the input sequence length was established at 50 frames. For the Human3.6 dataset, the output was configured to 10 frames, whereas for the 3DPW dataset, it was set to 25 frames, for this dataset is more changeable and needs more future ground truths to generate more transitive future predictions. Regarding the feature representation of the human body, each joint is described by three-dimensional coordinates, represented as C = 3 × K . Specifically, the Human3.6 dataset employs 22 joints, resulting in a feature dimension of 66. We kept the same number of joints for 3DPW.
As is elaborated in Section 3.1, for the 2D-DE, we built a method which can yield the difference of the given series, while we padded one frame at the beginning of the generated series to ensure the same data dimensions. We performed this method P times and thus obtained P series, which in reality meant velocity, acceleration, and more slight changes of human motion. Then, we transformed the temporary skeleton series to frequency field data utilizing DCT. For feature extraction of each difference layer, we adopted the network of the multi-layer perceptron (MLP) block, where the latent layer size was 128. Then, these layers were concatenated and merged with another MLP block, where the latent layer size was 128 × P . Finally, the difference feature output was concatenated with the attention-based feature output, mixed with a 256-latent-sized MLP block, and transformed into temporary skeleton-based series as the ultimate prediction.
For feature extraction from each difference layer, we utilized a multi-layer perceptron (MLP) network, where the latent layer size was set to 128. The outputs of these layers were then concatenated and merged using another MLP block, with a latent layer size of 128 × P . The final difference features were combined with the attention-based features, passed through an MLP block with a latent size of 256, and ultimately converted back into a temporal skeleton-based sequence utilizing IDCT, providing the final prediction.
According to Section 3.2, the Temporal-wise Attention learning process is integrated with the attention mechanism-based motion feature abstracting method. Practically, the q u e r y and k e y lengths were fixed at 10 frames, denoted as M = 10 . The mappings H Q (·) and H K (·) project Q and K i into 256 dimensions for attention weight computation. The v a l u e length was configured to 20, meaning that the model predicted the most pertinent 20 frames based on the last 10 frames of the input sequence. During the padding phase, the final frame of the input sequence was replicated 10 times to generate a total of 20 frames. These frames were then concatenated with the preceding 20 frames, forming a 40-frame input for the GCNs.
In the learning phase of the Spatial-wise Feature Extractor, the dimensions of W were established as 256 × 256 , while A had dimensions of C × C . We applied 18 GCN blocks in a stacked arrangement, connected via residual structures to improve generalization performance and accelerate model concentration.
To train the network, the Adam optimizer was employed with a batch size of 32. The model, comprising 15.34 million parameters, was implemented using the PyTorch framework and trained over 50 epochs on one NVIDIA RTX 4070 GPU, with the initial learning rate set to 0.001.

4.4. Comparison with State-of-the-Art Methods

In this section, we systematically evaluate the performance of various human pose estimation methods, with a particular focus on comparing them to our newly proposed model. To ensure a comprehensive analysis, we report the results of our model across both short-term (0∼500 ms) and long-term (500∼1000 ms) prediction intervals, as these time frames provide insights into the model’s performance across different temporal horizons.
For the Human3.6M dataset, our model is trained using the past 50 frames to predict the subsequent 10 frames. Once the next 10 frames are predicted, these frames are iteratively fed back into the model as input to generate further predictions. This recursive approach allows for the evaluation of our model’s ability to maintain accuracy over multiple steps of prediction. We note that the SkDiff depth represents the number of iterations of SkDiff. Among the results of the experiments, we find that the trained model achieves the lowest MPJPE when the SkDiff depth is 1.
On the 3DPW dataset, a similar methodology is adopted, with the model utilizing the past 50 frames to predict the next 25 frames. This setting provides a broader time window for prediction, enabling a more thorough examination of the model’s robustness over longer-term forecasts.

4.4.1. Results on Human3.6M

The performance of our model on the Human3.6M dataset, as presented in Table 1 and Table 2, showcases its superiority in both short-term and long-term 3D motion prediction, measured by mean per-joint position error (MPJPE). Our model not only competes effectively with state-of-the-art methods like TFAN [17] and EqMotion [18] but also demonstrates distinct advantages in capturing both immediate and sustained motion trends.
Table 1 provides a comprehensive overview of the short-term prediction results across various actions on the Human3.6M dataset. Our model achieves notable accuracy, particularly excelling in the ‘posing’ and ‘discussion’ actions. For instance, our model records the lowest MPJPE for the ‘discussion’ action across all time intervals, indicating its precision in predicting subtle and repetitive movements. Similarly, for the ‘posing’ action, our model surpasses all competitors at shorter intervals, reducing MPJPE by 37.3% and 36.8% on 160 ms and 320 ms intervals, respectively, compared to TFAN [17]. This suggests that our model effectively captures complex, nuanced changes in action dynamics over time.
In terms of long-term predictions, as shown in Table 2, our model consistently outperforms others, achieving the best accuracy for both 560 ms and 1000 ms averages. Notably, when comparing the relative percentage reduction in the ‘posing’ action, our model reduces MPJPE by 29.8% compared to TFAN [17] at 560 ms. These results underscore the model’s ability to maintain high predictive accuracy over extended periods, capturing complex temporal dependencies that other models struggle with. Additionally, our model also delivers competitive results in actions such as ‘eating’, ‘directions’, and ‘walking together’, further validating its generalization capabilities across diverse motion categories.
The data from Table 1 and Table 2 confirm that the prediction error of our proposed method tends to decrease with time, which means it has the capacity to fix the generation error, in that our model leverages SkDiff to mitigate this degradation. The 2D-DE, in particular, excels at extracting detailed transition features from generated difference information of the time series. Therefore, it is able to identify unreasonable predictions and reduce their weight, resulting in more precise long-term predictions.
Furthermore, by integrating features from both the temporal and spatial dimensions, our model is adept at finding the correlation between temporal and spatial information, which is crucial for accurate predictions over mid-to-long-term periods (320–1000 ms). Among the state-of-the-art models, ours consistently achieves the highest average accuracy in these mid-to-long-term predictions, demonstrating its strong adaptability and robustness across various temporal scales. For overall evaluation, we consider taking the average MPJPE across all listed temporal scales, thus finding that our model’s average MPJPE is 4.7% lower than that of the second-best model April-GCN [14].

4.4.2. Results on 3DPW

Table 3 provides a comprehensive comparison of our model’s 3D spatial predictions on the 3DPW dataset, demonstrating its superior performance over existing state-of-the-art approaches across both short-term and long-term prediction intervals. Unlike other models, which often face difficulties in balancing performance across different temporal horizons, our model consistently achieves outstanding results across both short-term and long-term predictions using a unified framework.
A closer analysis of the data presented in Table 3 further highlights the significant advantages of our model in both short-term (200 ms) and long-term (1000 ms) predictions. Our approach reduces the MPJPE by 10.6% at the 200 ms interval and by 31.6% at the 1000 ms interval compared to the second-best model DPnet [42]. The long-term prediction result also proves that our model is able to decrease the generation error owing to 2D-DE.
These results underscore the effectiveness of our model in maintaining high predictive accuracy across varying databases, demonstrating its generalization in capturing complex 3D motion patterns in both man-made and natural scenarios. We also evaluate our model overall with the same method as in the analysis of the performance on Human3.6M, and it turns out that our model has an average MPJPE 26.6% lower than the second-best model DPnet [42].

4.5. Visualization Analysis

To further validate our model, we selected four typical poses from the prediction results for visualization. As illustrated in Figure 3, we compared our approach with HisRep [13], used as a baseline. It is clear that our model demonstrates better performance by comparing the ground truth similarity between the predictions yielded by HisRep and our model. Our model proves effective for slight motion changes, as it captures transition features obtained by 2D-DE. In the visualizations shown in Figure 3a,d, it is evident that the baseline tends to generate rapid changes in predictions regardless of the history of motion changing characteristics. In contrast, our model effectively identifies the motion transition features and naturally generates high-quality motions aligned with them.
Our method also excels in slight adjustments, as shown in Figure 3b,c. It captures subtle movements, with the skeletons transitioning smoothly and maintaining pace with the motion. It is shown that our model has the capability of correcting the generation error, which could occur during upper or lower body motion prediction.
From the visualization results, it is evident that our model can distinguish between different motion classes, allowing for more tailored responses to historical skeletal sequences. It is highlighted that our model is better at handling the smooth transition of motion, avoiding large differences between reality and prediction.

4.6. Ablation Study

We conduct ablation studies to evaluate the influence of the numbers of GCLs in AST-DE, different configurations, and the SkDiff Extractor latent feature size of our approach on the Human3.6M dataset.

4.6.1. Effect of the Number of GCLs in AST-DE

We separately tested the impact of GCN layers in the AST-DE on model performance when using MLP and Add as the SkDiff Mixer (Figure 4). The data demonstrate that the model achieves high predictive accuracy when the number of GCLs is set to 18. We believe this is because as the number of layers in the model increases, it can extract more spatial features of the human skeleton, which is beneficial for learning structural information. In addition, when using Add as a fusion method, the model performs better than using MLP. We speculate that addition increases the information content of features in the same dimension and adds detailed information.

4.6.2. Effect of Different Configurations

To better understand the impact of various configurations on the model’s performance, we conducted several experiments where we varied whether AST-DE and 2D-DE were used, the depth of SkDiff, the padding frame used in SkDiff, and the type of Mixer used for the 2D-DE and AST-DE. Our best results were achieved when both the AST-DE and the 2D-DE were present, the SkDiff depth was set to 1, the last frame was used for padding, and the Mixer was Add. The ablation results in Table 4 demonstrate the impact of different configurations on the model’s performance, and our analysis is presented below.
Effect of components: To examine the effect of the components, we first evaluated scenarios where the model used only one of the components. Subsequently, we tested configurations that included both components. Our findings are summarized as follows:
  • Enabling only the AST-DE significantly improves model performance, particularly across various time horizons. This suggests that the AST-DE effectively captures the correlation between temporal sequence attention features and spatial graphical features.
  • When the AST-DE is absent and the model relies solely on the 2D-DE for prediction, the performance deteriorates markedly. This indicates that the differential information alone causes a loss of skeletal data and fails to model complete and effective motion semantics.
  • Integrating the 2D-DE with the AST-DE enhances the model’s long-term prediction capabilities. As noted earlier, our approach is effective in correcting long-term errors using the 2D-DE. As shown in Table 4, the combined model outperforms the models using a single component, reducing MPJPE by 1.4 at 1000 ms.
Effect of SkDiff depth: As the SkDiff depth increases from 1 to 3, model performance varies across time horizons. Generally, a depth of 1 or 3 yields better results, while a depth of 2 slightly increases errors. This suggests that the 2-SkDiff model is more sensitive to noise, whereas the 3-SkDiff model has a higher capacity to mitigate noise and capture more detailed motion trends for prediction. This is supported by the experimental result where the depth is 3, the padding is “first”, and the mix method is “Add”.
Effect of padding frame position: The choice of padding frame position (“last” vs. “first”) also impacts model performance. As mentioned in Section 3.3, repeating the last frame may enhance the model’s ability to eliminate inaccurate predictions, while repeating the first frame emphasizes recent motion dynamics. The results indicate that using the “last” frame for padding generally leads to better performance than using the “first” frame, likely due to increased robustness against noise at the end of the SkDiff block. However, this trend is not always consistent, as the initial weights can also affect the outcome; hence, the “first” frame padding strategy may also yield competitive results.
Effect of SkDiff Mixer type: When comparing the use of MLP (multi-layer perceptron) and Add as the mix method for the two components, we observe that the MLP approach generally achieves higher MPJPE values for shorter time horizons (80 ms to 320 ms) but lower MPJPE errors for longer time horizons (560 ms to 1000 ms). This suggests that MLPs are more effective at learning complex nonlinear relationships but may suffer from overfitting, leading to less accurate long-term predictions. Consequently, we replaced the MLP with Add, which theoretically better compensates for the AST-DE, given their linear relationship. Although the best results were obtained with the Add configuration, we observed that the overall performance difference between the two Mixer types was relatively small. This finding suggests that a simple Add might not be sufficient to extract effective motion semantics.

4.6.3. Effect of the Number of MLP Channels in SKDiff Block

When we chose the Add as SkDiff Mixer to fuse features, we conducted ablation experiments on the number of channels in the MLP (Table 5). The experimental results show that the model’s prediction results are the most accurate when the number of channels is 128. We speculate that as the number of channels increases, the model overly focuses on detailed features, resulting in overfitting.

4.6.4. Effect of the Number of Channels in SkDiff Extractor and SkDiff Mixer

In the 2D-DE, we tested the number of channels C d for the SkDiff Extractor and SkDiff Mixer, experimenting with C d = 128, 256, and 512. As can be seen from Figure 5, when C d = 128 , the model achieves the best prediction performance across all time intervals. We believe this could be attributed to a balance between model complexity and computational efficiency. With too many channels, the model might become overly complex, which could result in overfitting to the training data or increased computational cost without a corresponding improvement in predictive accuracy.

5. Conclusions

This paper presents Dynamic Differencing-based Hybrid Networks (2DHnet) for improved human motion prediction. Traditional models and recent deep learning approaches often struggle to capture rapid changes in motion dynamics and optimize spatial–temporal features effectively. To address these challenges, 2DHnet introduces two key modules: the Dynamic Differential Dependencies Extractor (2D-DE), which captures dynamic features such as velocity and acceleration; and the Attention-based Spatial–Temporal Dependencies Extractor (AST-DE), which enhances spatial–temporal correlations. Our 2DHnet provides a more comprehensive representation of human motion by combining these modules in a dual-branch architecture. Experimental results on the Human3.6M and 3DPW datasets demonstrate that 2DHnet achieves significant improvements over state-of-the-art methods, with average MPJPE reductions of 4.7% and 26.6%, respectively. Even though this work highlights the importance of motion dynamics learning, 2DHnet may lack transparency in its decision-making processes. Understanding why the model makes certain predictions can be challenging, which may hinder its acceptance in applications requiring high levels of interpretability. Addressing these limitations in future work could involve simplifying the model architecture and improving data augmentation techniques.

Author Contributions

Methodology, J.Z. and R.J.; software, R.J. and C.L.; data curation, R.J. and C.L.; visualization, C.L. and J.Z.; writing—original draft preparation, R.J. and C.L.; writing—review and editing, J.Z.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets used in this research are available open source at the following links: Human3.6M dataset at http://vision.imar.ro/human3.6m (accessed on 20 September 2024); 3DPW dataset at https://virtualhumans.mpi-inf.mpg.de/3DPW/license.html (accessed on 20 September 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ghafir, I.; Prenosil, V.; Svoboda, J.; Hammoudeh, M. A survey on network security monitoring systems. In Proceedings of the 2016 IEEE 4th International Conference on Future Internet of Things and Cloud Workshops (FiCloudW), Vienna, Austria, 22–24 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 77–82. [Google Scholar]
  2. MacKenzie, I.S. Human-Computer Interaction: An Empirical Research Perspective; Morgan Kaufmann: Cambridge, MA, USA, 2012. [Google Scholar]
  3. Weinland, D.; Ronfard, R.; Boyer, E. A survey of vision-based methods for action representation, segmentation and recognition. Comput. Vis. Image Underst. 2011, 115, 224–241. [Google Scholar] [CrossRef]
  4. Wang, J.M.; Fleet, D.J.; Hertzmann, A. Gaussian process dynamical models for human motion. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 30, 283–298. [Google Scholar] [CrossRef] [PubMed]
  5. Brand, M.; Hertzmann, A. Style machines. In Proceedings of the 27th Annual Conference on Computer Graphics and Interactive Techniques, New Orleans, LA, USA, 23–28 July 2000; pp. 183–192. [Google Scholar]
  6. Zhong, J.; Cao, W. Geometric algebra-based multiscale encoder-decoder networks for 3D motion prediction. Appl. Intell. 2023, 53, 26967–26987. [Google Scholar] [CrossRef]
  7. Fragkiadaki, K.; Levine, S.; Felsen, P.; Malik, J. Recurrent Network Models for Human Dynamics. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 4346–4354. [Google Scholar]
  8. Jain, A.; Zamir, A.R.; Savarese, S.; Saxena, A. Structural-rnn: Deep learning on spatio-temporal graphs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5308–5317. [Google Scholar]
  9. Martinez, J.; Black, M.J.; Romero, J. On human motion prediction using recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2891–2900. [Google Scholar]
  10. Shi, J.; Zhong, J.; Cao, W. Multi-semantics Aggregation Network based on the Dynamic-attention Mechanism for 3D Human Motion Prediction. IEEE Trans. Multimed. 2023, 26, 5194–5206. [Google Scholar] [CrossRef]
  11. Li, M.; Chen, S.; Zhao, Y.; Zhang, Y.; Wang, Y.; Tian, Q. Dynamic multiscale graph neural networks for 3d skeleton based human motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 214–223. [Google Scholar]
  12. Mao, W.; Liu, M.; Salzmann, M.; Li, H. Learning Trajectory Dependencies for Human Motion Prediction. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Los Alamitos, CA, USA, 27 October–2 November 2019; pp. 9488–9496. [Google Scholar]
  13. Wei, M.; Miaomiao, L.; Mathieu, S. History Repeats Itself: Human Motion Prediction via Motion Attention. In Proceedings of the Europe Conference on Computer Vision ECCV, Online, 23–28 August 2020. [Google Scholar]
  14. Gu, B.; Tang, J.; Ding, R.; Liu, X.; Yin, J.; Zhang, Z. April-GCN: Adjacency Position-velocity Relationship Interaction Learning GCN for Human motion prediction. Knowl.-Based Syst. 2024, 292, 111613. [Google Scholar] [CrossRef]
  15. Zhong, J.; Cao, W. Geometric algebra-based multiview interaction networks for 3D human motion prediction. Pattern Recognit. 2023, 138, 109427. [Google Scholar] [CrossRef]
  16. Cao, W.; Li, S.; Zhong, J. A dual attention model based on probabilistically mask for 3D human motion prediction. Neurocomputing 2022, 493, 106–118. [Google Scholar] [CrossRef]
  17. Du, X.; Wang, Y.; Li, Z.; Yan, S.; Liu, M. TFAN: Twin-Flow Axis Normalization for Human Motion Prediction. IEEE Signal Process. Lett. 2024, 31, 486–490. [Google Scholar] [CrossRef]
  18. Xu, C.; Tan, R.T.; Tan, Y.; Chen, S.; Wang, Y.G.; Wang, X.; Wang, Y. Eqmotion: Equivariant multi-agent motion prediction with invariant interaction reasoning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1410–1420. [Google Scholar]
  19. Guo, W.; Du, Y.; Shen, X.; Lepetit, V.; Alameda-Pineda, X.; Moreno-Noguer, F. Back to mlp: A simple baseline for human motion prediction. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 4809–4819. [Google Scholar]
  20. Li, C.; Zhang, Z.; Lee, W.S.; Lee, G.H. Convolutional Sequence to Sequence Model for Human Dynamics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  21. Li, S.; Li, W.; Cook, C.; Zhu, C.; Gao, Y. Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  22. Liu, J.; Wang, G.; Duan, L.Y.; Abdiyeva, K.; Kot, A.C. Skeleton-Based Human Action Recognition with Global Context-Aware Attention LSTM Networks. IEEE Trans. Image Process. 2018, 27, 1586–1599. [Google Scholar] [CrossRef] [PubMed]
  23. Tang, J.; Wang, J.; Hu, J.F. Predicting human poses via recurrent attention network. Vis. Intell. 2023, 1, 18. [Google Scholar] [CrossRef]
  24. Zhong, C.; Hu, L.; Zhang, Z.; Ye, Y.; Xia, S. Spatial-Temporal Gating-Adjacency GCN for Human Motion Prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  25. He, Z.; Zhang, L.; Wang, H. An initial prediction and fine-tuning model based on improving GCN for 3D human motion prediction. Front. Comput. Neurosci. 2023, 17, 1145209. [Google Scholar]
  26. Fu, J.; Yang, F.; Dang, Y.; Liu, X.; Yin, J. Learning Constrained Dynamic Correlations in Spatiotemporal Graphs for Motion Prediction. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 14273–14287. [Google Scholar] [CrossRef] [PubMed]
  27. Martínez-González, A.; Villamizar, M.; Odobez, J.M. Pose Transformers (POTR): Human Motion Prediction with Non-Autoregressive Transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  28. Mi, L.; Ding, R.; Zhang, X. Skeleton-based human motion prediction via spatio and position encoding transformer network. In Proceedings of the International Conference on Artificial Intelligence, Virtual Reality, and Visualization (AIVRV 2022), Chongqing, China, 2–4 September 2022; SPIE: Bellingham, WA, USA, 2023; Volume 12588, pp. 186–191. [Google Scholar]
  29. Zhao, M.; Tang, H.; Xie, P.; Dai, S.; Sebe, N.; Wang, W. Bidirectional transformer gan for long-term human motion prediction. ACM Trans. Multimed. Comput. Commun. Appl. 2023, 19, 163. [Google Scholar] [CrossRef]
  30. Meneses, M.; Matos, L.; Prado, B.; Carvalho, A.; Macedo, H. SmartSORT: An MLP-based method for tracking multiple objects in real-time. J. -Real-Time Image Process. 2021, 18, 913–921. [Google Scholar] [CrossRef]
  31. Cao, G.; Huang, W.; Lan, X.; Zhang, J.; Jiang, D.; Wang, Y. MLP-DINO: Category Modeling and Query Graphing with Deep MLP for Object Detection. In Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence (IJCAI-24), Jeju, Republic of Korea, 3–9 August 2024. [Google Scholar]
  32. Chen, S.; Xie, E.; Ge, C.; Chen, R.; Liang, D.; Luo, P. Cyclemlp: A mlp-like architecture for dense visual predictions. IEEE Trans. Pattern Anal. Mach. Intell. 2021. Volume abs/2107.10224. Available online: https://arxiv.org/abs/2107.10224 (accessed on 6 December 2024).
  33. Boughrara, H.; Chtourou, M.; Amar, C.B. MLP neural network based face recognition system using constructive training algorithm. In Proceedings of the International Conference on Multimedia Computing & Systems, Tangiers, Morocco, 10–12 May 2012. [Google Scholar]
  34. Shahreza, H.O.; Hahn, V.K.; Marcel, S. MLP-Hash: Protecting Face Templates via Hashing of Randomized Multi-Layer Perceptron. arXiv 2022, arXiv:2204.11054. [Google Scholar]
  35. Bouazizi, A.; Holzbock, A.; Kressel, U.; Dietmayer, K.; Belagiannis, V. MotionMixer: MLP-based 3D Human Body Pose Forecasting. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22. International Joint Conferences on Artificial Intelligence Organization, Vienna, Austria, 23–29 July 2022; pp. 791–798. [Google Scholar]
  36. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  37. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  38. Von Marcard, T.; Henschel, R.; Black, M.J.; Rosenhahn, B.; Pons-Moll, G. Recovering accurate 3d human pose in the wild using imus and a moving camera. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 601–617. [Google Scholar]
  39. Dang, L.; Nie, Y.; Long, C.; Zhang, Q.; Li, G. Msr-gcn: Multi-scale residual graph convolution networks for human motion prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 11467–11476. [Google Scholar]
  40. Ma, T.; Nie, Y.; Long, C.; Zhang, Q.; Li, G. Progressively generating better initial guesses towards next stages for high-quality human motion prediction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6437–6446. [Google Scholar]
  41. Li, M.; Chen, S.; Zhang, Z.; Xie, L.; Tian, Q.; Zhang, Y. Skeleton-parted graph scattering networks for 3d human motion prediction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 18–36. [Google Scholar]
  42. Tang, J.; Zhang, J.; Ding, R.; Gu, B.; Yin, J. Collaborative multi-dynamic pattern modeling for human motion prediction. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3689–3700. [Google Scholar] [CrossRef]
Figure 1. Network architecture of our proposed method. The entire network structure is divided into two components: Dependencies Extractor (AST-DE) in pink blocks and Dynamic Differential Dependencies Extractor (2D-DE) in yellow block. Specifically, AST-DE is the main branches including Temporal-wise Attention and Spatial-wise Feature Extractor, while 2D-DE is a differential branch, which is used to extract differential information from sequences and merge it with features extracted from other branches. Temporal-wise Attention extracts temporal features of the sequence, dividing the input historical sequence into three parts. The green parentheses represent the q u e r y , while the red and blue parentheses represent the k e y and v a l u e , respectively. Spatial-wise Feature Extractor is used to extract features about the human body in spatial dimensions, mainly using graph convolutional networks.
Figure 1. Network architecture of our proposed method. The entire network structure is divided into two components: Dependencies Extractor (AST-DE) in pink blocks and Dynamic Differential Dependencies Extractor (2D-DE) in yellow block. Specifically, AST-DE is the main branches including Temporal-wise Attention and Spatial-wise Feature Extractor, while 2D-DE is a differential branch, which is used to extract differential information from sequences and merge it with features extracted from other branches. Temporal-wise Attention extracts temporal features of the sequence, dividing the input historical sequence into three parts. The green parentheses represent the q u e r y , while the red and blue parentheses represent the k e y and v a l u e , respectively. Spatial-wise Feature Extractor is used to extract features about the human body in spatial dimensions, mainly using graph convolutional networks.
Ai 05 00139 g001
Figure 2. The structure of SkDiff Extractor and Mixer in 2D-DE. SkDiff Extractors are based on multi-layer perceptron and consist of linear layers and an activation function (ReLU). Mixer is dynamic, corresponding to the data feature, which is chosen from concatenating the input data together before fusing features or adding all results of SkDiff Extractors.
Figure 2. The structure of SkDiff Extractor and Mixer in 2D-DE. SkDiff Extractors are based on multi-layer perceptron and consist of linear layers and an activation function (ReLU). Mixer is dynamic, corresponding to the data feature, which is chosen from concatenating the input data together before fusing features or adding all results of SkDiff Extractors.
Ai 05 00139 g002
Figure 3. Visualization results of 2DHnet. The visual demonstration includes four common motions: (a) motion of ‘discussion’, (b) motion of ‘eating’, (c) motion of ‘waiting’, and (d) motion of ‘posing’, all from the H3.6M dataset. The labels “GT”, “HisRep”, and “Ours” represent the ground truth, predictions from the HisRep method, and predictions from our approach, respectively. The predicted motion is highlighted with pink and blue, while the ground truth is colored with gray and lilac. We mix the prediction and ground truth together for better comparison, and it is clear that our method provides more accurate predictions, effectively capturing the nuances in human motion series.
Figure 3. Visualization results of 2DHnet. The visual demonstration includes four common motions: (a) motion of ‘discussion’, (b) motion of ‘eating’, (c) motion of ‘waiting’, and (d) motion of ‘posing’, all from the H3.6M dataset. The labels “GT”, “HisRep”, and “Ours” represent the ground truth, predictions from the HisRep method, and predictions from our approach, respectively. The predicted motion is highlighted with pink and blue, while the ground truth is colored with gray and lilac. We mix the prediction and ground truth together for better comparison, and it is clear that our method provides more accurate predictions, effectively capturing the nuances in human motion series.
Ai 05 00139 g003
Figure 4. Ablation tests on the number of GCLs in AST-DE.
Figure 4. Ablation tests on the number of GCLs in AST-DE.
Ai 05 00139 g004
Figure 5. Ablation tests on the number of channels in SkDiff Extractor and SkDiff Mixer.
Figure 5. Ablation tests on the number of channels in SkDiff Extractor and SkDiff Mixer.
Ai 05 00139 g005
Table 1. Short-term results on the 15 actions of H3.6M dataset. The average prediction results for all the actions are added at the end. The best results are highlighted in bold.
Table 1. Short-term results on the 15 actions of H3.6M dataset. The average prediction results for all the actions are added at the end. The best results are highlighted in bold.
MotionWalkingEatingSmokingDiscussion
Milliseconds 80 160 320 400 80 160 320 400 80 160 320 400 80 160 320 400
PGBIG [40]10.219.834.540.37.015.130.638.16.614.128.234.710.023.853.666.7
SPGSN [41]10.119.434.841.57.114.930.537.96.713.828.034.610.423.853.667.1
EqMotion [18]9.218.234.241.46.614.230.037.76.212.826.733.69.122.051.865.3
SIMLPE [19]9.9--39.65.9--36.16.5--36.39.4--64.3
TFAN [17]10.019.936.543.15.913.628.435.96.614.329.836.89.121.950.063.4
April-GCN [14]9.218.734.441.16.214.130.037.75.813.127.033.78.922.652.865.8
Ours9.518.933.038.15.713.327.935.25.912.323.930.18.217.734.143.4
MotionDirectionsGreetingPhoningPosing
Milliseconds80160320400801603204008016032040080160320400
PGBIG [40]7.217.640.951.515.234.171.687.18.318.338.748.410.725.760.076.6
SPGSN [41]7.417.239.850.314.632.670.686.48.718.338.748.510.725.359.976.5
EqMotion [18]6.615.839.050.012.730.069.186.37.717.338.248.49.323.759.677.5
SIMLPE [19]6.5--55.812.4--77.38.1--48.68.8--73.8
TFAN [17]6.517.143.154.912.328.462.476.97.917.538.348.38.622.155.472.0
April-GCN [14]6.216.440.150.813.032.071.787.87.417.338.047.89.023.759.577.0
Ours6.617.843.655.512.329.262.977.87.817.938.448.27.316.934.745.5
MotionPurchasesSittingSitting DownTaking Photo
Milliseconds80160320400801603204008016032040080160320400
PGBIG [40]12.528.760.173.38.819.242.453.813.927.957.471.58.418.942.053.3
SPGSN [41]12.828.661.074.49.319.442.353.614.227.756.870.78.818.941.552.7
EqMotion [18]11.326.659.874.18.318.141.153.013.226.356.070.48.117.841.052.7
SIMLPE [19]11.7--72.48.6--55.213.6--70.87.8--50.8
TFAN [17]11.727.659.073.18.719.242.854.413.629.157.570.37.717.539.750.5
April-GCN [14]11.026.859.273.08.018.141.753.313.026.756.570.87.818.041.552.9
Ours12.129.061.576.18.419.042.353.714.030.259.372.27.817.940.351.4
MotionWaitingWalking DogWalking TogetherAverage
Milliseconds80160320400801603204008016032040080160320400
PGBIG [40]8.920.143.654.318.839.373.786.48.718.634.441.010.322.747.458.5
SPGSN [41]9.219.843.154.1----8.918.233.840.910.422.347.158.3
EqMotion [18]7.917.941.553.116.635.972.085.78.117.132.439.29.420.946.257.9
SIMLPE [19]7.8--53.218.2--83.68.4--41.29.6--57.3
TFAN [17]7.617.640.851.718.038.171.084.08.418.035.743.79.521.546.057.3
April-GCN [14]7.718.441.652.316.737.071.384.08.017.633.240.09.221.446.657.8
Ours7.618.442.153.518.338.872.785.77.917.033.040.09.320.143.353.8
Table 2. Long-term prediction across 15 actions and average prediction results of the H3.6M dataset. The best results are highlighted in bold.
Table 2. Long-term prediction across 15 actions and average prediction results of the H3.6M dataset. The best results are highlighted in bold.
MotionWalkingEatingSmokingDiscussionDirectionsGreetingPhoningPosing
Milliseconds 560 1000 560 1000 560 1000 560 1000 560 1000 560 1000 560 1000 560 1000
PGBIG [40]48.156.451.176.046.569.587.1118.269.3100.4110.2143.565.9102.7106.1164.8
SPGSN [41]46.953.649.873.446.768.6--70.1100.5--66.7102.5--
EqMotion [18]50.758.850.776.245.467.387.1116.869.8101.2111.6144.265.8102.3111.2169.6
SIMLPE [19]46.855.749.674.547.269.385.7116.373.1106.799.8137.566.3103.3103.4168.7
TFAN [17]53.064.449.576.748.672.484.7116.372.4107.2100.9138.566.7105.5102.1166.8
April-GCN [14]49.454.850.674.646.568.888.0117.770.3100.0111.0142.665.5100.3109.7166.4
Ours46.860.748.874.042.365.566.6106.273.3104.5103.0140.966.1104.572.4137.8
MotionPurchasesSittingSitting DownTaking PhotoWaitingWalking DogWalking TogetherAverage
Milliseconds56010005601000560100056010005601000560100056010005601000
PGBIG [40]95.3133.374.4116.196.7147.874.3118.672.2103.4104.7139.851.964.376.9110.3
SPGSN [41]--75.0116.2--75.6118.273.5103.6----77.4109.6
EqMotion [18]97.4136.974.2116.096.9148.977.7122.473.5105.8104.4142.150.462.077.8111.4
SIMLPE [19]93.8132.575.4114.195.7142.471.0112.871.6104.6105.6141.250.861.575.7109.4
TFAN [17]96.1136.175.4114.095.2141.871.2113.970.4105.0107.1146.555.270.376.5111.7
April-GCN [14]96.8135.475.4117.298.7149.375.9119.172.3103.6103.2137.150.561.177.6109.9
Ours98.6136.974.5115.296.9143.372.7115.272.9105.7108.2143.951.065.572.9108.0
Table 3. Comparison of prediction results on the 3DPW dataset. The best results are highlighted in bold.
Table 3. Comparison of prediction results on the 3DPW dataset. The best results are highlighted in bold.
Milliseconds2004006008001000
DMGNN [11]37.367.894.5109.7123.6
LTD [12]35.667.890.6106.9117.8
MSR [39]37.871.393.9110.8121.5
PGBIG [40]35.367.889.6102.6109.4
DPnet [42]29.358.379.894.4104.1
April-GCN [14]30.461.888.098.2105.4
Ours26.244.759.566.971.2
Table 4. Ablation results on the existence of AST-DE or 2D-DE, SkDiff depth, the frame used in SkDiff padding and the Mixer type of 2D-DE and AST-DE. The best results of the ablation group are highlighted in bold. We also evaluate the effect TSAtt existence, which is demonstrated with the symbol in the table.
Table 4. Ablation results on the existence of AST-DE or 2D-DE, SkDiff depth, the frame used in SkDiff padding and the Mixer type of 2D-DE and AST-DE. The best results of the ablation group are highlighted in bold. We also evaluate the effect TSAtt existence, which is demonstrated with the symbol in the table.
TSAttDepthPaddingMixer80 ms160 ms320 ms400 ms560 ms1000 ms
---9.320.943.654.173.6109.7
1LastMLP362.2363.1367.1369.3550.2695.1
1LastMLP9.421.043.754.273.3108.3
2LastMLP9.721.845.456.175.4111.0
3LastMLP9.421.244.154.573.4108.1
1FirstMLP9.721.444.254.774.1109.5
2FirstMLP9.521.444.355.074.3109.8
3FirstMLP9.621.444.755.474.7109.6
1LastAdd9.320.943.353.872.9108.0
2LastAdd9.621.644.855.374.3110.5
3LastAdd9.521.544.354.873.7109.7
1FirstAdd9.621.644.555.073.9110.1
2FirstAdd9.521.544.455.074.3111.1
3FirstAdd9.421.243.753.972.8107.2
Table 5. Ablation tests on the number of MLP channels in SkDiff block when using Add Mixer. The best results in the ablation study group are highlighted in bold.
Table 5. Ablation tests on the number of MLP channels in SkDiff block when using Add Mixer. The best results in the ablation study group are highlighted in bold.
Number of Channels80 ms160 ms320 ms400 ms560 ms1000 ms
1289.320.943.353.872.9108.0
2569.521.444.154.573.4109.6
5129.621.644.755.474.6110.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ji, R.; Lu, C.; Zhong, J. Dynamic Differencing-Based Hybrid Network for Improved 3D Skeleton-Based Motion Prediction. AI 2024, 5, 2897-2913. https://doi.org/10.3390/ai5040139

AMA Style

Ji R, Lu C, Zhong J. Dynamic Differencing-Based Hybrid Network for Improved 3D Skeleton-Based Motion Prediction. AI. 2024; 5(4):2897-2913. https://doi.org/10.3390/ai5040139

Chicago/Turabian Style

Ji, Ruiya, Chengjie Lu, and Jianqi Zhong. 2024. "Dynamic Differencing-Based Hybrid Network for Improved 3D Skeleton-Based Motion Prediction" AI 5, no. 4: 2897-2913. https://doi.org/10.3390/ai5040139

APA Style

Ji, R., Lu, C., & Zhong, J. (2024). Dynamic Differencing-Based Hybrid Network for Improved 3D Skeleton-Based Motion Prediction. AI, 5(4), 2897-2913. https://doi.org/10.3390/ai5040139

Article Metrics

Back to TopTop