Next Article in Journal
Improvement in Strain Sensor Stability by Adapting the Metal Contact Layer
Previous Article in Journal
Analysis of the Impact of Detection Threshold Adjustments and Noise Uncertainty on Energy Detection Performance in MIMO-OFDM Cognitive Radio Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MEMe: A Mutually Enhanced Modeling Method for Efficient and Effective Human Pose Estimation

1
Key Laboratory of Optical Engineering, Chinese Academy of Sciences, Chengdu 610209, China
2
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
3
University of Chinese Academy of Sciences, Beijing 100039, China
4
School of Information and Communication Engineering, University of Electronic Science and Technology of China, Chengdu 610209, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(2), 632; https://doi.org/10.3390/s22020632
Submission received: 20 December 2021 / Revised: 6 January 2022 / Accepted: 11 January 2022 / Published: 14 January 2022
(This article belongs to the Topic Artificial Intelligence in Sensors)

Abstract

:
In this paper, a mutually enhanced modeling method (MEMe) is presented for human pose estimation, which focuses on enhancing lightweight model performance, but with low complexity. To obtain higher accuracy, a traditional model scale is largely expanded with heavy deployment difficulties. However, for a more lightweight model, there is a large performance gap compared to the former; thus, an urgent need for a way to fill it. Therefore, we propose a MEMe to reconstruct a lightweight baseline model, EffBase transferred intuitively from EfficientDet, into the efficient and effective pose (EEffPose) net, which contains three mutually enhanced modules: the Enhanced EffNet (EEffNet) backbone, the total fusion neck (TFNeck), and the final attention head (FAHead). Extensive experiments on COCO and MPII benchmarks show that our MEMe-based models reach state-of-the-art performances, with limited parameters. Specifically, in the same conditions, our EEffPose-P0 with 256 × 192 can use only 8.98 M parameters to achieve 75.4 AP on the COCO val set, which outperforms HRNet-W48, but with only 14% of its parameters.

1. Introduction

Since 2016, deep learning-based methods [1,2] have become a prime focus of research in 2D human pose estimation, greatly promoting the development of action recognition [3] and other human-centered applications [4,5]. Those deep learning models can be categorized as large models with high performances and small models with low accuracies, leading to a performance gap in Figure 1. To fill this gap between the complicated and lightweight models, this paper explores a general modeling method used to make the lightweight models “cross the gap”, i.e., with better performances than the big ones.
Traditionally, to overcome the challenges in the scale variances and keypoint occlusions, various classic large models are proposed, such as stacked hourglass [6], CPN [7], SimpleBaseline [8], and HRNet [9]. Stacked hourglass consists of multiple stacked hourglass-shaped modules with intermediate supervision, which is the first multi-scale representation network architecture in human pose estimation, but is complex and inefficient. To solve this problem, CPN cascades only two pyramid nets with ResNet as the backbone, where one is a global net and the other is a refine net, to make better auxiliary supervision. SimpleBaseline directly proposes a single-stage hourglass, where the features are downsampled by the ResNet to encode information, and then upsampled by deconvolution to decode the final output. As for HRNet, it is designed to maintain high-resolution representation by multi-scale parallel branches, which is more efficient and effective than ever, but still has largely redundant parameters and high complexity, and is not suitable for real deployment. In short, large models still need to be more efficient.
Recently, to gain a lightweight model with high accuracy, the main logic has been to use the single-stage architecture, and follow efficient models, such as MobileNet [10,11,12] and ShuffleNet [13,14]. A MobileNet-based method [15] generally upsamples the original convolution layer resolution and adds skip connections, but cannot remain a high-resolution representation, which could lead to poor performance. ShuffleNet-based methods, such as LiteHRNet [16], use the channel split–shuffle operation to reduce model complexity, which loses representation flexibility and model performance. In addition, the latest DANet [17] uses well-designed blocks with orthogonal attention and some special post-processing techniques to achieve the state-of-art among small models, but it still cannot cross the performance gap between large models and small models in Figure 1. Thus, small models need to be more effective.
In this paper, we propose a mutually enhanced modeling method (MEMe) to mutually enhance a lightweight model performance, but with low complexity. We transferred EfficientDet (EffDet) [18] into a human pose estimation to form an efficient human pose baseline named EffBase, by combining the following three modules: EfficientNet (EffNet), the original backbone of EffDet; BiFPNNeck, the original neck of EffDet; and final sum head (FSHead) commonly used for human pose estimation. Then, we used MEMe to reconstruct it into an efficient and effective pose net (EEffPose), as shown in Figure 2. Similarly, it contains three mutually enhanced modules: the MEMe-based backbone, efficient and effective net (EEffNet); the MEMe-based neck, total fusion neck (TFNeck); and the MEMe-based head, final attention head (FAHead).
For the backbone, EffNet is mutually enhanced into EEffNet, where the MEMe means the extended convolution and cross-scale fusion. In this way, it can extract more diverse multi-scale features, mutually merging cross-scale information in a more flexible mechanism than before, avoiding the limited and isolated multi-scale feature representations.
For the neck, BiFPNNeck is mutually enhanced into TFNeck by enhancing the repeating BiFPN blocks into total fusion blocks, where a single total fusion block includes three operations: AddUp, MixUp, and FixUp. It is a more sufficient and flexible feature fusion structure that can extract more powerful features and avoid weight-unbalanced problems compared with unidirectional CPN and bidirectional BiFPN.
For the head, FSHead is mutually enhanced into FAHead by using point-wise multiplication instead of sum fusion after the nearest upsampling. Specifically, the FAHead can be regarded as a cross attention mechanism where multi-scale outputs use point-wise multiplication to query and calibrate every corresponding pixels after upsampling instead of gathering together, avoiding the error accumulation due to nearest upsampling and the invalid fusion of information in different scales.
Our contributions are manifold:
  • We propose a MEME-based backbone (EEffNet) mutually enhanced by extended convolutions and cross-scale fusions, which bring a comprehensive multi-scale structure to extract more powerful features.
  • We propose a MEME-based neck (TFNeck) mutually enhanced by repeating several total fusion blocks containing AddUp, MixUp, and FixUp operations for all scales, which ensure the sufficient, flexible, and weight-balanced information flow.
  • We propose a MEME-based head (FAHead) mutually enhanced by Sigmoid and point-wise multiplication to query and calibrate multi-scale outputs, which take full advantage of the multi-branch structure and improve the final output.
  • By using MEMe, our EffBase crosses the performance gap and transforms into an efficient and effective pose net (EEffPose), which is superior, with less complexity than state-of-the-art (COCO [19] and MPII [20]) benchmarks. As shown in Figure 1, our EEffPose-P0 with 256 × 192 can use only 8.98 M parameters to achieve 75.4 AP on the COCO-val set, which outperforms HRNet-W48, but with only 14% of the cost.

2. Related Works

2.1. Efficient Backbone

Backbones like VGG, Hourglass, ResNet and HRNet are always used in human pose estimation. As for VGG backbone, OpenPose [21,22] uses VGG-18 to extract features and PAF to combine body keypoints. As for Hourglass backbone, Stacked Hourglass [6] propose a hourglass-shape encoder-decoder, Hourglass, to get heatmap output, which is used widely as common backbone such as in the AlphaPose [23]. As for ResNet backbone, [24] and CPN [7] use ResNet as backbone and decode the heatmap as the Hourglass while SimpleBaseline [8] uses several deconvolution layers firstly after the ResNet backbone to recover the output. As for HRNet backbone, a multi-scale but high-resolution representation is introduced by [9] firstly and then designed efficiently by [25].
However, to get a lightweight and efficient backbone, some efficient modeling methods were introduced in MobileNet [10,11,12], ShuffleNet [13,14], and EfficientNet [26]. Based on MobileNet’s depth-wise convolutions, DANet [17] proposes the OAB and SFU block to realize an efficient pose model. While using the channel shuffle operation in ShuffleNet, LiteHRNet [16] designs a lite block for a multi-scale model, reducing the complexity, but with a lot of loss in accuracy. EfficientNet—as an efficient backbone—is directly used in EfficientPose [27] for backbone, but fails to get acceptable accuracy.

2.2. Feature Fusion Neck

The feature pyramid network (FPN) [28] is the basic top–down information fusion neck, including many variants: PANet [29] adds another bottom–up information-flow layer; NAS-FPN [30] automatically generates the connection relationship among nodes; BiFPN [18] reduces some redundant connection and retains the top–down then bottom–up fusion path to fuse multi-scale features bidirectionally.
In 2D human pose estimation tasks, FPN-like fusion necks are the most common ones where CPN [7], MSPN [31], and Hourglass [6] fuse the low-level features and high-level features together, but only by an unidirectional information flow. There is no FPN-like fusion neck in HRNet [9] leading to the unbalanced feature extraction and the insufficient information fusion.

2.3. Final Output Head for Pose Estimation

In general, there are two categories of the final output head for pose estimation: regression head and heatmap head. The former directly regresses the mapping between the input image and coordination of joints [32]. The latter aims to recognize the joints at every pixel of the heatmap [33], which is widely used because of better results than the former. For example, the final sum head in HRNet is a heatmap head used to gather the multi-scale heatmaps as the final output.
Attention mechanisms in transformers [34] are now becoming more popular in a variety of fields, such as classification, objection detection, and target tracking [35]. Among them, DETR [36] reconstructs original feature information by using Q,K,V in detection, first, which is closely related to the field of pose estimation. Affected by this, TFPose [37] uses a transformer while PRCT [38] uses cascaded transformers as the attention regression head to detect the keypoints, but still cannot get a better result, with more parameters and complexities than HRNet. As for the attention heatmap head, polarized self-attention [39] uses the polarized mechanism to get the polarized channel and spatial attention to refine the output, while leading to the increase of complexity.

3. Methodology

We first transfered the EfficientDet [18] into an efficient baseline (EffBase) for human pose estimation, which is made up of original EffNet backbone, original BiFPN neck, and final sum head (FSHead). Then we used the mutually enhanced modeling method (MEMe) to enhance it into a brand-new efficient and effective pose (EEffPose) net with the enhanced EffNet (EEffNet), total fusion neck (TFNeck), and final attention head (FAHead), as shown in Figure 2. We presume that our method can succeed in improving the performance of the lightweight model, with low complexity. The formula is shown below:
E E f f P o s e ( · ) = M E M e ( E f f B a s e ( · ) ) E f f B a s e ( · ) = F S H e a d ( B i F P N ( E f f N e t ( · ) ) ) E E f f P o s e ( · ) = F A H e a d ( T F N e c k ( E E f f N e t ( · ) ) )
where EffNet, BiFPN and FSHead are the original backbone, neck, and head of EffBase, respectively. Meanwhile, EEffNet, TFNeck. and FAHead are the MEMe-based backbone, neck, and head of EEffPose, individually.

3.1. Enhanced EfficientNet

To make the optimal trade-off between the performance and model complexity, EffNet [26] is introduced as the original backbone, which is a high-efficiency model using the compound expansion method and neural architecture search technology to expand the model’s width, depth, and resolution. It is constructed efficiently by MobileBlock with outstanding performance and low model complexity. However, it cannot be directly used as the backbone here as multi-scale and high-resolution features are critical for human pose estimation.
For the backbone, we apply the MEMe to enhance the EfficientNet by adding extended convolutions and cross-scale fusions to get the enhanced EfficientNet (EEffNet). The extended convolutions enhance the backbone’s multi-scale representation while the cross-scale fusion mutually integrates the multi-scale features further, which transfer the VGG-style backbone into the mutually enhanced HRNet-style backbone. Compared with the former, the latter has more powerful feature extraction capability and high-resolution retaining ability, with multi-scale branches and cross-scale enhancement.
As shown in Figure 3, compared with EffNet’s one-scale structure, EEffNet is a multi-scale architecture where the extended Convs of EEffNet are cloned from the last convolution of the corresponding scale and the cross-scale fusion represents the total fusion layer, which will be introduced in the next subsection. For example, in the final stage, S 2 / S 3 (scale with sizes of 64/32), the features go through some extended convolutions, while S 4 (the scale with size of 16) are dealt by the same convolutions as the original EffNet, then they are integrated by the total fusion layer. By applying MEMe like this, our EEffNet is rebuilt from EffNet, which has a more powerful feature pyramid output.

3.2. Total Fusion Neck

Bidirectional FPN (BiFPN) from EfficientDet is a classical multi-scale feature fusion structure in object detection and it inspired us to design an N–N neck structure (N–N means that the input features of the N scales fuse into the output features of the same N scales), which have never been used in human pose estimation before. As shown in Figure 4a, BiFPN can realize the top–down and bottom–up feature fusion, where the consecutive and bidirectional path causes partial, restricted, and weight-unbalanced feature fusion.
Based on this, we propose a total fusion block where the feature pyramid from the backbone is mutually enhanced by a more sufficient, flexible, and a weight-balanced fusion strategy, including AddUp, MixUp, and FixUp. The main idea of the total fusion block is the connection relationship reflected by AddUp, MixUp, and FixUp rather than some magical operations. For example, AddUp emphasizes the need for dense connections between multi-scale features instead of unbalanced connections, such as BiFPN; MixUp is responsible for feature fusion and extraction, to get deeper representations; FixUp connects original-scale features to fix it. Finally, we cascaded the same numbers of repeated total fusion blocks as the number of BiFPN in EffBase to form the total fusion neck (TFNeck).
Formally, the equation of Total Fusion is:
s i o u t = T F ( S J i n , s i i n ) , J = { A l l S c a l e s } = F i x U p ( M i x U p ( A d d U p ( S J i n ) ) , s i i n )
where s i i n is i-th scale’s input, S J i n is the total set of s i i n , s i o u t is the i-th scale’s output, and J means all scale features. Moreover, they can be shown formally below:
s i a d d = A d d U p ( S J i n ) , J = { A l l S c a l e s } = M i s h ( j J w j · R e s i z e j i ( s j i n ) )
s i m i x = M i x U p ( s i a d d ) = M o b i l e B l o c k i ( s i a d d )
s i o u t = F i x U p ( s i m i x , s i i n ) = s i m i x + s i i n
where s j i n is j-th scale’s input of TF, s i a d d is i-th scale’s output of AddUp and also the input of MixUp. Mish is the activation function. Similarly, s i m i x is i-th scale’s output of MixUp and the input of FixUp, S J i n is the all of s j i n , and R e s i z e j i means resize j-th to i-th scale.
Combined with Figure 4, we explain our equation Equation (2), specifically in terms of the scale’s total fusion. First, for the AddUp (black arrow in Figure 4) operation of Equation (3), there are three scale inputs s j i n , j { 2 , 3 , 4 } , where all scale features are added up totally into each output of scale i t h , which is written as s i a d d . Then, for the MixUp (red arrow in Figure 4) operation of Equation (4), the input s i a d d independently passes a MobileBlock to get the output s i m i x . Finally, the FixUp (blue dotted arrow in Figure 4) operation of Equation (5) is an identity shortcut of i-th scale, which can ensure input and output feature resolution aligned, as well as keep a smoother gradient flow. To be specific, we extended the skip-connection operation in each scale s i i n to add the middle fusion output s i m i x into final s i o u t . By applying MEMe on BiFPN like this, our TFNeck can achieve a more sufficient, flexible, and weight-balanced feature fusion.

3.3. Final Attention Head

As shown intuitively in Figure 5, the way the FSHead gets the final output by summing or averaging the multi-scale outputs, may introduce the upsampling artificial noises. It may be an inappropriate cross-scale output ensemble by the linear combinations as they are in different semantic level outputs and could not achieve a better consistent output if the network effectively extracts diverse multi-scale information.
Therefore, combined with the attention mechanism rather than the liner combination ensemble, we used the Sigmoid function to activate both the high-level scale feature with more semantic information and the low-level scale feature usually with more detailed information to get multi-scale attention. Then, we point-wise multiply them together to make those attentions query and calibrate each other’s output, which will make better use of their inherent structural advantages in multi-scale receptive fields and circumvent the accumulation of cross-scale errors. After the final attention head, only the heatmap of the low-level scale validated by higher-level attention can be responded to, which means the output is still a low-level scale and high-resolution heatmap without upsampling noise, but with the help of higher-scale attention.
For clear comparison, their formulas can be written as:
H 2 = F S H e a d ( S J o u t ) = j C o n v B N j ( s j o u t )
H 2 = F A H e a d ( S J o u t ) = j σ ( C o n v B N j ( s j o u t ) )
where input of FAHead S J is the universe of outputs by the previous neck, H 2 means the output of FAHead, and σ represents the activation function of Sigmoid.
Equation (6) means the traditional sum fusion that directly fuses the features from the different feature spaces and is harmful for location information representation. By applying a MEMe like this, our Equation (7) utilizes the Sigmoid activation function and Point-Wise Multiplication to realize the cross attention of the S J leading to a calibrated output.

3.4. Instantiation

In this paper, to make a fair comparison, we used MEMe to reconstruct the EffBase into EEffPose-P0 for 256 × 192 and EEffPose-P2 for 384 × 288 on COCO. For the detailed structure information, EEffPose-P0/P2’s backbone EEffNet-B0/B2 were transferred from EffNet-B0/B2. As for the TFNeck of EEffPose-P0/P2, the feature pyramid of EEffNet-B0/B2 go through the same number of repeated total fusion blocks as EfficientDet-D0/D2. As for the head of EEffPose-P0/P2, they are sampled to the corresponding keypoint heatmap, whose size is the same as S 2 by convention.
Finally, our EEffPose-P0/P2 has the parameters of 8.98 M/16.7 M, and a GFLOPs of 4.94 G/23.0 G. Moreover, compared with HRNet-W32/W48, whose parameters are 28.5 M/63.6 M, and GFLOPs are 7.7 G/35.4 G, our MEMe-based models are of low-cost in the expanded deployments.

4. Experiments and Analysis

In this paper, we used MS COCO and MPII as the evaluation benchmarks to show the performance of our MEMe-based EEffPose and to verify the necessity of MEMe. To investigate the effectiveness of specific modules of the model and demonstrate the versatility of MEMe, we designed the ablation study on COCO val and conducted a series of visualizations for further explanation. Then, by comparing with the state-of-the-art, including small and large models, our MEMe-based EEffPose showed a superior performance and the best trade-off between performance and complexity. Finally, we demonstrated some qualitative presentations for samples in COCO and MPII, as shown in Figure 6 and Figure 7.
Implementation Details. Our experiments were trained on 4 NVIDIA A100 GPUs with PyTorch 1.9, CUDA 11.1, and cuDNN 8.0.5. For the system, it was a Ubuntu 16.04 running on an Intel [email protected] GHz CPU and 64 G RAM memory.
Datasets. The COCO dataset [19] is a rich dataset containing more than 90 targets, 0.3 billion images, and 2.5 billion labels. We trained our model on a MS COCO 2017 training set, which included 57,000 images and 150,000 person instances with 17 keypoints. The validation set including 5000 images was used to validate training performance. The results on the test-dev set (20,000 images) were also list, to make a more fair and convincing comparison between our model and other SOTA methods. Moreover, we report results on the MPII dataset [20]. MPII is a popular benchmark for single-person 2D pose estimation, which has 25,000 images. In total, there are 29,000 annotated poses for training, and another 7000 poses for testing.
Training. In the training process, each GPU occupies 16 samples. The Adam optimizer is used with ReduceLROnPlateau, where the initial learning rate is 0.001, the descending factor is 0.3, and descending patience is 5. Augmentation operations are used for each input image, including scale, rotation, flipping, half body data augmentation (only for COCO), and cropping, which is the same as HRNet [9] and SimpleBaseline [8]. The scaling factor is sampled between ± 0.35 / ± 0.30 and the the rotating factor is between ± 45 °/ ± 25 ° for COCO and MPII, respectively. Finally, the cropping operation gets a size of 256/384 for EEffPose-P0 and EEffPose-P2, while keeping the aspect ratio to 3:4/4:4 for COCO and MPII.
Testing. Testing is a top–down process, where the human target is extracted first and then keypoints are detected. Moreover, this testing process is the same as the HRNet and SimpleBaseline. The model outputs a heatmap where all keypoint detected results and flipped image results are averaged. Finally, the final position is achieved by a quarter of the pixel’s shift from the main wave peak.
Evaluation. Mean average precision (mAP) from object keypoint similarity (OKS) is used as the evaluation metric on COCO, where OKS uses the Euclidean distance between the predicted keypoints and ground-truths to evaluate the similarity of keypoint pairs.The head-normalized probability of correct keypoint (PCKh) is the evaluation metric on MPII, which can detect whether the keypoints locate in the ground-truth adjacent range. In our paper, we demonstrate the outcomes of AP, AP .5 (IOU > 0.5), AP .75 (IOU > 0.75), AP (M) (for the middle targets), AP (L) (for the large targets), and AR on COCO and PCKh on MPII.

4.1. Ablation Study

We use our EEffPose-P0 for the ablation study, which is conducted on the COCO val set with an input size of 256 × 192. To study the effects of MEMe at different stages, we design eight ablation models that alternately use MEMe on the backbone, neck, and head. We report the metrics and their convergence curves for the analysis. All experiments are performed in the same configuration, except for the maximum learning epoch and the decay of learning rate due to the use of ReduceLROnPlateau, which reduces the learning rate when the validation AP stops improving. It is still a fair comparison and the learning rate decay curve can further reflect the convergence speed and difficulty of the model.
MEMe-based Backbone (EEffNet). From Table 1, the following results can be seen: first, Order 2 gets a gain of 2.7 AP and 2.0 AR compared with Order 1, proving the effectiveness of using the EEffNet alone. Moreover, compared with Order 3, Order 6 increases the AP and AR scores by 2.1 and 1.8, showing that the MEMe backbone is still effective after using the TFNeck. Furthermore, Order 7 improves the AP and AR scores of Order 4 by 2.4 and 2.1, indicating that the MEMe backbone is still helpful on the basis of the FAHead. Finally, Order 8 shows 2.2 and 1.9 improvements on AP and AR scores than Order 5, demonstrating that the MEMe backbone can still make progress after using both the TFNeck and the FAHead. Meanwhile, a corroborating conclusion can be obtained from Figure 8 that Order 2, 6, 7 and 8 (solid lines) has higher AP scores during training than Order 1, 3, 4 and 5 (dotted lines). The above conclusions show that our enhanced backbone (EEffNet) can greatly increase the model capacity and improve performance.
MEMe-based Neck (TFNeck). From Table 1, the following results can be seen: first, Order 3 gets a gain of 3.8 AP and 1.9 AR compared with Order 1, proving the effectiveness of using the TFNeck alone. Moreover, compared with Order 2, Order 6 increases the AP and AR scores by 3.0 and 1.7, showing that the TFNeck is still effective after using the EEffNet. Furthermore, Order 5 improves the AP and AR scores of Order 4 by 0.9 and 0.8, indicating that the TFNeck is still helpful on the basis of the FAHead. Finally, Order 8 shows 0.7 and 0.6 improvements on AP and AR scores than Order 7, demonstrating that the TFNeck can still make progress after using both the EEffNet and the FAHead. Meanwhile, Figure 8 corroborates that Order 3, 6, 5 and 8 (thick lines) has steadier and more sustainable convergence than Order 1, 2, 4 and 7 (thin lines) instead of falling into local optima too soon. The above conclusions show that our enhanced neck (TFNeck) can maintain a stable and sustainable convergence and prevent falling into a local optimal situation.
MEMe-based Head (FAHead). From Table 1, the following results can be seen: First, Order 4 gets a gain of 4.5 AP and 2.7 AR compared with Order 1, proving the effectiveness of using the FAHead alone; Moreover, compared with Order 2, Order 7 increases the AP and AR scores by 4.2 and 2.8, showing that the FAHead is still effective after using the EEffNet. Furthermore, Order 5 improves the AP and AR scores of Order 3 by 1.6 and 1.6, indicating that the FAHead is still helpful on the basis of the TFNeck. Finally, Order 8 shows 1.9 and 1.7 improvements on AP and AR scores than Order 6, demonstrating that the FAHead can still make progress after using both the EEffNet and the TFNeck. Meanwhile, Figure 8 reveals that Order 4, 7, 5 and 8 (red lines) has faster and better convergence than Order 1, 2, 3 and 6 (blue lines). The above conclusions show that our enhanced head (EEffNet) can boost the model to get a fast and excellent convergence during training time.
Visualization and Analysis. Figure 9 is used to clearly demonstrate the performance of our proposed three modules. Firstly, to show the effect of EEffNet, by the comparison of (a, b), the S 2 and S 3 of some keypoints, such as K 14 and K 15 obviously depict that our proposed EEffNet generates a more clear Gaussian Heatmap with a more powerful information extraction ability. Secondly, to show the performance of TFNeck, by the comparison of (a, b) and (c), (a, b)’s S 4 of all keypoints shows the nearly blank heatmap (the bright color here is due to the pseudo-color conversion), which cannot extra a proper feature. While (c)’s K 14 and K 15 can get a better feature with our TFNeck, which proves the necessity of flexible information flow and weight-balanced feature fusion. Finally, to show the performance of FAHead, by comparison of (c, d), the phenomenon is that all scale outputs of (d) are the enhanced Gaussian heatmaps, just with different sizes, except K 16 and K 17 . It proves that our FAHead, based on the cross attention mechanism, can generate better Gaussian heatmaps and reject the pseudo-shadow points, such as K 16 and K 17 in (a, c).

4.2. Comparisons to the State-of-the-Art

COCO val. In Table 2, for the EEffPose-P0 of 256 × 192, it achieves a score of 75.4 AP, which outperforms not only other small models, but also the big ones, such as HRNet-W48 with 256 × 192. For the EEffPose-P2 of 384 × 288, it achieves a 76.7 AP score and is higher than HRNet-W48, reaching the SOTA performance. Specifically, compared with the LiteHRNet-18/30, EEffPose-P0/P2 increases the AP by 10.6/6.3, with an acceptable increment in parameters. Similarly, compared with the DANet, EEffPose-P0/P2 increases the AP score by 4.4/3.3. Next, compared with other lightweight models, ShuffleNetV2 and MobileNetV2, both EEffPose-P0/P2 improve over 10 gains. Compared to large networks, such as HRNet, EEffPose-P0 can also achieve a higher AP score with a much lower model complexity. Especially, EEffPose-P0 with 256 × 192 has a higher AP score than that of HRNet-W48 with 384 × 288 using only 14% of its parameters. In Figure 1, it shows that our MEMe-based EEffPose reaches the optimal trade-off even with the best AP than others and enhances the lightweight model’s performance to cross the gap without too much pain.
COCO test-dev. Table 3 depicts the experimental results of our MEMe-based EEffPose and other methods. Compared with the small networks, such as LiteHRNet, DANet, MobileNet, our proposed EEffPose-P0/P2 reach the SOTA performance with an AP score of 74.2/75.9. Compared with the large networks in the same input resolution level, our proposed EEffPose has a higher AP score than Hourglass, CPN, HRNet, and SimpleBaseline, with a much lower computation cost succeeding in finding the best trade-off.
MPII val. Table 4 shows the results of our EEffPose and other models with the input of 256 × 256. EEffPose-P0/P2 get 90.38/90.84 [email protected], respectively, which outperform the large model, such as HRNet-W32/W48, and the small models, such as MobileNetV2, MobileNetV3, ShuffleNetV2 and LiteHRNet. Furthermore, our EEffPose-P2 can still make progress while HRNet-W48 cannot, due to the saturated dataset performance [9].

5. Conclusions and Discussions

In this paper, we proposed the mutually enhanced modeling method (MEMe) to enhance the lightweight model EffBase into a highly efficient EEffPose, which enhanced the original EffNet, BiPFN, FSHead into the mutually enhanced EEffNet, TFNeck, and FAHead, to fill the gap between the high accuracy of the large models and the low accuracy of the small models. Extensive experiments prove that our proposed EEffPose reaches the state-of-the-art, with low cost, and our MEMe plays a key role in this.
Moreover, our proposed MEMe is a general enhancing modeling method that can be applied to other lightweight models in human pose estimations, to improve their accuracy. Moreover, the MEMe can also be applied to many other vision tasks, such as segmentation, object detection, object tracking, etc., to improve accuracy with a low model complexity. The rise of HRNet brings the basic architecture of the multi-scale feature representation while our proposed MEMe not only fills the performance gap, but also leads to evolution among mutually enhancing multi-scale features.

Author Contributions

Conceptualization, J.L. and J.Z.; methodology, J.L., Z.W. and J.Z.; software, J.L.; validation, J.Z.; formal analysis, J.L.; investigation, J.L.; resources, B.Q., J.Z. and H.Y.; writing—original draft preparation, J.L. and Z.W.; writing—review and editing, J.L., B.Q. and J.Z.; visualization, J.L.; supervision, B.Q., J.Z. and H.Y.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The code of EEffPose will be released on GitHub before 1 July 2022: https://github.com/igodogi/EEffPose and can also be early available by contact if interested; for MSCOCO dataset: https://cocodataset.org; for MPII dataset: http://human-pose.mpi-inf.mpg.de/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zheng, C.; Wu, W.; Yang, T.; Zhu, S.; Chen, C.; Liu, R.; Shen, J.; Kehtarnavaz, N.; Shah, M. Deep Learning-Based Human Pose Estimation: A Survey. arXiv 2021, arXiv:2012.13392. [Google Scholar]
  2. Chen, Y.; Tian, Y.; He, M. Monocular human pose estimation: A survey of deep learning-based methods. Comput. Vis. Image Underst. 2020, 192, 102897. [Google Scholar] [CrossRef]
  3. Yang, W.; Zhang, J.; Cai, J.; Xu, Z. Shallow Graph Convolutional Network for Skeleton-Based Action Recognition. Sensors 2021, 21, 452. [Google Scholar] [CrossRef] [PubMed]
  4. Badiola-Bengoa, A.; Mendez-Zorrilla, A. A Systematic Review of the Application of Camera-Based Human Pose Estimation in the Field of Sport and Physical Exercise. Sensors 2021, 21, 5996. [Google Scholar] [CrossRef] [PubMed]
  5. Stenum, J.; Cherry-Allen, K.M.; Pyles, C.O.; Reetzke, R.D.; Vignos, M.F.; Roemmich, R.T. Applications of Pose Estimation in Human Health and Performance across the Lifespan. Sensors 2021, 21, 7315. [Google Scholar] [CrossRef] [PubMed]
  6. Newell, A.; Yang, K.; Deng, J. Stacked Hourglass Networks for Human Pose Estimation. In Proceedings of the Computer Vision—ECCV 2016, Amsterdam, The Netherlands, 11–14 October 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 483–499. [Google Scholar] [CrossRef] [Green Version]
  7. Chen, Y.; Wang, Z.; Peng, Y.; Zhang, Z.; Yu, G.; Sun, J. Cascaded Pyramid Network for Multi-person Pose Estimation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7103–7112. [Google Scholar] [CrossRef] [Green Version]
  8. Xiao, B.; Wu, H.; Wei, Y. Simple Baselines for Human Pose Estimation and Tracking. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 472–487. [Google Scholar] [CrossRef] [Green Version]
  9. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep High-Resolution Representation Learning for Human Pose Estimation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 5686–5696. [Google Scholar] [CrossRef] [Green Version]
  10. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  11. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  12. Howard, A.; Sandler, M.; Chen, B.; Wang, W.; Chen, L.C.; Tan, M.; Chu, G.; Vasudevan, V.; Zhu, Y.; Pang, R.; et al. Searching for MobileNetV3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar] [CrossRef]
  13. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 6848–6856. [Google Scholar] [CrossRef] [Green Version]
  14. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 122–138. [Google Scholar] [CrossRef] [Green Version]
  15. Debnath, B.; Yamaguchi, M.; Behera, A. Adapting MobileNets for mobile based upper body pose estimation. In Proceedings of the 2018 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; IEEE Computer Society: Los Alamitos, CA, USA, 2018; pp. 1–6. [Google Scholar] [CrossRef] [Green Version]
  16. Yu, C.; Xiao, B.; Gao, C.; Yuan, L.; Zhang, L.; Sang, N.; Wang, J. Lite-HRNet: A Lightweight High-Resolution Network. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–21 June 2021; IEEE Computer Society: Los Alamitos, CA, USA, 2021; pp. 10435–10445. [Google Scholar] [CrossRef]
  17. Luo, Z.; Wang, Z.; Cai, Y.; Wang, G.; Wang, L.; Huang, Y.; Zhou, E.; Tan, T.; Sun, J. Efficient Human Pose Estimation by Learning Deeply Aggregated Representations. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; IEEE Computer Society: Los Alamitos, CA, USA, 2021; pp. 1–6. [Google Scholar] [CrossRef]
  18. Tan, M.; Pang, R.; Le, Q.V. EfficientDet: Scalable and Efficient Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–18 June 2020; IEEE Computer Society: Los Alamitos, CA, USA, 2020; pp. 10778–10787. [Google Scholar] [CrossRef]
  19. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014, Zurich, Switzerland, 6–12 September 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Springer International Publishing: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar] [CrossRef] [Green Version]
  20. Andriluka, M.; Pishchulin, L.; Gehler, P.; Schiele, B. 2D Human Pose Estimation: New Benchmark and State of the Art Analysis. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3686–3693. [Google Scholar] [CrossRef]
  21. Cao, Z.; Simon, T.; Wei, S.; Sheikh, Y. Realtime Multi-person 2D Pose Estimation Using Part Affinity Fields. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Los Alamitos, CA, USA, 2017; pp. 1302–1310. [Google Scholar] [CrossRef] [Green Version]
  22. Martinez, G.; Raaj, Y.; Idrees, H.; Xiang, D.; Joo, H.; Simon, T.; Sheikh, Y. Single-Network Whole-Body Pose Estimation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 6981–6990. [Google Scholar] [CrossRef] [Green Version]
  23. Fang, H.S.; Xie, S.; Tai, Y.W.; Lu, C. RMPE: Regional Multi-person Pose Estimation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2353–2362. [Google Scholar] [CrossRef] [Green Version]
  24. Bulat, A.; Kossaifi, J.; Tzimiropoulos, G.; Pantic, M. Toward fast and accurate human pose estimation via soft-gated skip connections. In Proceedings of the 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), Buenos Aires, Argentina, 18–22 May 2020; pp. 8–15. [Google Scholar] [CrossRef]
  25. Neff, C.; Sheth, A.; Furgurson, S.; Tabkhi, H. EfficientHRNet: Efficient Scaling for Lightweight High-Resolution Multi-Person Pose Estimation. arXiv 2020, arXiv:2007.08090. [Google Scholar]
  26. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, ICML 2019, Long Beach, CA, USA, 9–15 June 2019; Chaudhuri, K., Salakhutdinov, R., Eds.; PMLR: London, UK, 2019; Volume 97, pp. 6105–6114. [Google Scholar]
  27. Groos, D.; Ramampiaro, H.; Ihlen, E.A. EfficientPose: Scalable single-person pose estimation. Appl. Intell. 2021, 51, 2518–2533. [Google Scholar] [CrossRef]
  28. Lin, T.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Los Alamitos, CA, USA, 2017; pp. 936–944. [Google Scholar] [CrossRef] [Green Version]
  29. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; IEEE Computer Society: Los Alamitos, CA, USA, 2018; pp. 8759–8768. [Google Scholar] [CrossRef] [Green Version]
  30. Ghiasi, G.; Lin, T.; Le, Q.V. NAS-FPN: Learning Scalable Feature Pyramid Architecture for Object Detection. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 7029–7038. [Google Scholar] [CrossRef] [Green Version]
  31. Li, W.; Wang, Z.; Yin, B.; Peng, Q.; Du, Y.; Xiao, T.; Yu, G.; Lu, H.; Wei, Y.; Sun, J. Rethinking on Multi-Stage Networks for Human Pose Estimation. arXiv 2019, arXiv:1901.00148. [Google Scholar]
  32. Toshev, A.; Szegedy, C. DeepPose: Human Pose Estimation via Deep Neural Networks. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 1653–1660. [Google Scholar] [CrossRef] [Green Version]
  33. Chen, X.; Yuille, A.L. Articulated Pose Estimation by a Graphical Model with Image Dependent Pairwise Relations. In Advances in Neural Information Processing Systems; Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K.Q., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2014; Volume 27. [Google Scholar]
  34. Lin, T.; Wang, Y.; Liu, X.; Qiu, X. A Survey of Transformers. arXiv 2021, arXiv:2106.04554. [Google Scholar]
  35. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. arXiv 2021, arXiv:2012.12556. [Google Scholar]
  36. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-End Object Detection with Transformers. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 213–229. [Google Scholar] [CrossRef]
  37. Mao, W.; Ge, Y.; Shen, C.; Tian, Z.; Wang, X.; Wang, Z. TFPose: Direct Human Pose Estimation with Transformers. arXiv 2021, arXiv:2103.15320. [Google Scholar]
  38. Li, K.; Wang, S.; Zhang, X.; Xu, Y.; Xu, W.; Tu, Z. Pose Recognition with Cascade Transformers. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 21–24 June 2021; IEEE Computer Society: Los Alamitos, CA, USA, 2021; pp. 1944–1953. [Google Scholar] [CrossRef]
  39. Liu, H.; Liu, F.; Fan, X.; Huang, D. Polarized Self-Attention: Towards High-quality Pixel-wise Regression. arXiv 2021, arXiv:2107.00782. [Google Scholar]
  40. Zhang, Z.; Tang, J.; Wu, G. Simple and Lightweight Human Pose Estimation. arXiv 2020, arXiv:1911.10346. [Google Scholar]
Figure 1. Demonstration of the performance gap (in gray) and comparison of our models (in red) with others. In the figure, the model with less parameters numbers and higher AP is more efficient and effective, respectively. Efficient baseline (EffBase) means our baseline model transformed from EfficientDet. While efficient and effective pose (EEffPose) is reconstructed from EffBase by using our mutually enhanced modeling method (MEMe). It is noteworthy that our MEMe-based models (EEffPose-P0 and its expanded version EEffPose-P2), cross the performance gap, archive the state-of-art among small models, and even surpass HRNet—the typical representative of large models.
Figure 1. Demonstration of the performance gap (in gray) and comparison of our models (in red) with others. In the figure, the model with less parameters numbers and higher AP is more efficient and effective, respectively. Efficient baseline (EffBase) means our baseline model transformed from EfficientDet. While efficient and effective pose (EEffPose) is reconstructed from EffBase by using our mutually enhanced modeling method (MEMe). It is noteworthy that our MEMe-based models (EEffPose-P0 and its expanded version EEffPose-P2), cross the performance gap, archive the state-of-art among small models, and even surpass HRNet—the typical representative of large models.
Sensors 22 00632 g001
Figure 2. An overview of our MEMe-based pose net, efficient and effective pose (EEffPose). Our EEffPose contains three modules: enhanced EfficientNet (EEffNet), total fusion neck (TFNeck) and final attention head (FAHead). In figure, S j means the j-th scale feature in the feature pyramid. In particular, H 2 represents the heatmap output with the size of a quarter of the input.
Figure 2. An overview of our MEMe-based pose net, efficient and effective pose (EEffPose). Our EEffPose contains three modules: enhanced EfficientNet (EEffNet), total fusion neck (TFNeck) and final attention head (FAHead). In figure, S j means the j-th scale feature in the feature pyramid. In particular, H 2 represents the heatmap output with the size of a quarter of the input.
Sensors 22 00632 g002
Figure 3. EffNet vs. MEMe-based EEffNet. (a) The EffNet. (b) The mutually enhanced EEffNet. Compared with (a), (b) has a more powerful multi-scale representation by using our Extended Convs (gray box), and a more flexible information flow by using our Cross-scale fusion (red arrow). As a MEMe-base backbone, our EEffNet shows how to use MEMe to mutually enhance a backbone.
Figure 3. EffNet vs. MEMe-based EEffNet. (a) The EffNet. (b) The mutually enhanced EEffNet. Compared with (a), (b) has a more powerful multi-scale representation by using our Extended Convs (gray box), and a more flexible information flow by using our Cross-scale fusion (red arrow). As a MEMe-base backbone, our EEffNet shows how to use MEMe to mutually enhance a backbone.
Sensors 22 00632 g003
Figure 4. BiFPN vs. MEMe-based TotalFusion. (a) the BiFPN, Bidirectional FPN. (b) The mutually enhanced TotalFusion. For TotalFusion, the feature pyramid from the backbone is mutually enhanced by AddUp (black), MixUp (red), and FixUp (blue), where ’total’ is the key point compared with BiFPN. As a MEMe-base neck, our TotalFusion shows how to use MEMe to mutually enhance a neck.
Figure 4. BiFPN vs. MEMe-based TotalFusion. (a) the BiFPN, Bidirectional FPN. (b) The mutually enhanced TotalFusion. For TotalFusion, the feature pyramid from the backbone is mutually enhanced by AddUp (black), MixUp (red), and FixUp (blue), where ’total’ is the key point compared with BiFPN. As a MEMe-base neck, our TotalFusion shows how to use MEMe to mutually enhance a neck.
Sensors 22 00632 g004
Figure 5. FSHead vs. MEMe-based FAHead. (a) Final sum head. (b) The mutually enhanced final attention head. The FAHead’s output is mutually enhanced by the Sigmoid and point-wise multiplication to query and calibrate each other, which is a simple but solid way of cross-attention. As a MEMe-base head, our FAHead shows how to use MEMe to mutually enhance a head.
Figure 5. FSHead vs. MEMe-based FAHead. (a) Final sum head. (b) The mutually enhanced final attention head. The FAHead’s output is mutually enhanced by the Sigmoid and point-wise multiplication to query and calibrate each other, which is a simple but solid way of cross-attention. As a MEMe-base head, our FAHead shows how to use MEMe to mutually enhance a head.
Sensors 22 00632 g005
Figure 6. Presentation for COCO: including viewpoint change, occlusion, and multiple persons.
Figure 6. Presentation for COCO: including viewpoint change, occlusion, and multiple persons.
Sensors 22 00632 g006
Figure 7. Presentation for MPII: including viewpoint change, occlusion, and multiple persons.
Figure 7. Presentation for MPII: including viewpoint change, occlusion, and multiple persons.
Sensors 22 00632 g007
Figure 8. The convergence curves and LR self-adaption decay curves of ablation models.
Figure 8. The convergence curves and LR self-adaption decay curves of ablation models.
Sensors 22 00632 g008
Figure 9. Visualization of the final outputs. (a) The outputs of Order 1, which means the baseline EffBase. (b) The outputs of Order 2 where the backbone is MEMe-based. (c) The outputs of Order 6 whose backbone and neck are both MEMe-based. (d) The outputs of Order 8, which is our final MEMe-based model EEffPose. K N is the heatmap of keypoint while H J * and H 2 are the multi-scale outputs, and the final one integrated as Equation (6) for (ac) or Equation (7) for (d). For a better comparative observation, we upsampled H J * as the same scale as H 2 . The S 0 +H 0 means S 0 (the input image) overlaps with all keypoint heatmaps to demonstrate the detection results of the keypoints.
Figure 9. Visualization of the final outputs. (a) The outputs of Order 1, which means the baseline EffBase. (b) The outputs of Order 2 where the backbone is MEMe-based. (c) The outputs of Order 6 whose backbone and neck are both MEMe-based. (d) The outputs of Order 8, which is our final MEMe-based model EEffPose. K N is the heatmap of keypoint while H J * and H 2 are the multi-scale outputs, and the final one integrated as Equation (6) for (ac) or Equation (7) for (d). For a better comparative observation, we upsampled H J * as the same scale as H 2 . The S 0 +H 0 means S 0 (the input image) overlaps with all keypoint heatmaps to demonstrate the detection results of the keypoints.
Sensors 22 00632 g009
Table 1. The AP and AR of ablation models on the COCO val set with the input size of 256 × 192.
Table 1. The AP and AR of ablation models on the COCO val set with the input size of 256 × 192.
OrderMEMe-Based BackboneMEMe-Based NeckMEMe-Based HeadAP↑AR↑
167.875.7
270.577.7
371.677.6
472.378.4
573.279.2
673.579.4
774.780.5
875.481.1
It is worth mentioning that Order 1 depicts our baseline EffBase and Order 8 represents our EEffPose. In MEMe-based Backbone: ✗ means original backbone (EffNet), while ✓ means our backbone (EEffNet). In MEMe-based Neck: ✗ means original neck (BiFPN), while ✓ means our neck (TFNeck). In MEMe-based Head: ✗ means original head (FSHead), while ✓ means our head (FAHead).
Table 2. Comparisons to the state-of-the-art on the COCO val set.
Table 2. Comparisons to the state-of-the-art on the COCO val set.
MethodBackboneInput SizeParamsGFLOPsAP↑AP .5AP .75AP (M)AP (L)AR↑
SmallLiteHRNet [16]LiteHRNet-18256 × 1921.1 M0.264.886.773.062.170.571.2
LiteHRNet-30384 × 2881.8 M0.770.488.777.767.576.376.2
DANet [17]DANet-72256 × 1923.4 M1.071.089.678.167.676.276.6
DANet-102256 × 1925.8 M1.873.490.680.569.978.678.8
LPN [40]ResNet-152256 × 1927.4 M1.871.089.278.667.877.776.8
ShuffleNetV2 [14]ShuffleNetV2256 × 1927.6 M1.2859.985.466.356.666.266.4
ShuffleNetV2384 × 2887.6 M2.8763.686.570.559.570.769.7
MobileNetV2 [11]MobileNetV2256 × 1929.6 M1.4864.687.472.361.171.270.7
MobileNetV2384 × 2889.6 M3.3367.387.974.362.874.772.9
OursEEffPose-P0256 × 1928.98 M4.9475.490.582.671.782.281.1
EEffPose-P2384 × 28816.7 M23.076.790.583.172.983.682.1
Large8-S Hourglass [6]Hourglass256 × 19225.2 M26.266.9-----
CPN [7]ResNet-50256 × 19227.0 M6.2068.6-----
HRNet [9]HRNet-W32256 × 19228.5 M7.7074.490.581.970.881.079.8
HRNet-W32384 × 28828.5 M17.375.890.682.771.982.881.0
HRNet-W48256 × 19263.6 M15.775.190.682.271.581.880.4
HRNet-W48384 × 28863.6 M35.476.390.882.972.383.481.2
SimpleBaseline [8]ResNet-50256 × 19234.0 M8.9070.488.678.367.177.276.3
ResNet-152384 × 28868.6 M35.674.389.681.170.579.779.7
Table 3. Comparisons to the state-of-the-art on the COCO test set.
Table 3. Comparisons to the state-of-the-art on the COCO test set.
MethodBackboneInput SizeParamsGFLOPsAP↑AP .5AP .75AP (M)AP (L)AR↑
SmallLiteHRNet [16]LiteHRNet-30384 × 2881.8 M0.7069.790.777.566.975.075.4
DANet [17]DANet-72256 × 1923.4 M1.070.591.878.267.575.076.7
DANet-88256 × 1925.3 M1.471.591.979.568.576.077.6
LPN [40]ResNet-152256 × 1927.4 M1.871.089.278.667.877.776.8
MobileNetV2 [11]MobileNetV2384 × 2889.6 M3.3366.890.074.062.673.372.3
OursEEffPose-P0256 × 1928.98 M4.9474.292.082.170.979.980.0
EEffPose-P2384 × 28816.7 M23.075.992.483.472.481.981.3
Large8-S Hourglass [6]Hourglass256 × 19225.2 M26.266.9-----
CPN  [7]Res-Inception384 × 288--72.191.480.068.777.278.5
HRNet [9]HRNet-W32256 × 19228.5 M7.7073.592.281.970.279.279.0
HRNet-W32384 × 28828.5 M17.374.992.582.871.380.980.1
HRNet-W48256 × 19263.6 M15.774.292.482.470.979.779.5
HRNet-W48384 × 28863.6 M35.475.592.583.371.981.580.5
SimpleBaseline [8]ResNet-152384 × 28868.6 M35.673.791.981.170.380.079.0
Table 4. Comparisons to the state-of-the-art on the MPII val set.
Table 4. Comparisons to the state-of-the-art on the MPII val set.
ModelParamsGFLOPs[email protected]
EfficientPose I  [27]0.7 M1.6785.2
HRNet-W16 [16]1.3 M0.7280.2
LiteHRNet-18 [16]1.1 M0.2786.1
LiteHRNet-30 [16]1.8 M0.4287.0
EfficientPose II [27]1.7 M7.7088.2
EfficientPose III [27]3.2 M23.489.5
EfficientPose IV [27]6.6 M72.989.8
ShuffleNetV2 [14]7.6 M1.7082.8
MobileNetV2 [11]9.6 M1.9785.4
HRNet-W32 [9]28.5 M10.2590.3
HRNet-W48 [9]63.6 M47.2390.28 *
EEffPose-P08.98 M6.5990.4
EEffPose-P216.72 M30.790.9
* is a re-implementing result for the unreleased result.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Wang, Z.; Qi, B.; Zhang, J.; Yang, H. MEMe: A Mutually Enhanced Modeling Method for Efficient and Effective Human Pose Estimation. Sensors 2022, 22, 632. https://doi.org/10.3390/s22020632

AMA Style

Li J, Wang Z, Qi B, Zhang J, Yang H. MEMe: A Mutually Enhanced Modeling Method for Efficient and Effective Human Pose Estimation. Sensors. 2022; 22(2):632. https://doi.org/10.3390/s22020632

Chicago/Turabian Style

Li, Jie, Zhixing Wang, Bo Qi, Jianlin Zhang, and Hu Yang. 2022. "MEMe: A Mutually Enhanced Modeling Method for Efficient and Effective Human Pose Estimation" Sensors 22, no. 2: 632. https://doi.org/10.3390/s22020632

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop