Next Article in Journal
Optimization of Process Parameters for Turning Hastelloy X under Different Machining Environments Using Evolutionary Algorithms: A Comparative Study
Next Article in Special Issue
Dilated Filters for Edge-Detection Algorithms
Previous Article in Journal
New Trends in Biosciences
Previous Article in Special Issue
An Efficient Text Detection Model for Street Signs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation

1
AI Graduate School, Ulsan National Institute of Science and Technology, Ulsan 44919, Korea
2
Department of Computer Science and Engineering, Ulsan National Institute of Science and Technology, Ulsan 44919, Korea
3
Railway Safety Research Division, Korea Railroad Research Institute, Uiwang-si 16105, Korea
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2021, 11(20), 9724; https://doi.org/10.3390/app11209724
Submission received: 31 August 2021 / Revised: 11 October 2021 / Accepted: 12 October 2021 / Published: 18 October 2021

Abstract

:
Three-dimensional human pose and shape estimation is an important problem in the computer vision community, with numerous applications such as augmented reality, virtual reality, human computer interaction, and so on. However, training accurate 3D human pose and shape estimators based on deep learning approaches requires a large number of images and corresponding 3D ground-truth pose pairs, which are costly to collect. To relieve this constraint, various types of weakly or self-supervised pose estimation approaches have been proposed. Nevertheless, these methods still involve supervision signals, which require effort to collect, such as unpaired large-scale 3D ground truth data, a small subset of 3D labeled data, video priors, and so on. Often, they require installing equipment such as a calibrated multi-camera system to acquire strong multi-view priors. In this paper, we propose a self-supervised learning framework for 3D human pose and shape estimation that does not require other forms of supervision signals while using only single 2D images. Our framework inputs single 2D images, estimates human 3D meshes in the intermediate layers, and is trained to solve four types of self-supervision tasks (i.e., three image manipulation tasks and one neural rendering task) whose ground-truths are all based on the single 2D images themselves. Through experiments, we demonstrate the effectiveness of our approach on 3D human pose benchmark datasets (i.e., Human3.6M, 3DPW, and LSP), where we present the new state-of-the-art among weakly/self-supervised methods.

1. Introduction

Tremendous progress has been made on estimating 3D human poses and shapes from a single image [1,2,3,4,5,6,7,8,9,10,11,12]. In this context, deep learning-based approaches have been successful over the last decades [1,2,3,4,5,6,7,8]. It was initially decided to estimate the 2D/3D skeletal representation of the human bodies [1,2,3,4]. The skeletal representation is efficient to describe key characteristics of the human motions, although the shape information is lacking. Recently, Kanazawa et al. [5] proposed the estimation of both poses and shapes of the human bodies by incorporating a differentiable 3D mesh representation—i.e., a skinned multi-person linear model (SMPL) [13]—in the deep learning framework. More recently, several frameworks that improve the 3D mesh estimation network [5] have been proposed to deal with temporal consistency [6], multi-person cases [7], domain differences [8], and so on.
One of the fundamental issues in constructing pose estimation frameworks is that they consume large numbers of 2D or 3D ground-truth poses (i.e., x, y, and z coordinate values of the 3D human bodies) for a given 2D RGB (i.e., red, green, blue) or 2.5D depth input image to secure good accuracy in the mesh estimation task. Many researchers have proposed million-scale data pairs to properly train such frameworks [14]. However, it is challenging to acquire large-scale datasets containing diverse variations and quality 3D annotations. Manually annotating such 3D coordinate values is non-trivial, and it takes a great deal of time and manual effort. One team attempted to relieve the issue by using synthetic datasets based on graphics engines [15,16]. However, the appearance of images obtained from the graphics engines showed an observable gap to the real samples [17], and it is well known that the models trained by pure synthetic datasets do not generalize well to the real testing datasets [18].
In this paper, we attempted to relieve the issue of insufficient data by proposing self-supervision losses to train the 3D human pose estimation framework without explicit 2D/3D skeletal ground-truths. Our self-supervision losses consist of four types of supervision whose ground-truths are defined based on single 2D images: jigsaw puzzling [19], rotation prediction [20], image in-painting [21], and image projection [22] tasks (as illustrated in Figure 1). In the jigsaw puzzling task, the image is divided into the D × D tiles and shuffled; then, the jigsaw puzzle solver is trained to estimate the order of the tiles. In the rotation prediction task, the image is rotated and the rotation predictor is trained to estimate the rotation degrees. In the image in-painting task, sub-patches of the image are erased and the image in-painting decoder is learned to generate the erased patches with the aid of the adversarial loss [23]. Lastly, in the image projection task, the neural 3D mesh renderer is used to differentially generate the 2D images by projecting the estimated 3D meshes. This task directly provides gradients to the estimated 3D meshes and updates the parameters of the 3D mesh estimator, while the former three tasks indirectly enrich the feature vector extracted from the 2D images via the feature extractor. Via the combination of the proposed four losses, the 3D mesh estimator produces more accurate 3D human meshes.
Via a series of experiments, we have observed that the proposed self-supervised losses with additional image databases are able to enrich the pre-trained 3D human pose estimator and achieve state-of-the-art accuracy among self/weakly/semi-supervised works. The major contributions of this paper are summarized as follows:
  • We construct a 3D human pose and shape estimation framework that could be trained by 2D single image-level self-supervision without the use of other forms of supervision signals, such as explicit 2D/3D skeletons, video-level, or multi-view priors;
  • We propose four types of self-supervised losses based on the 2D single images themselves and introduce a method to effectively train the entire networks. In particular, we investigate which are the most promising combinations of losses to effectively achieve the 2D single image-level self-supervision for 3D human mesh reconstruction;
  • The proposed method outperforms the competitive 3D human pose estimation algorithms, proving that leveraging single 2D images could be used for strong supervision to train networks for the 3D mesh reconstruction task.
We have made our code and data publicly available at https://github.com/JunukCha/SSPSE accessed on October 13 2021.

2. Related Works

In this section, we review the recent literature on 3D human pose estimation and 3D human pose and shape estimation works. Then, we analyze recent weak/semi-supervised approaches designed for 3D human pose and shape estimation that are closely related to ours.

2.1. Three-Dimensional Human Pose Estimation

Many recent studies have focused on estimating 2D [11] or 3D keypoint locations on the human body [12]. Normally, these keypoint locations include major body joints such as the wrists, ankles, elbows, neck, shoulders, and knees. The architectures of 2D pose detectors have been designed to map images into the 2D pose vector more accurately. Wei et al. [2] proposed a sequential architecture composed of convolutional neural networks (CNNs) with multiple sizes of receptive fields. Zhou et al. [3] proposed a method using both a 3D geometric prior and temporal smoothness prior to treat considerable uncertainties in 2D joint locations. Newell et al. [1] proposed stacked hourglass networks based on the successive architecture, which is composed of pooling and upsampling layers. As we are living in the 3D space, understanding poses in the 3D space is the natural extension to 2D pose estimation. Three-dimensional pose estimation aims to locate the key 3D joints of the human body from an image or a video, which are either in the single-view or multi-view setting. Most approaches are based on the two-stage approaches that first predict 2D joint locations using 2D pose detectors [1,2,3,24] and then predict 3D joint locations from the 2D joints or features by performing regression [4,25,26,27] or model fitting [3,28,29,30,31,32]. Xu et al. [24] proposed a graph stacked hourglass network for 3D human pose estimation, which consists of four stacked hourglasses. Estimated 2D joints are fed into the network, and the 3D pose is predicted. Martinez et al. [4] proposed a simple framework composed of fully connected layers to lift a 2D pose to a 3D angular pose. Moreno et al. [25] proposed a framework composed of a 2D joint detector based on CNN and inferred 3D poses from the detected 2D joints. These approaches have made great progress in improving the performance of 3D human pose estimation.

2.2. Three-Dimensional Human Pose and Shape Estimation

Followed by progresses in the field of 3D human pose estimation, simultaneously estimating the human body pose and shape has become a recent trend [10,33,34,35,36,37,38,39,40,41]. There are two main approaches for estimating 3D human poses and shapes (i.e., meshes): one is the optimization-based methods [9,42,43], and the other is the regression-based methods [5,6,7,8,36,37]. SMPLify [9] is one of the representative optimization-based methods. It first estimates the 2D skeletons from the images and fits the graphical 3D body model (i.e., SMPL [13]) to it using the optimization method. Kanazawa et al. [5] proposed the deep learning-based regression framework. This is an end-to-end framework for reconstructing a 3D human mesh from a single RGB image. Omran et al. [36] proposed neural body fitting. This framework generates 12 semantic human body parts with a semantic segmentation CNN. From these semantic human body parts, pose and shape parameters of SMPL are estimated and a 3D human mesh is reconstructed. Pavlakos et al. [37] proposed a framework that estimates heatmaps related to pose parameters of SMPL and silhouettes related to shape parameters of SMPL simultaneously. Using these heatmaps and silhouettes, a 3D human mesh is reconstructed. A method [44] has also been introduced that exploits the merits of both optimization-based and regression-based methods for 3D mesh estimation. The temporal dynamics of the human body and shape have also been incorporated in the framework of [6]. The framework used to overcome the adaptation of a pre-trained model of human mesh estimation to the out-of-domain field was proposed in [8]. In addition, non-parametric body mesh reconstruction methods have been proposed [38,45]. Tan et al. [38] proposed an encoder–decoder network: first, a decoder is trained to predict a body silhouette from the SMPL parameter input; second, an encoder is trained on a real image and corresponding silhouette pairs while the decoder is fixed. Lin et al. [45] proposed end-to-end human pose and mesh reconstruction with a transformer [46].

2.3. Weakly/Semi-Supervised Learning in 3D Human Mesh Estimation

For most human pose estimation methods [36,37,43,47,48,49,50], supervised learning prevails; however, securing the 3D mesh ground truth is non-trivial. Weakly or semi-supervised methods are designed to solve the issue of the lack of quality annotation by using the available easier annotations. In the context of the 3D human mesh estimation task, semi-supervised learning uses 3D skeletons that are coarser than 3D meshes, while weakly-supervised learning uses either 2D annotation [36,51,52] or pseudo-3D annotation [53,54,55,56]. In [36], 2D keypoint annotations are exploited to estimate SMPL body model parameters from CNNs to recover human 3D meshes following predicted body part segmentation masks. In [51], two anatomically inspired losses are proposed and used with a weakly supervised learning framework to jointly learn from large-scale in-the-wild 2D and indoor/synthetic 3D data. The authors of [52] suggest weakly-supervised solutions by adopting multi-view consistency loss in 2.5D pose representations leveraged by independent 2D annotated samples. In [53,54,55,56], multi-view geometry is used to resolve 3D depth ambiguity. We developed our original methods in the self-supervised setting only using the input 2D images; however, to compare with these weakly or semi-supervised approaches, we extended our model to be additionally learned by 2D or 3D skeletons if they are available. In this manner, we compared our method to the state-of-the-art self/weakly/semi-supervised approaches.

3. Method

Our self-supervised 3D human mesh estimation framework is detailed in this section and illustrated in Figure 2. In this work, we used the network architecture of [5,6] as our baseline 3D mesh reconstruction network, which uses the ResNet-50 [57] architecture as the feature extractor and predicts SMPL [13] parameters using it. Our aim is to increase the accuracy of the baseline 3D mesh reconstruction network using the single 2D image-level self-supervised losses and its training strategy. In the remainder of this section, we explain the process at a more detailed level. Prior to that, we begin the discussion by introducing our network architectures.

3.1. Network Architectures

Our feature extractor f Feat and mesh estimator f Mesh were taken from the previous mesh reconstruction method [5], and they were responsible for extracting features and estimating the 3D human meshes. Three networks (jigsaw puzzle solver f Jigsaw , rotation predictor f Rot , part in-painter f Inp ) were additionally involved to develop the self-supervision loss solving three different image manipulation tasks, and two additional networks (texture network f Tex , neural mesh renderer f NMR ) were involved to develop the self-supervision loss solving the image projection task. Finally, the discriminator f Disc was involved to capture the distribution of real 2D images containing human images. More details on individual networks are described in the remainder of this subsection.
Feature extractor f Feat : X F R 2048 × 1 . Similar to recent human mesh recovery works [5,6] that estimate 3D human mesh model (i.e., SMPL [13]) parameters from the RGB images x X , we involved the ResNet-50 [57] architecture as our feature extractor. It generates 2048-dimensional feature vectors f F from the input image x , whose size is resized to a 224 × 224 × 3 dimensional array.
Mesh estimator f Mesh : F M R 6890 × 13 , 776 . After extracting the feature vector f F , we mapped it to the corresponding 85-dimensional SMPL [13] parameters h H R 85 × 1 (we used 3, 10 , and 72-dimensional vectors as camera, shape, and pose parameters, respectively, in the same manner as in [5]). These parameters were used to differentially generate the corresponding human body meshes m M with 6890 vertices and 13,776 faces using the SMPL layer (refer to [5,13] for more details).
Jigsaw puzzle solver f Jigsaw : F × F × F × F × F O R C × 1 . The jigsaw puzzle solver was used to solve the first self-supervision task, called the jigsaw puzzling task. In this task, we divided the input 2D image x into D × D tiles and permuted them to generate a jigsaw puzzling image x jigsaw . Then, we input D 2 tiles combined with the original image into the feature extractor f Feat by resizing all images into 224 × 224 × 3 dimensional images. The resultant features from five images were concatenated, retaining the permuted orders, and mapped into the vector o O , whose entry denotes the probability for the permutation order of the tiles in the whole images. The class number C = ( D × D ) ! is equivalent to the number of permutations for D × D tiles. We found that D = 2 best encodes the human bodies, balancing it from the background regions (see Figure 1); thus, C = ( 2 × 2 ) ! = 24 was used throughout the experiment.
Rotation predictor f Rot : F R R R × 1 . The rotation predictor was used to solve the second self-supervision task, called the rotation prediction task. In this task, we rotated the input 2D image x with R fixed angles θ to generate a new image x rot . Humans are in the upright position in images with 0 degrees. We first applied f Feat on the input image x rot to obtain the feature vector f . Then, the rotation predictor f Rot was applied on f to predict the vector r R , which contained the probability for R possible angles. In this work, we set R = 4 by setting the rotation angles to 0 , 90 , 180 , and 270 for simplicity; however, we were able to achieve a meaningful accuracy improvement using this simple setting.
Part in-painter f Inp : F P R P × P × 3 . The part in-painter was used to solve the third self-supervision task, called the part in-painting task. In this task, we erased a P × P × 3 -sized patch p from the center of an input image x , resulting in x hole . Then, the part in-painter f Inp was responsible for predicting p ^ resembling p . We set P = 64 considering the patch size compared to the size of the human bodies in images (Figure 1d describes the example in-painted patches).
Neural mesh renderer f NMR : T × M X × S R 224 × 224 × 3 × R 224 × 224 . The neural mesh renderer [22] was employed to solve the last self-supervision task, called the image projection task. In this task, 3D meshes were projected to the 2D images by an operation similar to graphics rendering [22]. Either the RGB human images x X or a binary segmentation mask s S could be generated from the 3D meshes m . To render RGB images, an additional texture array t T was required that described RGB values for faces of the 3D meshes. To infer the texture array, we additionally involved the texture network f Tex , as explained below.
Texture network f Tex : F T R 13 , 776 × 3 . As there are 13 , 776 faces in the SMPL [13] model, the dimension of the texture array t is 13 , 776 × 3 , as it represents RGB values corresponding to each mesh face. Furthermore, we first inferred the 13 , 776 × 3 dimensional array t from f Tex and differentially reshaped this towards a 13 , 776 × Z × Z × Z × 3 dimensional array. Z = 2 was used as the visual quality based on this minimum dimension, which was sufficient for our purpose.
Discriminator f Disc : x [ 0 , 1 ] . The discriminator was involved to capture the distribution of the real human images and give gradients to the networks to make the realistic images, in a similar manner to [23]. The discriminator f Disc played a crucial role, especially when training the part in-painter f Inp and texture network f Tex , as these tasks require training networks to generate the realistic patches or textured images.
Table 1, Table 2 and Table 3 present the detailed structure of our network architectures. For f r cnn , f NMR , and f Feat , we employed the implementation and weights from [5,22,58], respectively. For the mesh estimator f Mesh as explained in Table 4, we involved the SMPL layer of [5], which can output the SMPL human body meshes from its estimated pose (72), shape (10), and camera (3) parameters.

3.2. Training Method

Our aim during the training stage was to enrich both feature extractor f Feat and mesh estimator f Mesh , which were responsible for estimating 3D meshes. Our losses were proposed to improve the two networks (i.e., feature extractor f Feat and mesh estimator f Mesh ) based on the 2D single image-level self-supervised losses, which were designed to solve three image manipulation tasks (i.e., jigsaw puzzle, rotation prediction, and part in-painting) and one image projection task. In this subsection, we explain the entire training strategy and introduce the individual losses that we used.

3.2.1. Summary of the Entire Training Process

The entire training process was divided into two stages: (1) pre-training stage and (2) mesh training stage.
At the (1) pre-training stage, five networks (i.e., jigsaw puzzle solver f Jigsaw , rotation predictor f Rot , part in-painter f Inp , discriminator f Disc , and texture network f Tex ) were pre-trained using the corresponding losses defined in Equations (2)–(6), respectively. We pre-trained these networks to constitute them for the self-supervision losses used in the subsequent mesh training stage.
Then, at the (2) mesh training stage, we trained the feature extractor f Feat and mesh estimator f Mesh , which were responsible for the 3D mesh estimation using the loss L defined as follows:
L ( f Feat , f Mesh , f Disc ) = L IM ( f Feat ) + L NR ( f Feat , f Mesh , f Disc )
where the image manipulation loss L IM and neural rendering loss L NR are defined in Equations (7) and (9), respectively. The actual improvement of the 3D mesh reconstruction was obtained during this stage. Furthermore, the discriminator f Disc was further trained by discriminating between real and generated images to provide richer supervision.
The entire training process is summarized in the Algorithm 1.

3.2.2. Pre-Training Stage

The aim of this stage was to train five different networks ( f Jigsaw , f Rot , f Inp , f Disc , and f Tex ) that were used to constitute the losses in the subsequent “mesh training” stage. We trained the networks by running T 1 = 10 epochs of training with the Adam optimizer on the losses L jigsaw , L rot , L inp , L adv and L RGB (Equations (2)–(6)) with learning rates of 0.01 , 5 × 10 4 , 1 × 10 4 , and 1 × 10 4 , and 0.01 , respectively.
Jigsaw puzzle loss L jigsaw . As in [19], the jigsaw puzzling task could be formulated as the standard classification task. We applied the cross-entropy loss L jigsaw for the jigsaw puzzle solver f Jigsaw to make its output close to the permutation order:
L jigsaw ( f Jigsaw | x , x jigsaw ) = c = 1 C y c , jigsaw log ( o ^ c )
where y c , jigsaw is the c-th dimensional element of the one-hot encoded vector from the permutation ground-truth, and o ^ c is the c-th dimensional response of the f Jigsaw network’s output.
Rotation prediction loss L rot . The rotation prediction could be also formulated as the standard classification task as in [20]. Images were rotated with four possible angles θ , 0, 90, 180, and 270, and then mapped to the classification label set { 1 , 2 , 3 , 4 } . To achieve this, the cross-entropy loss L rot was applied on the output of the rotation predictor f Rot :
L rot ( f Rot | x rot ) = c = 1 4 y c , rot log ( r ^ c )
where y c , rot is the c-th dimensional element of the one-hot encoded vector from the rotation ground-truth, and r ^ c is the c-th dimensional response of the f Rot network’s output.
Algorithm 1: The summary of our entire training process
Applsci 11 09724 i001
Part in-painting loss L inp . The part in-painting task could be framed as the image generation task as in [21]. We erased a patch p from the center of the RGB images x to make an image with a hole x hole . The image x hole was inputted to the f Inp ( f Feat ( x hole ) ) , and the part in-painter f Inp was trained to reconstruct the output p ^ = f Inp ( f Feat ( x hole ) ) . We trained the network using both the mean square error (MSE) loss and the GAN-type adversarial loss by the discriminator f Disc to reconstruct a realistic patch p ^ . The MSE loss was enforced to make the reconstructed patch p ^ look similar to the original patch p , and GAN-type adversarial loss was enforced to make the reconstructed image x ^ inp that combined x hole and p ^ look more realistic. f Inp was pre-trained using the following rules:
L inp ( f Inp | x hole ) = λ inp p p ^ 2 2 + λ disc f Disc ( x ^ inp ) ) 1 2 2
where x ^ inp is the in-painted image that fills the hole in x hole with the estimated p ^ , and λ inp and λ disc are set at 0.999 and 0.001 , respectively.
Adversarial loss L adv . The discriminator f Disc was trained to discriminate original images x from images x ^ inp —that is, the combination of x hole and p ^ —outputted from the f Inp using the LSGAN [23] objective, as follows:
L adv ( f Disc | x , x ^ inp ) = f Disc ( x ) ) 1 2 2 + f Disc ( x ^ inp ) 0 2 2 .
RGB rendering loss L RGB . When projecting 3D meshes into the RGB images, RGB values for each vertex of 3D mesh were required. The texture network f Tex was pre-trained to infer such RGB values using the loss, defined as follows:
L RGB ( f Tex | x ) = x ^ render x render 2 2
where x render and x ^ render denote the pseudo ground-truth obtained from f r cnn ( x ) x and rendered meshes f NMR ( f Mesh ( f Feat ( x ) ) , f Tex ( x ) ) , respectively. The operation ⊙ denotes the pixel-wise multiplication and f r cnn denotes the pre-trained Mask-RCNN [58] network. The rationale behind the use of the Mask-RCNN [58] for obtaining the pseudo ground-truth x render is that 3D mesh reconstruction is more challenging than the 2D segmentation mask prediction.

3.2.3. Mesh Training Stage

At this stage, we trained the feature extractor f Feat , mesh estimator f Mesh , and the discriminator f Disc with the aid of pre-trained networks obtained from the pre-training stage. We ran the training for T 2 = 10 epochs using the loss L (Equation (1)) with a learning rate of 5 × 10 6 . The learning rate was decreased using the exponential learning rate decay, whose decay rate was set to 0.99 for each epoch. In the remainder of this subsection, we elaborate the sub-loss of the loss L: L IM and L NR .
Image manipulation loss L IM . This loss consisted of sub-losses reflecting three self-supervision tasks (i.e., solving jigsaw puzzling, rotation prediction and part in-painting tasks). To enrich the feature extractor f Feat using jigsaw puzzling, rotation prediction, and part in-painting tasks, we used the loss L IM that combined losses defined in Equations (2)–(4). We used the same form of losses as Equations (2)–(4) while changing the target training network to f Feat :
L IM ( f Feat ) = L jigsaw ( f Feat | x , x jigsaw ) + L rot ( f Feat | x rot ) + L inp ( f Feat | x hole ) .
Neural rendering loss L NR . This loss reflected the last self-supervision task (i.e., solving image projection task). Given a 2D image x , 3D human meshes are estimated by the sequential combination of f Mesh and f Feat : f Mesh ( f Feat ( x ) ) . Then, we can render estimated 3D meshes m to images x ^ render and segmentation masks x ^ mask using the neural mesh renderer f NMR . L RGB and L m a s k are responsible for making two rendered images (i.e., x ^ render and x ^ mask ) close to their original inputs, respectively. Additional losses L disc and L real are defined to train the discriminator and obtain more realistic images. The combination of four sub-losses is called the neural rendering loss L NR :
L NR ( f Feat , f Mesh , f Disc ) = L RGB ( f Feat , f Mesh | x ) + L real ( f Feat , f Mesh | x ) + L mask ( f Feat , f Mesh | x ) + L disc ( f Disc | x , x ^ render , bg , x ^ inp )
where x ^ render , bg is the image that combines the rendered image x ^ render and a random background image x bg . L RGB ( f Feat , f Mesh | x ) is in the same form as L RGB ( f Tex | x ) in Equation (6); however, the target training network is changed to f Feat and f Mesh . L mask ( f Feat , f Mesh | x ) , L real ( f Feat , f Mesh | x ) , and L disc ( f Disc | x , x ^ render , x ^ inp ) are newly defined as follows:
L real ( f Feat , f Mesh | x ) = f Disc ( x ^ render , bg ) 1 2 2 ,
L mask ( f Feat , f Mesh | x ) = x mask x ^ mask 2 2 ,
L disc ( f Disc | x , x ^ render , x ^ inp ) = f Disc ( x ) ) 1 2 2 + f Disc ( x ^ render , bg ) 2 2 + f Disc ( x ^ inp ) 2 2
where x mask is the binary mask estimated from the Mask-RCNN [58]. Using Equation (11), the discriminator f Disc is trained to discriminate the original images x as real while the rendered images x ^ rendered and in-painted images x ^ inp are determined as fakes.
Joint loss L joints . When 2D or 3D skeleton joints are available, this loss is used to close the keypoints regressed from 3D mesh vertices to their ground-truth locations (in the SMPL [13] model, the model that geometrically regresses 3D skeletons from mesh vertices is accompanied). This is an optional loss that is not used for the self-supervised setting (i.e., Ours (self-sup)); however, we used this loss for the weakly-supervised setting (i.e., ours (weakly-sup)); or for the semi-supervised setting (i.e., ours (semi-sup)), in which either of 2D or 3D skeletons are available, respectively.
Incorporation of different human images. As we constituted our framework with a self-supervised loss that did not require any 2D/3D skeleton annotations, we were able to train our network with any RGB human images outside of our training dataset. We additionally involved 2D images from the MPI-INF-3DHP [59], MPII [60], COCO [61], and LSP [62] datasets, which have been involved in the weakly/semi-supervised settings, plus our own collections of wild YouTube data. When detecting the humans in the images, we used the Mask-RCNN network [58] to inspect if there was a human bounding box or not. The results in Table 5 using “Full w/o Y” (row 4) and “Full” (row 5) datasets can be compared to see the performance improvement by the incorporation of human images, where “Full” and “Full w/o Y” denote the model trained by datasets with and without our own collections of wild YouTube images, respectively. We could observe that the accuracy was improved by the incorporation of human images.

4. Experiments

In this section, we explain the datasets and evaluation method and then analyze the obtained results in a qualitative and quantitative manner. We also provide ablation results by varying the parameters of our framework.

4.1. Dataset

We trained our model using the following datasets: Human3.6M [14], MPI-INF-3DHP [59], MPII [60], COCO [61], LSP [62], and a collection of wild YouTube data consisting of 62k, 96k, 14k, 28k, 1k, and 3k data entities, respectively. We tested using the following datasets: Human3.6M [14], 3DPW [63], and LSP [62].
The Human3.6M dataset is a widely used dataset with paired images and 2D/3D pose annotations. This is an in-studio dataset and consists of four multi-view images; however, we did not use multi-view images for training. Like [5], we used subjects S1, S5, S6, S7, S8 for training and S9 and S11 for testing. We computed the mean per joint position error (MPJPE) and Procrustes analysis-MPJPE (PA-MPJPE) on the Human3.6M test dataset to evaluate the 3D pose estimation results.
MPI-INF-3DHP has 2D images and corresponding 2D/3D skeleton annotation pairs. This is an in-studio dataset and consists of eight multi-view images; however, we did not use multi-view images for training. The MPII dataset and COCO dataset consist of indoor and outdoor images with only 2D annotations.
The LSP dataset is a wild athletic dataset that consists of wild images with 2D annotations. This dataset is used for evaluation, with their (part-)segmentation ground truths secured by [43]. The 3DPW dataset is an in-the-wild outdoor dataset and is used only for evaluation purposes. We further collected images from YouTube of jogging or dancing motions, with diverse backgrounds and viewpoints (the dataset is available in our Github repository).

4.2. Evaluation Method

We used two approaches to evaluate the estimated 3D poses: MPJPE and PA-MPJPE. MPJPE is the mean Euclidean distance between the ground truth and prediction for a joint. PA-MPJPE is Procrustes analysis-MPJPE, which aligns the estimated 3D poses to their ground-truths by a similarity transformation called Procrustes analysis before computing the MPJPE. We also used two approaches to evaluate the estimated 3D shapes: accuracy and F1 score for FG-BG and part-segmentation masks. We measured their accuracy and F1 score by a pixel-wise comparison between estimated masks and their ground truths.

4.3. Results

For the Human3.6M dataset, we compare our method to recent fully/semi/weakly-/self-supervised methods [5,36,37,43,47,48,49,50,64] that output the 3D human poses by protocol-2 in Table 6. Our method outperformed previous methods in all three types of supervision settings (i.e., self/weakly/semi-supervised). Kundu et al. [64]’s method is the competitive self-supervised learning framework using the video priors; however, our method outperformed it even without any video priors. Li et al. [54]’s method is the only baseline that outperforms our method and uses the self-supervised setting. However, the work is not directly comparable to ours, as its setting is easier as it estimates only skeletal joints and using multi-view priors. We provide full 3D meshes inferring both pose and shape and use only 2D images as the supervision signals.
To show the generalization ability of our method, we evaluated it on 3DPW, the data of which were never seen in our training process. For the 3DPW dataset, we compared the accuracy of our method to those of recent state-of-the-art approaches [4,5,9,47,51,64] in Table 7. For the LSP dataset, as the dataset does not contain 3D pose annotations, we only measured the accuracy and F1 score of FG-BG segmentation and six-part segmentation, respectively, for comparison with other algorithms. We compared our method to state-of-the-art optimized-based methods [9,37,65] and regression-based methods [5,49,50,64] in Table 8. Optimization-based methods tend to have better performance on segmentation tasks than regression-based methods; however, our method outperformed most of them even in the self-supervised setting (i.e., Ours (self-sup)); our approach was also recorded as the best when involving more supervisions (i.e., Ours (weakly-sup) and Ours (semi-sup)).
In Figure 3, we visualized qualitative examples of our method on Human3.6M, 3DPW, and LSP datasets. We overlaid predicted meshes, joints, part-segmentation masks, and neural rendered images on the input images. Even though our purpose was not to estimate textures at the testing stage; we could observe that our method faithfully estimates textures for unseen testing images.

4.4. Ablation Study

To analyze the effectiveness of our entire training objective (Equation (1)), we conducted ablation experiments on our losses, as presented in Table 5. In our initial experiment, we used the combination of three images’ manipulation task losses (i.e., jigsaw puzzling, rotation prediction, and part in-painting) and one RGB rendering loss L RGB for the “mesh training stage”. However, given only these four types of losses, we observed that 3D meshes were trained in the wrong way, inferring camera scale parameters to be very small, as this might be an easier way to reduce the RGB rendering loss L RGB . To relieve this, we proposed the addition of more constraints in our loss by the segmentation mask loss L mask , defined in Equation (10): we observed that the uses of loss L mask and the discriminator loss L adv are helpful for preventing this phenomenon. In Table 5, by comparing results from “Ours (self-sup)- L real - L mask ”, “Ours (self-sup)- L real ”, and “Ours (self-sup)”, we could conclude that results improved as more supervision was used. The corresponding qualitative results are also shown in Figure 4. Furthermore, we experimented on different configurations of datasets for training in the same table (Table 5): Rows 3–4 of Table 5 show the accuracy using fewer datasets for training where “H36M” and “Full w/o Y” denote the Human3.6M dataset and full datasets without the collection of wild YouTube data, respectively. “Full” denotes the full training data that we described in Section 4.1, including the collection of wild YouTube data compared to “Full w/o Y”. We can see that involving more image data results in improved accuracy.
We further analyzed the effectiveness of the individual image-level supervision (i.e., L jigsaw , L rot , and L inp ) in the Table 9. We conducted the experiment by applying one of our self-supervision losses for our self/weakly and semi-supervised baselines. For example, “Ours (self-sup)- L rot - L inp ”, “Ours (self-sup)- L jigsaw - L inp ”, and “Ours (self-sup)- L rot - L jigsaw ” denote the network trained without a rotation prediction loss L rot and a part in-painting loss L inp , the network trained without a jigsaw puzzle loss L jigsaw and a part in-painting loss L inp , and the network trained without a rotation prediction loss L rot and a jigsaw puzzle loss L jigsaw for our self-supervision setting, respectively. From these experiments, we can see that (1) none of three losses clearly prevail over the other remaining, and (2) applying all the losses consistently outperforms the results when applying one of the losses. Thus, we need to combine three image-level self-supervisions (i.e., L jigsaw , L rot , and L inp ) together in our main experiments.

5. Conclusions

In this paper, we present a self-supervised learning framework for recovering human 3D meshes from a single RGB image. We proposed the use of the combination of three image manipulation task losses and one neural rendering loss to enrich the feature space and boost the mesh reconstruction accuracy. We also find that the combination of RGB rendering, mask rendering, and adversarial losses is essential to properly achieve image-level self-supervision without using video or multi-view priors. Experiments were conducted on three popular benchmarks, and our algorithm achieved the best accuracy among competitive self/weakly/semi and fully-supervised algorithms. Via these results, we could conclude that single 2D image-level supervision could be a strong supervision method for training 3D human pose and shape estimation. While we showed promising results using the proposed 2D single image-level self-supervision losses, avoiding the effort of collecting other types of supervision signals, some small inconvenience remains in our framework in terms of manually setting hyper-parameters. Future work should explore the possibilities of developing a fully automatic pipeline that resolves this inconvenience by including schemes to select the optimal hyper-parameters for our losses (e.g., tile number in jigsaw puzzle loss, rotation angles for rotation prediction loss and etc.).

Author Contributions

Conceptualization, J.C., W.-H.P. and S.B.; Methodology, J.C. and M.S.; Software, J.C. and M.S.; Validation, J.C., M.S. and S.B.; Formal Analysis, C.L., S.L. (Seongyeong Lee) and S.L. (Seungeun Lee); Writing—Original Draft Preparation, M.S., W.-H.P. and S.B.; Writing—Review and Editing, M.S., W.-H.P. and S.B.; Visualization, C.L., S.L. (Seongyeong Lee) and D.K.; Supervision, W.-H.P. and S.B.; Project Administration, S.B.; Funding Acquisition, S.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a grant from the R&D Program of the Korea Railroad Research Institute, Republic of Korea.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  2. Wei, S.E.; Ramakrishna, V.; Kanade, T.; Sheikh, Y. Convolutional pose machines. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  3. Zhou, X.; Zhu, M.; Leonardos, S.; Derpanis, K.G.; Daniilidis, K. Sparseness meets deepness: 3d human pose estimation from monocular video. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  4. Martinez, J.; Hossain, R.; Romero, J.; Little, J.J. A simple yet effective baseline for 3d human pose estimation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  5. Kanazawa, A.; Black, M.J.; Jacobs, D.W.; Malik, J. End-to-end Recovery of Human Shape and Pose. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  6. Kocabas, M.; Athanasiou, N.; Black, M.J. VIBE: Video inference for human body pose and shape estimation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  7. Sun, Y.; Bao, Q.; Liu, W.; Fu, Y.; Black, M.J.; Mei, T. CenterHMR: Multi-person center-based human mesh recovery. arXiv 2020, arXiv:2008.12272. [Google Scholar]
  8. Guan, S.; Xu, J.; Wang, Y.; Ni, B.; Yang, X. Bilevel online adaptation for out-of-domain human mesh reconstruction. arXiv 2021, arXiv:2103.16449. [Google Scholar]
  9. Bogo, F.; Kanazawa, A.; Lassner, C.; Gehler, P.; Romero, J.; Black, M.J. Keep it SMPL: Automatic estimation of 3D human pose and shape from a single image. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  10. Pavlakos, G.; Choutas, V.; Ghorbani, N.; Bolkart, T.; Osman, A.A.; Tzionas, D.; Black, M.J. Expressive body capture: 3d hands, face, and body from a single image. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  11. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  12. Yang, W.; Ouyang, W.; Wang, X.; Ren, J.; Li, H.; Wang, X. 3d human pose estimation in the wild by adversarial learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5255–5264. [Google Scholar]
  13. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; Black, M.J. SMPL: A skinned multi-person linear model. ACM Trans. Graph. 2015, 34, 248. [Google Scholar] [CrossRef]
  14. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  15. Varol, G.; Laptev, I.; Schmid, C.; Zisserman, A. Synthetic humans for action recognition from unseen viewpoints. arXiv 2019, arXiv:1912.04070. [Google Scholar]
  16. Varol, G.; Romero, J.; Martin, X.; Mahmood, N.; Black, M.J.; Laptev, I.; Schmid, C. Learning from synthetic humans. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  17. Shrivastava, A.; Pfister, T.; Tuzel, O.; Susskind, J.; Wang, W.; Webb, R. Learning from simulated and unsupervised images through adversarial training. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  18. Rad, M.; Oberweger, M.; Lepetit, V. Feature mapping for learning fast and accurate 3D pose inference from synthetic images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  19. Noroozi, M.; Favaro, P. Unsupervised learning of visual representations by solving jigsaw puzzles. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  20. Gidaris, S.; Singh, P.; Komodakis, N. Unsupervised representation learning by predicting image rotations. arXiv 2018, arXiv:1803.07728. [Google Scholar]
  21. Pathak, D.; Krahenbuhl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context encoders: Feature learning by inpainting. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  22. Kato, H.; Ushiku, Y.; Harada, T. Neural 3D mesh renderer. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  23. Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.K.; Wang, Z.; Smolley, S.P. Least squares generative adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  24. Xu, T.; Takano, W. Graph Stacked Hourglass Networks for 3D Human Pose Estimation. arXiv 2021, arXiv:2103.16385. [Google Scholar]
  25. Moreno-Noguer, F. 3d human pose estimation from a single image via distance matrix regression. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  26. Chen, C.H.; Ramanan, D. 3d human pose estimation= 2d pose estimation+ matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7035–7043. [Google Scholar]
  27. Rayat, M.; Hossain, I.; Little, J.J. Exploiting temporal information for 3D pose estimation. arXiv 2017, arXiv:1711.08585. [Google Scholar]
  28. Akhter, I.; Black, M.J. Pose-conditioned joint angle limits for 3D human pose reconstruction. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  29. Ramakrishna, V.; Kanade, T.; Sheikh, Y. Reconstructing 3d human pose from 2d image landmarks. In Proceedings of the 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012. [Google Scholar]
  30. Fish Tung, H.Y.; Harley, A.W.; Seto, W.; Fragkiadaki, K. Adversarial inverse graphics networks: Learning 2D-to-3D lifting and image-to-image translation from unpaired supervision. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  31. Sanzari, M.; Ntouskos, V.; Pirri, F. Bayesian image based 3d pose estimation. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
  32. Zhou, X.; Zhu, M.; Leonardos, S.; Daniilidis, K. Sparse representation for 3D shape estimation: A convex relaxation approach. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1648–1661. [Google Scholar] [CrossRef] [PubMed]
  33. Agarwal, A.; Triggs, B. Recovering 3D human pose from monocular images. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 44–58. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Balan, A.O.; Black, M.J. The naked truth: Estimating body shape under clothing. In Proceedings of the 10th European Conference on Computer Vision, Marseille, France, 12–18 October 2008. [Google Scholar]
  35. Guler, R.A.; Kokkinos, I. Holopose: Holistic 3d human reconstruction in-the-wild. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  36. Omran, M.; Lassner, C.; Pons-Moll, G.; Gehler, P.; Schiele, B. Neural body fitting: Unifying deep learning and model based human pose and shape estimation. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018. [Google Scholar]
  37. Pavlakos, G.; Zhu, L.; Zhou, X.; Daniilidis, K. Learning to estimate 3D human pose and shape from a single color image. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  38. Tan, J.K.V.; Budvytis, I.; Cipolla, R. Indirect Deep Structured Learning for 3D Human Body Shape and Pose Prediction. In Proceedings of the British Machine Vision Conference (BMVC), London, UK, 4–7 September 2017; BMVA Press: Durham, UK, 2017; pp. 15.1–15.11. [Google Scholar]
  39. Tung, H.Y.F.; Tung, H.W.; Yumer, E.; Fragkiadaki, K. Self-supervised learning of motion capture. arXiv 2017, arXiv:1712.01337. [Google Scholar]
  40. Saito, S.; Huang, Z.; Natsume, R.; Morishima, S.; Kanazawa, A.; Li, H. Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  41. Saito, S.; Simon, T.; Saragih, J.; Joo, H. Pifuhd: Multi-level pixel-aligned implicit function for high-resolution 3d human digitization. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  42. Huang, Y.; Bogo, F.; Lassner, C.; Kanazawa, A.; Gehler, P.V.; Romero, J.; Akhter, I.; Black, M.J. Towards accurate marker-less human shape and pose estimation over time. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017. [Google Scholar]
  43. Lassner, C.; Romero, J.; Kiefel, M.; Bogo, F.; Black, M.J.; Gehler, P.V. Unite the people: Closing the loop between 3d and 2d human representations. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  44. Kolotouros, N.; Pavlakos, G.; Black, M.J.; Daniilidis, K. Learning to reconstruct 3D human pose and shape via model-fitting in the loop. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  45. Lin, K.; Wang, L.; Liu, Z. End-to-end human pose and mesh reconstruction with transformers. arXiv 2021, arXiv:2012.09760. [Google Scholar]
  46. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Long Beach, CA, USA, 2017; pp. 6000–6010. [Google Scholar]
  47. Kanazawa, A.; Zhang, J.Y.; Felsen, P.; Malik, J. Learning 3d human dynamics from video. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  48. Arnab, A.; Doersch, C.; Zisserman, A. Exploiting temporal context for 3D human pose estimation in the wild. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  49. Kolotouros, N.; Pavlakos, G.; Daniilidis, K. Convolutional mesh regression for single-image human shape reconstruction. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  50. Pavlakos, G.; Kolotouros, N.; Daniilidis, K. Texturepose: Supervising human mesh estimation with texture consistency. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
  51. Dabral, R.; Mundhada, A.; Kusupati, U.; Afaque, S.; Sharma, A.; Jain, A. Learning 3d human pose from structure and motion. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018. [Google Scholar]
  52. Iqbal, U.; Molchanov, P.; Kautz, J. Weakly-Supervised 3D Human Pose Learning via Multi-view Images in the Wild. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  53. Chen, C.H.; Tyagi, A.; Agrawal, A.; Drover, D.; Stojanov, S.; Rehg, J.M. Unsupervised 3d pose estimation with geometric self-supervision. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  54. Li, Y.; Li, K.; Jiang, S.; Zhang, Z.; Huang, C.; Da Xu, R.Y. Geometry-driven self-supervised method for 3d human pose estimation. AAAI Conf. Artif. Intell. 2020, 34, 11442–11449. [Google Scholar] [CrossRef]
  55. Kocabas, M.; Karagoz, S.; Akbas, E. Self-supervised learning of 3d human pose using multi-view geometry. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  56. Rhodin, H.; Salzmann, M.; Fua, P. Unsupervised geometry-aware representation for 3d human pose estimation. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018. [Google Scholar]
  57. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  58. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017. [Google Scholar]
  59. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3d human pose estimation in the wild using improved cnn supervision. In Proceedings of the 2017 International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017. [Google Scholar]
  60. Andriluka, M.; Pishchulin, L.; Gehler, P.; Schiele, B. 2d human pose estimation: New benchmark and state of the art analysis. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  61. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014. [Google Scholar]
  62. Johnson, S.; Everingham, M. Clustered Pose and Nonlinear Appearance Models for Human Pose Estimation. In Proceedings of the British Machine Vision Conference, BMVC 2010, Aberystwyth, UK, 31 August–3 September 2010. [Google Scholar]
  63. von Marcard, T.; Henschel, R.; Black, M.J.; Rosenhahn, B.; Pons-Moll, G. Recovering accurate 3d human pose in the wild using imus and a moving camera. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018. [Google Scholar]
  64. Kundu, J.N.; Rakesh, M.; Jampani, V.; Venkatesh, R.M.; Babu, R.V. Appearance Consensus Driven Self-supervised Human Mesh Recovery. In Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  65. Varol, G.; Ceylan, D.; Russell, B.; Yang, J.; Yumer, E.; Laptev, I.; Schmid, C. Bodynet: Volumetric inference of 3d human body shapes. In Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018. [Google Scholar]
  66. Chang, I.; Park, M.G.; Kim, J.; Yoon, J.H. Multi-View 3D Human Pose Estimation with Self-Supervised Learning. In Proceedings of the 2021 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju Island, Korea, 13–16 April 2021. [Google Scholar]
Figure 1. Example image-level self-supervision: We applied three types of image manipulation and one image projection task to train our human mesh estimation network. Each row shows examples from the Human3.6M dataset, and each column corresponds to (a) input RGB images, (b) shuffled patches used for the jigsaw puzzling task, (c) rotated RGB images used for rotation prediction task, (d) output images from part inpainting task, and (e) neural rendered images from estimated 3D meshes and textures, respectively.
Figure 1. Example image-level self-supervision: We applied three types of image manipulation and one image projection task to train our human mesh estimation network. Each row shows examples from the Human3.6M dataset, and each column corresponds to (a) input RGB images, (b) shuffled patches used for the jigsaw puzzling task, (c) rotated RGB images used for rotation prediction task, (d) output images from part inpainting task, and (e) neural rendered images from estimated 3D meshes and textures, respectively.
Applsci 11 09724 g001
Figure 2. A schematic diagram of the proposed image-level self-supervised 3D human pose and shape estimation framework. First, we prepared three types of images x jigsaw , x hole , and x rot from the original image x . Second, we pre-trained four networks denoted as green boxes (i.e., f Jigsaw , f Inp , f Rot , and f Tex ) and a network in a purple box (i.e., f Disc ) using five corresponding losses (i.e., L jigsaw , L inp , L rot , L RGB , and L adv , respectively). Two networks f NMR and f r cnn with pink boxes were fixed after employing the implementation of [22] and initializing weights from [58], respectively. Finally, in the mesh training stage, we trained f Feat and f Mesh in blue boxes, which were used to infer the 3D human meshes from 2D images. The discriminator f Disc in the purple box was also further trained during the mesh training stage. Blue, black, and green arrows denote routes used for testing, training, and supervision signals, respectively. Optional loss L joints was not used for the self-supervised setting (i.e., Ours (self-sup)); it was used via 2D skeletons for the weakly-supervised setting (i.e., Ours (weakly-sup)) and was used via both 2D and 3D skeletons for the semi-supervised setting (i.e., Ours (semi-sup)).
Figure 2. A schematic diagram of the proposed image-level self-supervised 3D human pose and shape estimation framework. First, we prepared three types of images x jigsaw , x hole , and x rot from the original image x . Second, we pre-trained four networks denoted as green boxes (i.e., f Jigsaw , f Inp , f Rot , and f Tex ) and a network in a purple box (i.e., f Disc ) using five corresponding losses (i.e., L jigsaw , L inp , L rot , L RGB , and L adv , respectively). Two networks f NMR and f r cnn with pink boxes were fixed after employing the implementation of [22] and initializing weights from [58], respectively. Finally, in the mesh training stage, we trained f Feat and f Mesh in blue boxes, which were used to infer the 3D human meshes from 2D images. The discriminator f Disc in the purple box was also further trained during the mesh training stage. Blue, black, and green arrows denote routes used for testing, training, and supervision signals, respectively. Optional loss L joints was not used for the self-supervised setting (i.e., Ours (self-sup)); it was used via 2D skeletons for the weakly-supervised setting (i.e., Ours (weakly-sup)) and was used via both 2D and 3D skeletons for the semi-supervised setting (i.e., Ours (semi-sup)).
Applsci 11 09724 g002
Figure 3. Qualitative examples of three databases (Human3.6M, 3DPW, and LSP). (a) Input image; (bd), (eg), and (hj) denote our results obtained from self/weakly and semi-supervised models, respectively. (b,e,h) show images with joints and 3D meshes, (c,f,i) show images with part-segmentation masks, and (d,g,j) show neural mesh rendered images, respectively.
Figure 3. Qualitative examples of three databases (Human3.6M, 3DPW, and LSP). (a) Input image; (bd), (eg), and (hj) denote our results obtained from self/weakly and semi-supervised models, respectively. (b,e,h) show images with joints and 3D meshes, (c,f,i) show images with part-segmentation masks, and (d,g,j) show neural mesh rendered images, respectively.
Applsci 11 09724 g003
Figure 4. Qualitative examples depending on losses: (a) input, (b) results from ‘Ours (self-sup)- L real - L mask ’, (c) results from ‘Ours (self-sup)- L real ’, and (d) results from ‘Ours (self-sup)’. Results become better involving more supervision.
Figure 4. Qualitative examples depending on losses: (a) input, (b) results from ‘Ours (self-sup)- L real - L mask ’, (c) results from ‘Ours (self-sup)- L real ’, and (d) results from ‘Ours (self-sup)’. Results become better involving more supervision.
Applsci 11 09724 g004
Table 1. Architecture of jigsaw puzzle solver f Jigsaw . Inputs are concatenated feature maps of four tiles of x jigsaw and an original image x , obtained from f Feat .
Table 1. Architecture of jigsaw puzzle solver f Jigsaw . Inputs are concatenated feature maps of four tiles of x jigsaw and an original image x , obtained from f Feat .
LayerOperationKernelDimensionality
Input: Feature map-10,240 × 1
1Linear + ReLU + Dropout (0.5)- 2048 × 1
2Linear- 24 × 1
Table 2. Architecture of rotation predictor f Rot . Input is a feature map of the rotated image from f Feat ( x rot ) .
Table 2. Architecture of rotation predictor f Rot . Input is a feature map of the rotated image from f Feat ( x rot ) .
LayerOperationKernelDimensionality
Input: Feature map-10,240 × 1
1Linear + ReLU + Dropout(0.5)- 2048 × 1
2Linear- 24 × 1
Table 3. Architecture of part in-painter f Inp . Input is a feature map of the center hole image from f Feat ( x hole ) .
Table 3. Architecture of part in-painter f Inp . Input is a feature map of the center hole image from f Feat ( x hole ) .
LayerOperationKernelDimensionality
Input: Feature map- 2048 × 1
1ConvT. + B.N. + ReLU 4 × 4 4 × 4 × 512
2ConvT. + B.N. + ReLU 4 × 4 8 × 8 × 256
3ConvT. + B.N. + ReLU 4 × 4 16 × 16 × 128
4ConvT. + B.N. + ReLU 4 × 4 32 × 32 × 64
5ConvT. + Tanh 4 × 4 64 × 64 × 3
Table 4. Architecture of mesh estimator f Mesh .
Table 4. Architecture of mesh estimator f Mesh .
LayerOperationKernelDimensionality
Input: Feature map + Pose, Shape, Camera param. - 2205
1Linear + Dropout (0.5)-1024
2Linear + Dropout (0.5)-1024
3-1Linear-72
3-2Linear-10
3-3Linear-3
4SMPL layer [5]- 6890 × 3
Table 5. Ablation study on Human3.6M to analyze the effectiveness of the objective functions (first 2 rows) and results of weakly-supervised and semi-supervised approaches (last 2 rows). We also conducted an ablation study depending on the different training dataset compositions: “Full” denotes the full training dataset that we described in Section 4.1. Rows 3–4 show the ablation results using different training datasets, where “H36M” and “Full w/o Y” denote the Human3.6M dataset and full datasets without the collection of wild YouTube data, respectively. We use protocol-2 for “H36M”.
Table 5. Ablation study on Human3.6M to analyze the effectiveness of the objective functions (first 2 rows) and results of weakly-supervised and semi-supervised approaches (last 2 rows). We also conducted an ablation study depending on the different training dataset compositions: “Full” denotes the full training dataset that we described in Section 4.1. Rows 3–4 show the ablation results using different training datasets, where “H36M” and “Full w/o Y” denote the Human3.6M dataset and full datasets without the collection of wild YouTube data, respectively. We use protocol-2 for “H36M”.
MethodsDatasetMPJPE(↓)PA-MPJPE(↓)
Ours (self-sup)— L real L mask Full96.862.1
Ours (self-sup)— L real Full95.760.3
Ours (self-sup) H36M90.455.4
Ours (self-sup) Full w/o Y86.554.9
Ours (self-sup) Full84.254.4
Ours (weakly-sup) Full80.752.7
Ours (semi-sup) Full65.844.9
Table 6. Evaluation with Human3.6M(Protocol-2). Methods in the first 10 rows use equivalent 2D and 3D pose supervision. Methods in rows 11–14 and rows 15–18 are weakly-supervised and self-supervised approaches, respectively. * indicates methods that output only 3D joints, not 3D meshes.
Table 6. Evaluation with Human3.6M(Protocol-2). Methods in the first 10 rows use equivalent 2D and 3D pose supervision. Methods in rows 11–14 and rows 15–18 are weakly-supervised and self-supervised approaches, respectively. * indicates methods that output only 3D joints, not 3D meshes.
MethodsPA-MPJPE(↓)
Fully-supervised & Semi-supervised
Lassner et al. [43]93.9
Pavlakos et al. [37]75.9
Omran et al. [36]59.9
HMR [5]56.8
Temporal-HMR [47]56.9
Arnab et al. [48]54.3
Kolotouros et al. [49]50.1
TexturePose [50]49.7
Kundu et al. [64]48.1
Ours (semi-sup) 44.9
Weakly-supervised
HMR unpaired [5]66.5
Kundu et al. [64]58.2
Iqbal et al. [52]54.5
Ours (weakly-sup) 52.7
Self-supervised
Kundu et al. [64]90.5
Chang et al. [66] (multi-view-sup)77.0
Kundu et al. [64] (multi-view-sup)74.1
Ours (self-sup) 54.4
* Li et al. [54] (multi-view-sup)45.7
Table 7. Evaluation with wild 3DPW dataset in a fully-unseen setting. Unlike Temporal-HMR, we do not use any temporal supervision. Methods in the first seven rows use equivalent 2D and 3D pose supervision.
Table 7. Evaluation with wild 3DPW dataset in a fully-unseen setting. Unlike Temporal-HMR, we do not use any temporal supervision. Methods in the first seven rows use equivalent 2D and 3D pose supervision.
MethodsMPJPE(↓)PA-MPJPE(↓)
Martinez et al. [4]-157.0
SMPLify [9]199.2106.1
TP-NET [51]163.792.3
HMR [5]130.076.7
Temporal-HMR [47]127.180.1
Kundu et al. [64] (semi-sup)125.878.2
Ours (semi-sup) 106.066.0
Kundu et al. [64] (weakly-sup)153.489.8
Ours (weakly-sup) 107.469.1
Kundu et al. [64] (self-sup)187.1102.7
Ours (self-sup) 124.788.5
Table 8. Evaluation of FG-BG and six-part segmentation with LSP test set. First 4 rows: optimization-based methods. Last 10 rows: regression-based methods.
Table 8. Evaluation of FG-BG and six-part segmentation with LSP test set. First 4 rows: optimization-based methods. Last 10 rows: regression-based methods.
MethodsFG-BG Seg.Part Seg.
Acc. (↑)F1 (↑)Acc. (↑)F1 (↑)
Optimization-based
SMPLify oracle [9]92.170.8888.820.67
SMPLify [9]91.890.8887.710.64
SMPLify on [37]92.170.8888.240.64
Bodynet [65]92.750.84--
Fully-supervised & Semi-supervised
HMR [5]91.670.8787.120.60
Kolotouros et al. [49]91.460.8788.690.66
TexturePose [50]91.820.8789.000.67
Kundu et al. [64]91.840.8789.080.67
Ours (semi-sup)93.280.9089.900.70
Weakly-supervised
HMR unpaired [5]91.300.8687.000.59
Kundu et al. [64]91.700.8787.120.60
Ours (weakly-sup)93.310.9089.860.70
Self-supervised
Kundu et al. [64]91.460.8687.260.64
Ours (self-sup)92.620.8987.720.63
Table 9. Ablation study on Human3.6M, 3DPW, and LSP to analyze the effectiveness of image manipulation task losses.
Table 9. Ablation study on Human3.6M, 3DPW, and LSP to analyze the effectiveness of image manipulation task losses.
MethodsHuman3.6M3DPWLSP
BG-FG Seg.Part Seg.
MPJPE (↓)PA-MPJPE (↓)MPJPE (↓)PA-MPJPE (↓)Acc. (↑)Acc. (↑)
Ours (self-sup) - L rot - L inp 90.155.5126.789.592.8088.15
Ours (self-sup) - L jigsaw - L inp 93.160.0137.598.492.8688.06
Ours (self-sup) - L rot - L jigsaw 91.258.2135.998.792.7687.70
Ours (self-sup) 84.254.4124.788.92.6287.72
Ours (weakly-sup) - L rot - L inp 83.753.0112.169.393.3989.78
Ours (weakly-sup) - L jigsaw - L inp 86.354.7110.568.293.3789.76
Ours (weakly-sup) - L rot - L jigsaw 84.153.9109.668.293.4589.88
Ours (weakly-sup) 80.752.7107.469.193.3189.86
Ours (semi-sup) - L rot - L inp 67.746.0105.864.993.1889.84
Ours (semi-sup) - L jigsaw - L inp 68.445.9108.065.793.2389.79
Ours (semi-sup) - L rot - L jigsaw 67.646.5106.965.093.2289.88
Ours (semi-sup) 65.844.9106.066.093.2889.90
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cha, J.; Saqlain, M.; Lee, C.; Lee, S.; Lee, S.; Kim, D.; Park, W.-H.; Baek, S. Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation. Appl. Sci. 2021, 11, 9724. https://doi.org/10.3390/app11209724

AMA Style

Cha J, Saqlain M, Lee C, Lee S, Lee S, Kim D, Park W-H, Baek S. Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation. Applied Sciences. 2021; 11(20):9724. https://doi.org/10.3390/app11209724

Chicago/Turabian Style

Cha, Junuk, Muhammad Saqlain, Changhwa Lee, Seongyeong Lee, Seungeun Lee, Donguk Kim, Won-Hee Park, and Seungryul Baek. 2021. "Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation" Applied Sciences 11, no. 20: 9724. https://doi.org/10.3390/app11209724

APA Style

Cha, J., Saqlain, M., Lee, C., Lee, S., Lee, S., Kim, D., Park, W. -H., & Baek, S. (2021). Towards Single 2D Image-Level Self-Supervision for 3D Human Pose and Shape Estimation. Applied Sciences, 11(20), 9724. https://doi.org/10.3390/app11209724

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop