Next Article in Journal
Assessing Wear Characteristics of Sprayable, Diacetylene-Containing Sensor Formulations
Previous Article in Journal
TRPV4—A Multifunctional Cellular Sensor Protein with Therapeutic Potential
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Geodesic-Based Maximal Cliques Search for Non-Rigid Human Point Cloud Registration

School of Aeronautics and Astronautics, Sun Yat-Sen University, Guangzhou 510275, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(21), 6924; https://doi.org/10.3390/s24216924
Submission received: 23 September 2024 / Revised: 19 October 2024 / Accepted: 25 October 2024 / Published: 29 October 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Non-rigid point cloud registration holds significant importance for human body pose analysis in the fields of sports, medicine, gaming, etc. In this paper, we propose a non-rigid point cloud registration algorithm based on geodesic distance measurement, which can improve the accuracy of the registration for matching point pairs during non-rigid deformations. Firstly, a graph is constructed for two sets of point clouds using geodesic distance measurement considering that geodesic distance changes minimally during non-rigid deformation of the human body, which can preserve the point cloud matching information between corresponding points. Furthermore, a maximal clique search is employed to find combinations of matching pairs between point clouds. Finally, by driving the human body model parameters, sparse matching pairs are overlapped as much as possible to achieve non-rigid point cloud registration of the human body. The accuracy of the proposed algorithm is verified with FAUST and CAPE datasets.

1. Introduction

The rapid advancement of 3D sensing technologies, such as range sensors, depth cameras, and multi-view reconstructions, has revolutionized the way human bodies are captured and represented in digital form. However, effectively utilizing these point clouds for applications such as biomechanics, virtual reality, and ergonomic assessments necessitates precise alignment and comparison of data captured from different instances or subjects. Non-rigid registration of point clouds is such a fundamental process in computer vision, utilized to align or match two or more sets of point cloud data [1]. Unlike rigid registration, which presumes that the shapes being aligned remain unchanged aside from translations and rotations, non-rigid registration must deal with complex deformations and variations in shape [2].
Since non-rigid deformation is point-wise, the transformation between two unstructured point clouds can be highly arbitrary. To address this complexity, one category of methods makes reasonable assumptions based on physical characteristics or domain knowledge to guide the estimation of deformation. These assumptions can enhance the accuracy and stability of registration while reducing the search space for potential solutions. Examples of such assumptions include the local rigidity assumption [3,4], the continuity assumption [5,6], the As-Rigid-As-Possible (ARAP) assumption [7,8,9], and the Coherent Point Drift assumption [6,10,11]. However, these assumptions have inherent limitations and are only applicable to specific scenarios, making these methods highly dependent on data quality and particularly suitable for handling large deformations.
Another category of methods addresses the 3D registration problem by generating correspondences free of outliers. These correspondences are crucial for defining non-rigid deformations, as they track how different parts of an object move and change shape over time. Typically, this approach involves two sequential stages: first, producing sparse correspondences, and second, performing optimization. The optimized energy function usually encompasses alignment components that measure the divergence between the two point clouds after transformation, alongside regularization components that ensure the transformation field remains smooth [12,13].
However, a significant limitation of these methods is their heavy reliance on the accuracy of the initial sparse correspondences. In the context of rigid point clouds, correspondence without outliers has been extensively researched due to the less stringent requirements of rigid transformations. Techniques involving maximum and maximal cliques with Euclidean distance can generate more accurate correspondences [14,15,16]. In these approaches, the preliminary set of correspondences is conceptualized as a compatibility graph, where each vertex represents a single correspondence and each edge signifies a pair of compatible correspondences. Subsequently, maximum or maximal cliques are identified within the graph, and a node-guided clique vetting procedure is employed to meticulously associate each graph vertex with the corresponding maximal clique that contains it.
In this paper, we introduce a three-dimensional human pose estimation method based on non-rigid point cloud registration. The fundamental assumption of our approach is that the topological structure among data points on the human body surface remains unchanged during motion and deformation, resulting in only minor variations in the geodesic distances between any two points. Geodesic distance represents the shortest path between two points on a curved surface or manifold. This metric is particularly crucial for capturing relationships within point clouds, especially during non-rigid deformations. Unlike Euclidean distance, which may overlook the surface complexity of deformable shapes, geodesic distance preserves the intrinsic geometry of the underlying manifold. This property makes it highly effective in accurately representing the relationships between points in scenarios involving non-rigid deformation, such as when modeling articulated objects or deformable surfaces. Our algorithm comprises two main components: a second-order spatial compatibility graph construction method based on geodesic distance measurements and a maximal clique search algorithm to obtain sparse correspondences.
For both the human body model and the point cloud data to be registered, we construct a second-order spatial compatibility graph where geodesic distances serve as weights. This graph effectively associates matching pairs that adhere to our core assumption among the initial correspondences. The maximal clique search algorithm is then employed to generate multiple sets of matching pairs that satisfy the constraint conditions. To refine these matching pairs, we utilize a nonlinear optimization algorithm derived from inverse dynamics to construct a loss function between the model and data matching pairs. This optimization process evaluates multiple outcomes by comparing the root mean square distance errors of neighboring points of the optimized results and the actual data. Ultimately, this procedure yields an effective set of matching pairs and accurate human body pose estimations. The workflow of the algorithm is illustrated in Figure 1.
The proposed geodesic-based maximal clique algorithm offers a distinctive advantage in non-rigid registration. By leveraging geodesic distances, which remain relatively stable even under large deformations, our algorithm improves the accuracy of sparse correspondences. The maximal clique search further enhances this by ensuring that the selected correspondences are highly reliable, addressing a critical limitation of current methods, especially during large point cloud deformation. In summary, the main contributions of our work are as follows:
  • We propose a novel non-rigid point cloud registration method that leverages maximal clique searching within constructed graphs based on geodesic distance measurements. This innovative approach enables the accurate identification of correct correspondences in human point clouds by effectively filtering initial correspondences derived from simple feature matching. By utilizing maximal cliques, our method ensures robust correspondence selection, enhancing the precision and reliability of the registration process.
  • We have designed an efficient, model-based optimization procedure that automatically filters the matching pair sets generated by maximal cliques. This optimization framework significantly enhances the robustness of our non-rigid registration method, particularly in scenarios involving large motions. By systematically refining the correspondence pairs, our approach minimizes errors and maintains high registration accuracy, even under substantial deformations and dynamic movements.

2. Related Works

Estimating human body poses from depth images or scanned point cloud data and performing point cloud registration have long been focal points of research in computer vision [17,18]. The data acquired from depth cameras and scanners fall under the category of unstructured 3D point cloud data. These unordered point clouds lack semantic information and cannot effectively describe the observed scenes or human body states. To extract meaningful information from these data, it is essential to align these point clouds with parameterized human body models, such as the SCAPE model [19], Frankenstein model [20], SMPL model [21], SMPL-X model [22], and others. Point-cloud-based human body pose estimation and reconstruction can typically be classified into two main categories: those based on geometric optimization [23] and those based on deep learning [24].

2.1. Geometric Optimization Algorithms

Geometric optimization algorithms iteratively reduce the matching error between two sets of point clouds by adjusting model parameters. These algorithms are characterized by their reliance on specifically designed objective functions and constraints based on various assumptions. They utilize optimization techniques to align models with point clouds and leverage different types of prior knowledge to reduce the computational complexity associated with high-dimensional point cloud data. Given two point clouds—one as the source and the other as the target—the goal of non-rigid registration is to find an appropriate transformation that deforms the source shape to align with the target shape [25]. Without prior model constraints, the transformation between two unstructured point clouds can be highly arbitrary. Therefore, reasonable assumptions based on physical characteristics or domain knowledge are often necessary to guide the transformation estimation, improving registration accuracy and stability while reducing the search space for potential solutions.

2.1.1. Local Rigidity Assumption

The local rigidity assumption posits that, during deformation, point clouds remain relatively consistent within localized regions [3]. This assumption reduces the computational complexity caused by the large number of points in the cloud by constructing a deformation graph that binds local regions, significantly lowering computational overhead. For instance, the Non-Rigid Iterative Closest Point (N-ICP) algorithm [4] employs embedded deformation (ED) to describe deformations in different parts of the point cloud. Nodes in the deformation graph are adjusted through affine transformations to capture local point cloud position adjustments and deformations. This deformation model is integrated into the optimization process and combined with point-to-point correspondences between point clouds to gradually align them. However, N-ICP is sensitive to the initial estimate and often requires an initial alignment through rigid registration to achieve effective non-rigid registration, especially when dealing with significant global deformations.

2.1.2. Continuity Assumption

The continuity assumption in non-rigid point cloud registration considers deformations between point clouds as continuous, facilitating the establishment of point correspondences and the estimation of deformation models for better alignment. The Coherent Point Drift (CPD) algorithm utilizes the principle of motion coherence in point cloud transformation, describing point correspondence using a Gaussian mixture model [5]. Through iterative optimization, CPD effectively handles point cloud deformations and non-rigid transformations. In 2021, Hirose reinterpreted the CPD algorithm within a Bayesian framework, ensuring convergence and eliminating the uncertainties present in the original algorithm [6]. This reinterpretation represents motion coherence as a prior distribution, providing a clearer and more explicit interpretation of its parameters.

2.1.3. As-Rigid-As-Possible (ARAP) Assumption

The As-Rigid-As-Possible (ARAP) assumption [26] aims to maintain rigidity as much as possible during deformation, particularly for articulated deformations such as those of the human body [7]. The ARAP assumption minimizes excessive deformations and ensures that the relative positions between points on the same object do not change significantly. This assumption is particularly useful for point cloud registration of non-rigid shapes, especially when registering different poses of the human body, as the ARAP assumption ensures that the relative positions of the skeleton and key points remain consistent.
Geometric-optimization-based non-rigid registration methods offer flexibility and adaptability when dealing with various non-rigid deformations. They often do not require extensive prior knowledge or model construction, making them useful in specific scenarios. However, due to the lack of model constraints, these methods have high data quality requirements and are sensitive to noise. The choice of method depends on the specific application requirements and the quality of available data.

2.1.4. Deep Learning Algorithms

Deep-learning-based 3D human body pose estimation using point clouds typically involves two main steps: extracting point cloud features with deep neural networks and regressing 3D joint coordinates or shape and pose parameters for human body models. Human body models, such as the SCAPE [19], Frankenstein [20], SMPL [21], and SMPL-X [22] models, are widely utilized as strong priors in point cloud registration and human pose estimation. These models, driven by predefined joint points, significantly reduce the degrees of freedom compared to non-model-based non-rigid point cloud deformations.
In Li et al.’s method [27] for estimating a 3D human body pose from a single RGB-D image, they first estimate 2D human body poses using a convolutional neural network (CNN) and derive 3D joint coordinates from the corresponding depth map. Subsequently, they optimize the SMPL model parameters and pose parameters by constructing loss functions based on both 2D and 3D joint point matching errors. However, the highly nonlinear relationship between the SMPL model parameters—which describe the local relative motion of human joints in an axis–angle format—and the point cloud data poses a significant challenge for directly regressing SMPL parameters from features extracted by backbone networks. To bridge this gap, several approaches have been proposed to establish an indirect relationship between point clouds and joint parameters.
For the VoteHMR algorithm introduced by Liu et al. [28], inspiration was drawn from the successful application of Hough voting in point cloud object detection. This method extracts sparse points from visible regions and establishes their relationships with skeletal joints. Cascaded joints are then used to obtain occlusion-aware joint features, which explicitly encode the human body pose and implicitly capture local geometric relationships. These features facilitate the estimation of shape parameters. Similarly, Wang et al. [29] proposed the Piecewise Transformation Fields (PTF) method, which learns to map any point from the pose space to a normalized pose space, thereby establishing local point alignment features. These features are subsequently used to estimate rotation parameters for a parametric SMPL model, offering greater accuracy and robustness compared to directly regressing SMPL parameters from global features.
Jiang et al. [30] introduced perceptual awareness of the human skeleton into the deep neural network framework by learning the mapping from point cloud features to skeleton features and then the mapping from skeleton features to SMPL parameters. Since point cloud features extracted by PointNet++ are unordered, the extraction process needs to follow the order of the skeleton. To achieve this, an attention module (AM) is incorporated to encode the spatial relationships of point cloud features. Additionally, a local graph aggregation (GA) module based on graph convolution is adopted to mine the local contextual relationships of point clouds without increasing memory requirements. Finally, a Skeleton Graph Module (SGM) based on graph convolution is employed to better learn features related to joint rotation parameters.
In Ying et al.’s work [31], RGB images are used to assist depth images in 3D human pose estimation. A 2D heatmap of human joints is first estimated and fused with a depth image to obtain a point cloud enriched with joint probability distribution information. This point cloud is then downsampled and input into a 3D network for learning and generating point-by-point features. A dense prediction module subsequently obtains the 3D coordinates of the human body’s joints.
Despite their effectiveness, these deep-learning-based methods for point cloud registration and human pose estimation require large numbers of training data, which limits their transferability to different datasets or real-world scenarios with varying data distributions.

3. Methods

3.1. Summary

Given a human model and a point cloud obtained from 3D scanning or multi-view RGB-D cameras, the goal of non-rigid point cloud registration is to align the two sets of point clouds—the source point cloud and the target point cloud—to structure the unorganized and unordered data effectively. Let P s and P t denote the two sets of points, respectively. The non-rigid point cloud registration method proposed in this paper employs a geodesic-based maximal clique search algorithm to establish sparse correspondences between articulated and deformed human body point clouds. This method leverages a manipulable human body model to construct an inverse kinematics loss function based on these sparse matching points. Through iterative optimization, the registration between the model and the data point cloud is achieved, resulting in an accurate estimation of the corresponding human body pose.
Specifically, the proposed method comprises two main components:
(1)
Maximal Clique Search for Sparse Correspondence Generation
The first component involves identifying maximal cliques between the source and target point clouds, thereby generating multiple sets of matching pairs that adhere to geodesic distance constraints. This process can be further divided into three steps:
  • Feature Extraction and Initial Matching: Extract features from both the source and target point clouds and perform feature matching to generate initial correspondences.
  • Graph Construction with Second-Order Spatial Compatibility: Construct a compatibility graph using second-order spatial relationships among the matching pairs. The spatial compatibility measure is based on geodesic distances between point pairs.
  • Maximal Clique Search: Search for maximal cliques within the constructed graph that satisfy the geodesic distance constraints, thereby generating robust sets of matching pairs.
(2)
Inverse Kinematics Optimization for Pose Estimation
The second component quantitatively evaluates the generated maximal cliques by utilizing the matching pairs to constrain the inverse kinematics of the human body model. This involves obtaining the pose parameters of the human body model and assessing the alignment between the deformed model and the target data point cloud. The optimization process identifies the best-fitting correspondences, ensuring accurate pose estimation.

3.2. Geodesic-Based Maximal Cliques Search

3.2.1. Correspondence Initialization

(1)
Feature Extraction
First, for the given two sets of point clouds, it is necessary to detect key points or key features from the source and target point clouds. These key points are typically local maxima, curvature extrema, corners, edges, or other distinctive points within the point clouds. Commonly used point cloud features include Fast Harris 3D Point Features (FHPF), Signature of Histograms of Orientations (SHOT), and features extracted using deep learning.
(2)
Feature Matching
Feature matching involves establishing correspondences between key points in the source and target point clouds, starting from the extracted features from the point clouds. Typically, nearest neighbor (NN) algorithms or other matching algorithms are used to identify points in the two point clouds with similar feature descriptors. Once matching key point pairs are found, correspondences are established between the source and target point clouds. Each point in the source point cloud is associated with a point in the target point cloud. In this paper, we use the FHPF to establish the initial features, and the distance between the features of each pair is evaluated to determine the similarity of the point pair.

3.2.2. Second-Order Spatial Compatibility with Geodesic Distance Measurement

From Section 3.1, we obtain the initial point cloud correspondences C = ( c i ) i = 1 , 2 , , N , where c i = ( p i s , p i t ) represents a set of correspondences between the source and target point clouds. In order to select the correct correspondences within the point cloud relationships, it is necessary to define an affinity between the various correspondences, resulting in the formation of a graph structure.
In graph structures, undirected graphs are typically represented as G = ( V , E ) , where the vertex set is denoted as V = { v 1 , v 2 , v n } , indicating the presence of n vertices in the graph. The edge set is defined as E = { ( v i , v j ) | v i , v j V ) } and consists of pairs of vertices. For any two vertices v i and v j in the graph, if there is an edge connecting them, then ( v i , v j ) E ; otherwise, ( v i , v j ) E . In an undirected graph, edges have no direction, which means that if ( v i , v j ) E , then ( v j , v i ) E , and both ( v i , v j ) and ( v j , v i ) refer to the same edge.
(1)
Geodesic Distance Measurement
In point cloud registration, each correspondence c i represents a vertex in the undirected graph, and the relationships between vertices can be defined using different distance metrics. In rigid point cloud registration, common distance metrics include Euclidean distance, normal vector distance, and geodesic distance.
For rigid transformations, Euclidean distance and normal vector distance are used as constraints in point cloud registration, similar to the metrics used in the Iterative Closest Point (ICP) algorithm. However, these metrics are not suitable for non-rigid point cloud registration, particularly for human body deformations, as they fail to capture the non-rigid deformation relationships. The complex movements of the human body lead to local relative deformations, as shown in Figure 2, which increase the number of parameters required for registration. In this case, the Euclidean distances between corresponding points (Figure 3a,b) also change due to non-rigid movements. Figure 3c shows the changes in Euclidean distances between corresponding points.
In this paper, we employ geodesic distances between point clouds to assess the spatial compatibility of two sets of correspondences, as shown in Equation (1). The fundamental idea is that, for any two correctly matched correspondences c i and c j , the geodesic distance between the deformed original points p i s and p j s and the geodesic distance between the deformed points p i t and p j t should remain the same. Taking the non-rigid motion of the human body as an example in Figure 3, (d) represents the geodesic distances between points of the source point cloud, (e) represents the geodesic distances between points of the target point cloud, and (f) represents the difference in geodesic distances between corresponding points, which remains largely unaffected by non-rigid deformations.
S g e o d i s t ( c i , c j ) = | D g e o d i s t ( p i s p j s ) D g e o d i s t ( p i t p j t ) |
Computing geodesic distances on mesh surfaces is a fundamental problem in digital geometry processing and computer-aided design. Most existing exact algorithms partition the mesh edges into multiple interval windows, processing one window at a time. In this study, a vertex-oriented triangle propagation (VTP) algorithm is used to calculate geodesic distances, which is faster compared to conventional algorithms.
(2)
First-order spatial compatibility
First-order spatial compatibility measurement suggests that the smaller the spatial distance difference between two matching pairs, the greater their similarity. Since the theoretically correct correspondence pairs (inliers) should have a distance difference of 0 between them, the similarity between two inliers is relatively high, forming a clustering effect among inliers. First-order spatial compatibility can be calculated using a distance metric. The operator is a monotonically decreasing function, meaning the smaller the distance metric between two matching pairs, the higher their spatial compatibility. Common monotonically decreasing functions include the negative exponential function:
S C i j = exp S g e o d i s t ( c i , c j ) 2 d 2
Here, d is the distance parameter.
(3)
Second-order spatial compatibility
A second-order spatial compatibility measurement method was proposed in reference [32]. First-order compatibility suffers from ambiguity, meaning outliers can also have a high similarity with inliers. The first-order spatial compatibility measure is first binarized, and then, for any two compatible correspondence pairs, the number of correspondence pairs that are jointly compatible with them is calculated as their similarity. A hard thresholding is applied to the binarized first-order spatial compatibility, resulting in { 0 , 1 } N × N :
C i j = 0 if S C i j τ c m p 1 if S C i j > τ c m p
Here, τ c m p is a hard thresholding parameter which can be set closer to 1 to reduce computational complexity for large point clouds. Since inliers are mutually compatible, the similarity between any two inliers is at least the number of inliers in all correspondence pairs, which is not true for inliers and outliers. Therefore, the second-order spatial compatibility measure can be calculated from the binarized first-order spatial compatibility matrix as follows:
S C i j 2 = C i j · k = 1 N C i k · C k j

3.2.3. Maximal Cliques Search

Once the spatial compatibility graph is constructed, the next step is to search for maximal cliques within the graph. A maximal clique is a subset of vertices C V satisfying the following conditions:
  • For every pair of vertices v i , v j C , there is an edge ( v i , v j ) E , meaning that all correspondences in the clique are mutually compatible.
  • The clique is maximal, meaning that no additional vertices from the graph can be added to the clique without violating the clique property (i.e., without breaking the spatial compatibility condition).
The maximal cliques represent sets of correspondences that are mutually compatible based on the spatial constraints, and thus are likely to be valid correspondences. The Eppstein algorithm uses a recursive backtracking framework similar to the Bron–Kerbosch algorithm but introduces optimizations that make it more efficient. The core idea is to explore candidate sets of vertices that may form maximal cliques and prune the search space to avoid unnecessary computations.
Define R as the current set of vertices that form a growing clique, P as the set of candidate vertices that can still be added to R without violating the clique property, and X as the set of vertices that have already been considered for the clique. At each step, the algorithm attempts to add a vertex from P to R and then recursively explores the potential cliques that can be formed. If a clique cannot be expanded any further, it is reported as a maximal clique.
The general recursion is structured as follows:
MaxClique ( R , P , X )
If both P and X are empty, then R is a maximal clique. For each vertex v P , recursively call the following:
MaxClique ( R { v } , P N ( v ) , X N ( v ) )
where N ( v ) is the neighborhood of vertex v and the recursion continues until all vertices have been explored.
The finally obtained maximal cliques represent a consistency set, which is a collection of point pairs that are highly consistent under the graph constraints. To reduce the number of maximal cliques, only the maximal clique with the highest total weight is selected from all maximal cliques containing the same node, and duplicate maximal cliques are removed.

3.3. SMPL Inverse Kinematics with Sparse Correspondences

3.3.1. Rigid Transform Estimation

In the process of human body deformation during motion, the range and variations of the limbs are more pronounced, while the motion of the torso is relatively small. Therefore, making an initial estimation of the rigid body motion of the entire human body based on the torso can effectively reduce the impact of the multiple degrees of freedom of the human body on non-rigid pose matching. The selection of torso points is based on the SMPL (Skinned Multi-Person Linear) model, which provides a well-defined human body mesh with a known number of vertices, each associated with specific body parts, including the torso, limbs, and head. In the SMPL model, the points (vertices) corresponding to the torso region are predefined, allowing us to extract these points for rigid body transformation estimation. To estimate the rigid motion, we select correspondences from the point cloud that match the torso vertices in the SMPL model. By leveraging the prior segmentation information from the SMPL model, we can reliably isolate the torso points for use in the rigid body transformation estimation, thus improving the robustness of the non-rigid registration process as follows:
R 0 = arg min R i = 1 3 p ^ i R q ^ i 2 2
In this equation, R represents the overall rotation of the human body, and ( q ^ i , p ^ i ) denotes a pair of matching points in decentered coordinates. q ^ i = q i 1 N t q i and p ^ i = p i 1 N t p i , where N t is the number of matching points in the torso. Expressed in matrix form, it can be written as follows:
R 0 = arg min R P R Q F 2
The optimization problem for the Frobenius norm can be converted into an optimization problem for the trace of matrices as follows:
arg min R P R T F 2 = arg min trace ( ( P R T ) T ( P R T ) ) = arg max trace ( R T P T )
Denote A = T P T = U Λ V T , then R 0 = V U T .

3.3.2. Non-Rigid Registration with Sparse Correspondences

In the process of non-rigid human point cloud registration, we utilize a Skinned Multi-Person Linear (SMPL)-model-driven optimization strategy to align the SMPL human body model with the point cloud data. The optimization is achieved by constructing a model–data matching loss function, designed to optimize the SMPL model’s pose parameters to match the observed point cloud, thus handling the non-rigid deformations effectively.
The overall loss function is defined as follows:
E ( θ ) = E v + λ θ b E θ b
Here, E θ b represents the VPoser pose prior model, as introduced in SMPLify-X [22], which is used to regularize the human body pose and prevent unnatural deformations. VPoser learns a lower-dimensional representation of plausible human poses using a variational autoencoder (VAE), ensuring that the optimized pose parameters θ remain within the realm of realistic human motion. The term λ θ b is a weight coefficient that controls the influence of this regularization in the optimization.
The term E v is the matching loss between the model and the point cloud data. It measures the difference between corresponding points from the model and the data and is given by the following:
E v = ( p i , q i ) C ρ ( M i p i )
In this equation, ( p i , q i ) represents a sparse matching pair, where q i is a vertex on the SMPL model, and p i is the corresponding point in the point cloud data. After inverse kinematic (IK) iterations, M i represents the deformed position of the model point corresponding to q i . The difference between the point cloud data point p i and the deformed model point M i is penalized using the Geman–McClure robust penalty function ρ ( s ) = s 2 s 2 + σ , which is designed to mitigate the influence of outliers or noisy data by reducing the contribution of large errors.
To optimize the human body pose parameters θ , we use the L-BFGS algorithm, a quasi-Newton method well suited for high-dimensional optimization problems, including inverse kinematics. By iteratively minimizing the loss function, the algorithm aligns the SMPL model with the point cloud data, calculating nearest neighbor correspondences between the model points and the data points, and reducing the distances between matched pairs.
This SMPL-model-driven optimization strategy allows for accurate estimation of the deformed human body pose, even in scenarios with large deformations or rapid motion. By leveraging the predefined SMPL model, which includes anatomical knowledge of the human body, this method effectively reduces the impact of the body’s high degrees of freedom on the registration process, particularly in complex motion sequences.
After performing the maximal clique search and generating multiple sets of matching points between the model and the point cloud, the quality of each set of correspondences can be evaluated. One effective way to evaluate the matching accuracy is by using the nearest neighbor distance-based evaluation method, as shown in the following formula:
D k n n = i | M | min v i n M ( θ ) M i v 2
In this formula, D k n n represents the total sum of the squared distances between each point M i in the point cloud and its nearest neighbor v in the SMPL model M ( θ ) , where θ represents the current pose parameters of the SMPL model. The term min v M ( θ ) M i v 2 computes the squared Euclidean distance between the point M i and its nearest point v in the deformed SMPL model.
This evaluation metric provides a quantitative measure of how well the points from the model and the point cloud align. By minimizing D k n n , we can ensure that the SMPL model is closely matched to the observed point cloud, indicating successful optimization.

4. Experiments

4.1. Datasets

(1)
FAUST Dataset
The FAUST dataset (https://faust-leaderboard.is.tuebingen.mpg.de/download, accessed on 1 September 2023) consists of 300 high-resolution scans of real human bodies, involving 10 individuals and a wide range of poses and evaluation methods [33]. This dataset was collaboratively created by the Max Planck Institute for Intelligent Systems in Germany and the University of Padua in Italy. Its purpose is to provide a standard benchmark for evaluating and comparing various 3D human body reconstruction methods and mesh registration methods. For our experiments, we used the registration part of this dataset, which includes 100 different poses from various individuals in the form of triangular mesh data, to perform non-rigid registration and human body pose estimation experiments.
(2)
CAPE Dataset
The CAPE dataset (https://cape.is.tue.mpg.de/download.php, accessed on 1 September 2023) is a large-scale human body dataset created in collaboration between the Max Planck Institute for Intelligent Systems in Germany, the University of Tübingen in Germany, the Swiss Federal Institute of Technology in Zurich, and the University of Grenoble Alpes in France. This dataset consists of eight males and three females and covers a wide range of human body motions. Approximately 80,000 frames of 3D scan sequences were captured at 60 FPS using a high-precision scanner from 3dMD. These raw scan data were then fitted with SMPL human body models, and clothed human body mesh data were obtained using nonlinear optimization algorithms [34].

4.2. Sparse Correspondence Evaluation

To validate the accuracy of the sparse matching pairs extracted by the GBMAC, this section includes an analysis of the experimental results obtained for both the FAUST dataset and the CAPE dataset. A comparison is made between the Global Matching Diffusion Pruning (DP) algorithm [35] and the Geodesic-Based Maximal Clique Non-Rigid Registration method (GBMAC) proposed in this paper for sparse matching on dynamic human body data.
In the analysis of the experimental results, the performance of DP and the GBMAC in generating sparse matching pairs for dynamic human body data from both the FAUST and CAPE datasets is compared and assessed.

4.2.1. FAUST Dataset Sparse Correspondence Evaluation

The FAUST dataset’s registration section includes ten individuals performing ten actions, with similar poses across actions analyzed statistically by comparing the first action’s results with the others, as illustrated in Figure 4. Table 1 details the registration outcomes for pose pairs 1–2, 1–3, …, 1–10, using “Cor.” and “Correct Cor.” as shorthand for Correspondence and Correct Correspondence. As shown in Table 1, the pose pairs 7 and 8 output less correct correspondence compared with the other pose pairs. Pose pair 7 is the motion between pose 1 and 8 in Figure 4, which involves tremendous motion of both the upper and lower limbs, while the pair with the fewest correspondences is 1–9, which corresponds to pose pair 8 in Table 1. This is due to the symmetric nature of the human body. When the body undergoes a 180-degree orientation change, the correspondences cannot be correctly evaluated using geodesic-distance-based spatial compatibility.
Sparse matching experiments used the first mesh of each individual as the source and the remaining nine poses as targets, employing the DP algorithm and the GBMAC algorithm for generating matches. Initial pairs were formed using SHOT features and feature similarity.
The DP algorithm reduced initial matches based on a consistency threshold c 0 and a region radius σ set at 0.5 and 0.05 m, respectively. The GBMAC algorithm identified matches by finding similar points between source and target data, then constructing a spatial compatibility graph and locating matching cliques.
The GBMAC generated hypotheses through maximal clique searching and validated them with nonlinear optimization, resulting in the best matching sets. Figure 5 illustrates the maximal clique matching point sets and corresponding SMPL models for the first individual’s 1–7 pose pair, with the GBMAC producing 654 pairs, 629 of which were correct, and an average distance of 7.83 mm between the SMPL model and the data.
To provide a visual comparison of the matching results obtained by the two different methods, Figure 6 shows the sparse matching results for pose pairs 1–4, 1–7, 1–8, and 1–9 of individual 00032 from the FAUST dataset. Other data pairs, where the pose differences between source and target data are small for the selected pose pairs, are not displayed. In the figure, the left column represents the matching pairs obtained by the GBMAC algorithm presented in this paper, while the right column shows the results obtained by the DP algorithm. The lines between points illustrate the matching relationships provided by the algorithms, with red lines indicating results where the distance between matching pairs exceeded 5 cm when compared to actual matching point distances.
From Figure 6, it can be observed that the GBMAC can achieve sparse matching results similar to the DP algorithm when there is minimal motion in the body’s torso. However, when the body undergoes more substantial global movement (as seen in pose pair 1–9, fourth row of the figure), the GBMAC yields higher-quality point-to-point matches.

4.2.2. CAPE Dataset Evaluation

The CAPE dataset consists of 15 individuals wearing different types of clothing while performing various actions. For this experiment, individual 00096 was selected, and 20 poses from 10 action sequences were chosen for non-rigid registration and human pose estimation which included Soccer, Chicken Wings, Basketball, Squats, Flying Eagle, Punching, Bend Back and Front, ATU, and Pose Model. The same algorithms used for the FAUST dataset were employed to process these data. Table 2 details the registration results for the pose pairs in actions 1–10. Figure 7 provides visual representations of the matching results between the two pose pairs from four different action categories.

4.2.3. Comparison Between First-Order and Second-Order Methods

In order to illustrate the advantages and effectiveness of second-order methods over first-order methods, we compare the number of correct matches obtained by the two methods from the FAUST dataset in Figure 8. The result shows that the second-order method can produce more correct correspondence for the nine pose pairs. As a matter of fact, the first-order spatial compatibility can be disturbed by the ambiguity problem, which refers to situations where the clustering algorithm has difficulty determining the correct cluster assignments due to overlapping or unclear cluster boundaries. The second-order spatial compatibility measure can decrease the likelihood of the outlier being included in the consensus set and thus minimize the ambiguity issue. Consequently, by tackling this challenge, the second-order method has the potential to deliver superior performance in clustering endeavors.

4.2.4. Time Consumption Analysis

To account for the real-time capabilities of our method, we evaluated the time consumption for each step of the experiments performed on the FAUST dataset on a desktop PC with an Intel i9-9900K CPU (Intel Corporation, Santa Clara, CA, USA). Table 3 details the averaged runtime of the four processes in our method performed on the FAUST dataset. During the experiments, the maximal clique search is the most time-consuming process, taking around one minute using the igraph library. The runtime of the maximal clique search is related to the number of matches, which can be controlled by the compatibility threshold. In the experiments, the threshold was set to 0.99, and, for slight motions which have a large number of correspondences, the value was set to 0.999 to control the number of input matches.

4.3. Non-Rigid Registration Evaluation

In this section, the effectiveness of the method presented in this paper is quantitatively and qualitatively evaluated for pose estimation. The results are compared with those of the GBCPD [36], N-ICP [37], RPTS [12], Fast RNRR [13], and DP-SMPL [22,35] models for the FAUST and CAPE datasets.
For quantitative evaluation, the Root Mean Square Error (RMSE) between the results of different algorithms and the ground truth is computed using the following formula:
RMSE = v i V e i 2 | V |
Here, e i = v i v i g t represents the distance error between the registered points and their corresponding ground truth points, and V denotes the set of points being evaluated.

4.3.1. FAUST Dataset Evaluation

This section compares the non-rigid human body pose registration results obtained by various algorithms from the FAUST dataset. Figure 9 displays the 0–1 pose pair, with source and target data in (a) and (b) and registration outcomes in (c)–(h), using a color gradient to indicate errors (0 to 0.1 m). From the figure, we can see that the NICP algorithm tends to cause local distortions with large movements. Non-model-based methods like RPTS and Fast RNRR perform well with accurate matching points. The experiment used DP-generated matching pairs, which were less accurate than manually selected points. The GBCPD algorithm combines geodesic and Euclidean distances, leading to abdominal misalignment with hand movements. DP-SMPL, using DP-generated points and model-based optimization, performs well with accurate matches but suffers from mismatches.
Table 4 quantitatively analyzes pose matching across algorithms. Model-free methods like Fast RNRR and DP-SMPL perform well on simple pose variations but struggle with more complex ones, where model-based methods excel due to strong priors. The GBMAC, introduced here, leverages model-based priors and auto-filters matching pairs for good results across pose pairs, with an overall RMSE of 8.86 mm for the FAUST dataset.

4.3.2. CAPE Dataset Evaluation

This section presents an experimental comparison and analysis of non-rigid human body pose registration results using different algorithms on the CAPE dataset. Compared to the FAUST dataset, the CAPE dataset offers richer data on human body pose variations, and the subjects in the CAPE dataset are dressed, making the data more similar to real-world scenarios. The experiments compared the non-rigid pose registration results for the individual numbered 00096 in the CAPE dataset across 10 different actions. The GBMAC results for the “Soccer”, “Basketball”, and “Flying Eagle” action sequences are shown in Figure 10.
From the figures, it is evident that the CAPE dataset includes more complex human body deformation actions, making the registration process more challenging. Table 5 lists the errors in matching results for different algorithms in various action groups. It can be observed that, for complex human body deformations, both BCPD and NICP algorithms fail to yield satisfactory registration results. RPTS, Fast RNRR, and DP-SMPL, on the other hand, achieve better matching results in simpler action sequences (groups 2 and 5). The GBMAC algorithm introduced in this paper outperforms other algorithms in most action groups, with an average matching error of 22.70 mm across all action groups.

5. Conclusions

In this paper, we introduced the GBMAC algorithm for non-rigid human point cloud registration, designed to handle deformations by building a second-order spatial compatibility graph using geodesic distances. The algorithm performs a maximal clique search to identify sparse correspondences and generates multiple sets of matching pairs. In the optimization stage, the maximal cliques are evaluated based on the root mean square distance error of neighboring points, ultimately producing an effective set of matching pairs and human pose alignments.
Our method was rigorously tested on the FAUST and CAPE datasets through both sparse matching and non-rigid registration experiments. In generating matching pairs, the GBMAC outperformed other methods, providing superior constraint relationships for subsequent tasks. Specifically, our approach achieved non-rigid matching accuracies of 8.86 mm on the FAUST dataset and 22.70 mm on the CAPE dataset, surpassing existing algorithms.
The practical benefits of the proposed GBMAC algorithm are particularly evident in real-world applications involving non-rigid point cloud registration across consecutive frames. In scenarios where device frame rates are limited and rapid human motion causes significant movement between consecutive frames, the accuracy of other algorithms often deteriorates significantly. However, the GBMAC algorithm excels in these challenging conditions, maintaining high accuracy even during large-scale movements, making it highly suitable for dynamic environments and real-time applications.
However, the efficiency of the proposed algorithm needs to be enhanced to meet real-time performance requirements in practical applications. The most time-consuming part is the maximal clique search process, which remains a key research focus for further improvement. Additionally, the non-rigid registration results are highly affected by the directional change of human orientation. This direction information could be further explored to align the human orientation before non-rigid registration.
Despite these challenges, the GBMAC algorithm demonstrates strong potential for advancing non-rigid registration tasks. Future work will focus on optimizing the algorithm’s efficiency, particularly in the maximal clique search process, and improving its scalability to handle more dynamic and complex deformations across various application scenarios.

Author Contributions

Conceptualization, S.G.; methodology, S.G. and S.Z.; validation, S.G., G.X., and G.Z.; data curation, S.G.; writing—review and editing, S.G., G.X., and S.Z.; visualization, S.G.; supervision, X.Z.; project administration, X.Z.; funding acquisition, X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SMPLSkinned Multi-Person Linear model
ICPIterative Closest Point
CPDCoherent Point Drift
ARAPAs Rigid As Possible
FHPFFast Point Feature Histogram
SHOTSignature of Histograms of Orientations
SCSpatial Compatibility
GBMACGeodesic-Based Maximal Clique

References

  1. Marin, R.; Melzi, S.; Rodola, E.; Castellani, U. Farm: Functional automatic registration method for 3d human bodies. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2020; Volume 39, pp. 160–173. [Google Scholar]
  2. Ge, S.; Fan, G. Non-rigid articulated point set registration for human pose estimation. In Proceedings of the 2015 IEEE Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 5–9 January 2015; pp. 94–101. [Google Scholar]
  3. Amberg, B.; Romdhani, S.; Vetter, T. Optimal step nonrigid ICP algorithms for surface registration. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 7–22 June 2007; pp. 1–8. [Google Scholar]
  4. Cheng, S.; Marras, I.; Zafeiriou, S.; Pantic, M. Statistical non-rigid ICP algorithm and its application to 3D face alignment. Image Vis. Comput. 2017, 58, 3–12. [Google Scholar] [CrossRef]
  5. Hegde, S.B.; Gangisetty, S. An Evaluation of Feature Encoding Techniques for Non-Rigid and Rigid 3D Point Cloud Retrieval. In Proceedings of the BMVC, Cardiff, UK, 9–12 September 2019; p. 47. [Google Scholar]
  6. Hirose, O. A Bayesian formulation of coherent point drift. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2269–2286. [Google Scholar] [CrossRef] [PubMed]
  7. Zollhöfer, M.; Nießner, M.; Izadi, S.; Rehmann, C.; Zach, C.; Fisher, M.; Wu, C.; Fitzgibbon, A.; Loop, C.; Theobalt, C.; et al. Real-time non-rigid reconstruction using an RGB-D camera. ACM Trans. Graph. (ToG) 2014, 33, 1–12. [Google Scholar] [CrossRef]
  8. Qin, Z.; Yu, H.; Wang, C.; Peng, Y.; Xu, K. Deep Graph-Based Spatial Consistency for Robust Non-Rigid Point Cloud Registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 5394–5403. [Google Scholar]
  9. Zhao, M.; Yu, X.; Hu, X.; Wang, F.; Xu, L.; Ji, X. Bidirectional Patch-based Correlations with Local Rigidity for Global Nonrigid Registration. IEEE Trans. Circuits Syst. Video Technol. 2024, 1. [Google Scholar] [CrossRef]
  10. Liu, W.; Wu, H.; Chirikjian, G.S. LSG-CPD: Coherent point drift with local surface geometry for point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15293–15302. [Google Scholar]
  11. Gao, Q.; Zhao, Y.; Xi, L.; Tang, W.; Wan, T.R. Break and Splice: A Statistical Method for Non-Rigid Point Cloud Registration. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2023; Volume 42, p. e14788. [Google Scholar]
  12. Li, K.; Yang, J.; Lai, Y.K.; Guo, D. Robust non-rigid registration with reweighted position and transformation sparsity. IEEE Trans. Vis. Comput. Graph. 2018, 25, 2255–2269. [Google Scholar] [CrossRef] [PubMed]
  13. Yao, Y.; Deng, B.; Xu, W.; Zhang, J. Quasi-newton solver for robust non-rigid registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7600–7609. [Google Scholar]
  14. Lin, M.; Murali, V.; Karaman, S. A planted clique perspective on hypothesis pruning. IEEE Robot. Autom. Lett. 2022, 7, 5167–5174. [Google Scholar] [CrossRef]
  15. Lin, Y.K.; Lin, W.C.; Wang, C.C. K-closest points and maximum clique pruning for efficient and effective 3-d laser scan matching. IEEE Robot. Autom. Lett. 2022, 7, 1471–1477. [Google Scholar] [CrossRef]
  16. Zhang, X.; Yang, J.; Zhang, S.; Zhang, Y. 3D registration with maximal cliques. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 17745–17754. [Google Scholar]
  17. Bogo, F.; Black, M.J.; Loper, M.; Romero, J. Detailed full-body reconstructions of moving people from monocular RGB-D sequences. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 2300–2308. [Google Scholar]
  18. Zuo, X.; Wang, S.; Zheng, J.; Yu, W.; Gong, M.; Yang, R.; Cheng, L. Sparsefusion: Dynamic human avatar modeling from sparse rgbd images. IEEE Trans. Multimed. 2020, 23, 1617–1629. [Google Scholar] [CrossRef]
  19. Anguelov, D.; Srinivasan, P.; Koller, D.; Thrun, S.; Rodgers, J.; Davis, J. Scape: Shape completion and animation of people. In ACM SIGGRAPH 2005 Papers; ACM: New York, NY, USA, 2005; pp. 408–416. [Google Scholar]
  20. Joo, H.; Simon, T.; Sheikh, Y. Total capture: A 3d deformation model for tracking faces, hands, and bodies. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8320–8329. [Google Scholar]
  21. Loper, M.; Mahmood, N.; Romero, J.; Pons-Moll, G.; Black, M.J. SMPL: A skinned multi-person linear model. ACM Trans. Graph. (TOG) 2015, 34, 1–16. [Google Scholar] [CrossRef]
  22. Pavlakos, G.; Choutas, V.; Ghorbani, N.; Bolkart, T.; Osman, A.A.; Tzionas, D.; Black, M.J. Expressive body capture: 3d hands, face, and body from a single image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10975–10985. [Google Scholar]
  23. Gall, J.; Rosenhahn, B.; Brox, T.; Seidel, H.P. Optimization and filtering for human motion capture. Int. J. Comput. Vis. 2010, 87, 75–92. [Google Scholar] [CrossRef]
  24. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  25. Li, H.; Sumner, R.W.; Pauly, M. Global correspondence optimization for non-rigid registration of depth scans. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2008; Volume 27, pp. 1421–1430. [Google Scholar]
  26. Sorkine, O.; Alexa, M. As-rigid-as-possible surface modeling. In Symposium on Geometry Processing; Citeseer: Kitchener, ON, USA, 2007; Volume 4, pp. 109–116. [Google Scholar]
  27. Li, Z.; Heyden, A.; Oskarsson, M. Template based Human Pose and Shape Estimation from a Single RGB-D Image. In Proceedings of the ICPRAM, Prague, Czech Republic, 19–21 February 2019; pp. 574–581. [Google Scholar]
  28. Liu, G.; Rong, Y.; Sheng, L. VoteHMR: Occlusion-aware voting network for robust 3D human mesh recovery from partial point clouds. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, China, 20–24 October 2021; pp. 955–964. [Google Scholar]
  29. Wang, S.; Geiger, A.; Tang, S. Locally aware piecewise transformation fields for 3d human mesh registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7639–7648. [Google Scholar]
  30. Jiang, H.; Cai, J.; Zheng, J. Skeleton-aware 3d human shape reconstruction from point clouds. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5431–5441. [Google Scholar]
  31. Ying, J.; Zhao, X. RGB-D fusion for point-cloud-based 3D human pose estimation. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 3108–3112. [Google Scholar]
  32. Chen, Z.; Sun, K.; Yang, F.; Tao, W. Sc2-pcr: A second order spatial compatibility for efficient and robust point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13221–13231. [Google Scholar]
  33. Bogo, F.; Romero, J.; Loper, M.; Black, M.J. FAUST: Dataset and evaluation for 3D mesh registration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3794–3801. [Google Scholar]
  34. Zhang, C.; Pujades, S.; Black, M.J.; Pons-Moll, G. Detailed, accurate, human shape estimation from clothed 3D scan sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4191–4200. [Google Scholar]
  35. Tam, G.K.; Martin, R.R.; Rosin, P.L.; Lai, Y.K. Diffusion pruning for rapidly and robustly selecting global correspondences using local isometry. ACM Trans. Graph. 2014, 33, 4. [Google Scholar] [CrossRef]
  36. Hirose, O. Geodesic-Based Bayesian Coherent Point Drift. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5816–5832. [Google Scholar] [CrossRef] [PubMed]
  37. Tombari, F.; Salti, S.; Di Stefano, L. Unique signatures of histograms for local surface description. In Proceedings of the Computer Vision–ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part III 11. Springer: Berlin/Heidelberg, Germany, 2010; pp. 356–369. [Google Scholar]
Figure 1. Workflow of geodesic-based maximal clique search for non-rigid point cloud registration.
Figure 1. Workflow of geodesic-based maximal clique search for non-rigid point cloud registration.
Sensors 24 06924 g001
Figure 2. Illustration of the non-rigid deformation, the red and yellow lines represent the Euclidean and Geodesic distances between the wrist and waist points, respectively.
Figure 2. Illustration of the non-rigid deformation, the red and yellow lines represent the Euclidean and Geodesic distances between the wrist and waist points, respectively.
Sensors 24 06924 g002
Figure 3. Euclidean distances, geodesic distances, and differences between matched points before and after human body motion, (a,b) are the Euclidean distance among source point and target point, respectively. (c) is the difference of (a,b). (d,e) are the Geodesic distance among source point and target point, respectively. (f) is the difference of (d,e).
Figure 3. Euclidean distances, geodesic distances, and differences between matched points before and after human body motion, (a,b) are the Euclidean distance among source point and target point, respectively. (c) is the difference of (a,b). (d,e) are the Geodesic distance among source point and target point, respectively. (f) is the difference of (d,e).
Sensors 24 06924 g003
Figure 4. Illustration of FAUST dataset 01 Personnel Action [33].
Figure 4. Illustration of FAUST dataset 01 Personnel Action [33].
Sensors 24 06924 g004
Figure 5. SMPL model from different maximal cliques, (ac) are the three maximal clique sets after searching and (df) are the final results after non-rigid registration results with the correspondences from each set.
Figure 5. SMPL model from different maximal cliques, (ac) are the three maximal clique sets after searching and (df) are the final results after non-rigid registration results with the correspondences from each set.
Sensors 24 06924 g005
Figure 6. Qualitative comparison of matching pairs of GBMAC and DP algorithms obtained from the FAUST dataset.
Figure 6. Qualitative comparison of matching pairs of GBMAC and DP algorithms obtained from the FAUST dataset.
Sensors 24 06924 g006
Figure 7. Qualitative comparison of sparse correspondence from the CAPE dataset.
Figure 7. Qualitative comparison of sparse correspondence from the CAPE dataset.
Sensors 24 06924 g007
Figure 8. Comparison of the first-order and second-order methods applied to the FAUST dataset.
Figure 8. Comparison of the first-order and second-order methods applied to the FAUST dataset.
Sensors 24 06924 g008
Figure 9. Registration results of different algorithms for the FAUST dataset.
Figure 9. Registration results of different algorithms for the FAUST dataset.
Sensors 24 06924 g009
Figure 10. The non-rigid registration results for Soccer, Basketball, and Flying Eagle pose series from the CAPE dataset.
Figure 10. The non-rigid registration results for Soccer, Basketball, and Flying Eagle pose series from the CAPE dataset.
Sensors 24 06924 g010
Table 1. Results comparison of sparse correspondence in FAUST dataset.
Table 1. Results comparison of sparse correspondence in FAUST dataset.
Pose PairInitializationDiffusion PruningGBMAC
Cor.Correct Cor.Cor.Correct Cor.Cor.Correct Cor.
1298542442040827732700
2343794594394117621738
3318280478277615191483
4349298998498132243163
5342110281028102526352590
6297758958056720311998
72572939168982953
8229720366146109
9297649147646128122758
Ave.314570469969119871944
Table 2. Results comparison of sparse correspondence in CAPE dataset.
Table 2. Results comparison of sparse correspondence in CAPE dataset.
Pose PairInitializationDiffusion PruningGBMAC
Cor.Correct Cor.Cor.Correct Cor.Cor.Correct Cor.
12670295311281895836
2310554455053112881169
3301361262159816191509
4325177477275812691175
530543563633401073953
6272532033230312821170
726452372522221020908
8293970370769011891131
92553167182149901820
1023958711478298248
Ave.28354104213951084992
Table 3. Averaged runtime for 100 launches on FAUST datasets.
Table 3. Averaged runtime for 100 launches on FAUST datasets.
ProcessTime (s)
Initial Correspondences12.6
Graph Construction18.8
Maximal Clique Search67.5
Hypothesis Evaluation21.7
Total120.6
Table 4. Quantitative comparisons of non-rigid registration for the FAUST dataset (mm), the bold are the best results among these methods.
Table 4. Quantitative comparisons of non-rigid registration for the FAUST dataset (mm), the bold are the best results among these methods.
Pose PairsGBCPDNICPRPTSFast RNRRDP-SMPLGBMAC
156.5944.5911.9711.0414.468.03
223.2019.7310.119.655.937.29
3286.1965.1311.3412.409.588.59
432.3327.2310.089.986.427.06
517.6015.519.359.526.336.94
6406.3250.2710.8710.3324.538.13
7434.26157.4935.0731.8640.549.72
8241.91220.24183.95228.7470.0315.79
9438.5059.3410.8911.9418.948.19
Ave.215.2173.2832.6337.2721.868.86
Table 5. Quantitative comparisons of non-rigid registration for the CAPE dataset (mm), the bold are the best results among these methods.
Table 5. Quantitative comparisons of non-rigid registration for the CAPE dataset (mm), the bold are the best results among these methods.
Pose SeriesBCPDNICPRPTSFast RNRRDP-SMPLGBMAC
165.71225.82113.92100.07137.9920.71
2132.7971.7022.8422.4120.9123.71
3157.79190.16131.03113.31120.2015.15
4180.0771.1351.4546.1511.3312.97
5108.45138.9819.4317.0139.9517.84
6148.32165.4558.8959.3070.4519.22
7289.61245.75231.03240.29225.7424.60
8150.2196.4434.2132.8015.8820.19
9110.17284.97198.32204.83249.9521.82
10446.41457.97409.23411.68322.5250.82
Ave.178.95194.84127.04124.78121.4922.70
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gan, S.; Xu, G.; Zhuge, S.; Zhang, G.; Zhang, X. Geodesic-Based Maximal Cliques Search for Non-Rigid Human Point Cloud Registration. Sensors 2024, 24, 6924. https://doi.org/10.3390/s24216924

AMA Style

Gan S, Xu G, Zhuge S, Zhang G, Zhang X. Geodesic-Based Maximal Cliques Search for Non-Rigid Human Point Cloud Registration. Sensors. 2024; 24(21):6924. https://doi.org/10.3390/s24216924

Chicago/Turabian Style

Gan, Shuwei, Guangsheng Xu, Sheng Zhuge, Guoyi Zhang, and Xiaohu Zhang. 2024. "Geodesic-Based Maximal Cliques Search for Non-Rigid Human Point Cloud Registration" Sensors 24, no. 21: 6924. https://doi.org/10.3390/s24216924

APA Style

Gan, S., Xu, G., Zhuge, S., Zhang, G., & Zhang, X. (2024). Geodesic-Based Maximal Cliques Search for Non-Rigid Human Point Cloud Registration. Sensors, 24(21), 6924. https://doi.org/10.3390/s24216924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop