Next Article in Journal
Visualization of Demodulated Sound Based on Sequential Acoustic Ray Tracing with Self-Demodulation in Parametric Array Loudspeakers
Previous Article in Journal
Characteristics and Genesis of Pore–Fracture System in Alkaline Lake Shale, Junggar Basin, China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Point Cloud Registration Algorithm Based on Weighting Strategy for 3D Indoor Spaces

1
MEMS Center, Harbin Institute of Technology, Harbin 150001, China
2
Key Laboratory of Micro-Systems and Micro-Structures Manufacturing, Ministry of Education, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(12), 5240; https://doi.org/10.3390/app14125240
Submission received: 13 May 2024 / Revised: 3 June 2024 / Accepted: 13 June 2024 / Published: 17 June 2024

Abstract

:
Point cloud registration is a technology that aligns point cloud data from different viewpoints by computing coordinate transformations to integrate them into a specified coordinate system. Many cutting-edge fields, including autonomous driving, industrial automation, and augmented reality, require the registration of point cloud data generated by millimeter-wave radar for map reconstruction and path planning in 3D environments. This paper proposes a novel point cloud registration algorithm based on a weighting strategy to enhance the accuracy and efficiency of point cloud registration in 3D environments. This method combines a statistical weighting strategy with a point cloud registration algorithm, which can improve registration accuracy while also increasing computational efficiency. First, in 3D indoor spaces, we apply PointNet to the semantic segmentation of the point cloud. We then propose an objective weighting strategy to assign different weights to the segmented parts of the point cloud. The Iterative Closest Point (ICP) algorithm uses these weights as reference values to register the entire 3D indoor space’s point cloud. We also show a new way to perform nonlinear calculations that yield exact closed-form answers for the ICP algorithm in generalized 3D measurements. We test the proposed algorithm’s accuracy and efficiency by registering point clouds on public datasets of 3D indoor spaces. The results show that it works better in both qualitative and quantitative assessments.

1. Introduction

Various registration algorithms have significantly improved accuracy and efficiency, but their primary application is in the registration of small components. When facing complex 3D indoor spaces, existing registration algorithms struggle to attain a good balance between accuracy and efficiency. A 3D indoor space, like a room, contains objects of different shapes and sizes with large quantities of point clouds, which pose a challenge for both accuracy and efficiency. The idea behind this paper is to use a statistical weighting strategy to come up with a point cloud registration algorithm that is both accurate and quick for 3D indoor places. Point cloud registration typically serves 3D map reconstruction or navigation in 3D indoor spaces, necessitating higher weights for significant targets that impact navigation, tracking, or localization, such as chairs on the ground, tables in the room, or walls obstructing the path. We can appropriately reduce weights for other types of targets to save computational resources during registration, thereby enhancing the overall efficiency of the algorithm and the registration performance of key targets. This paper uses PointNet [1] for semantic segmentation, marking different targets in the point cloud of 3D indoor spaces. A statistical weighting strategy is introduced where the segmented parts of the point cloud are treated as evaluation criteria and various attributes of the point cloud are set as samples for evaluation. We then use the weights of different parts of the point cloud as reference values to determine the minimum metric of the nonlinear objective function in the Iterative Closet Point (ICP) registration algorithm. We use a new technique to obtain closed-form answers to the nonlinear objective function in a 3D generalized metric. This makes the registration more accurate. Finally, experiments with 3D indoor space datasets verify the accuracy and efficiency of the proposed algorithm, demonstrating significant improvements over the traditional point cloud registration algorithms.
We organize the remainder of this paper as follows: Section 2 introduces the weighting strategy, including the basic principles of objective weighting methods and a novel improvement to the calculation steps proposed in this paper. Section 3 describes the process of obtaining the ICP minimum metric solution with assigned weights. Section 4 analyzes the accuracy and efficiency of the proposed algorithm with registration experiments. Section 5 ultimately provides the conclusion and outlook.
Point cloud registration is a vital technique in computer vision. It introduces a spatial transformation that maps one image onto another and ensures a one-to-one correspondence between points in both images of the same spatial location to achieve information fusion [2]. Over the years, research into point cloud registration has made great progress. In 1992, Besl et al. proposed the Iterative Closest Point (ICP) algorithm [3]. By calculating the distances between corresponding points in the source point cloud and the target point cloud, a rotation–translation matrix RT is constructed. Then, the RT can be applied to transform the source point cloud and calculate the mean squared difference of the transformed points. If the mean squared difference satisfies a threshold condition, then the algorithm terminates. Otherwise, it continues iterating until the error meets the threshold condition or the maximum iteration count is reached. However, ICP is usually sensitive to initial values and may become stuck in local optima [4]. In 2009, Magnusson proposed the normal distributions transform (NDT) as a great 3D surface representation with applications in point cloud registration [5]. The NDT utilizes statistical probability methods to determine corresponding point pairs based on the normal distribution of point clouds, thereby calculating the transformation relationship between the source point cloud and the target point cloud. It is a discrete global registration algorithm that, in discretized coordinate space, maximizes the score of the probability density of the normal distribution calculated for the source points voxelized in the target point cloud. It employs probability density functions and optimal solving methods to compute the transformation relationship between the point clouds to be registered. Compared to ICP, the NDT significantly decreases consumption of hardware resources and enhances registration accuracy. However, since all point cloud data need to be discretized, the NDT has longer running times than ICP, leading to lower efficiency. In 2019, Yasuhiro et al. proposed the PointNetLK, a deep-learning-based point cloud registration algorithm [6]. In PointNetLK, PointNet is treated as a learnable “imaging” function and is utilized for point cloud registration by modifying the classic Lucas–Kanade (LK) algorithm used for image alignment to integrate with PointNet [7]. Unlike ICP and its variants, PointNetLK does not require large amounts of computation resources, offering advantages in terms of registration accuracy, robustness to initialization, and computational efficiency. On the other hand, due to the explicit encoding of the alignment process in the network architecture, PointNetLK shows significant generalization performance for unseen objects, which implies that the network only needs to learn point cloud representations without the necessity to learn the alignment task. In summary, existing point cloud registration algorithms can be broadly categorized into several types: iterative, optimization-based algorithms such as ICP and its variants Point-to-plane ICP, NICP, Go-ICP, and IMLS-ICP, etc. [8,9,10,11]; statistical, Gaussian distribution-based algorithms such as NDT and its variant MSKM-NDT [12]; and deep-learning-based algorithms such as PointNetLK, DeepVCP, and PCGAN [13,14].
In the last few years there have been other papers, such as [15], which provides an overview of traditional and learning-based methods for point cloud registration and outlier rejection. It discusses the limitations of existing approaches and sets the stage for the proposed PointDSC model. Ref. [15] discusses traditional point cloud registration algorithms, such as RANSAC and its variants, which are popular outlier rejection methods but have drawbacks in terms of slow convergence and low accuracy in cases with large outlier ratios. It also mentions the limitations of 3D descriptors in traditional methods due to irregular density and a lack of useful texture, highlighting the importance of geometric consistency, such as length constraints under rigid transformation, which are commonly utilized by traditional outlier rejection algorithms. In addition, it presents learning-based outlier rejection methods, such as DGR and 3DRegNet, which formulate outlier rejection as an inlier/outlier classification problem. These methods use operators like sparse convolution and pointwise MLP for feature embedding, but they overlook the spatial compatibility between inliers. The section also highlights the limitations of existing outlier rejection methods, as well as the need for spatial consistency in outlier rejection for robust 3D point cloud registration. In the related work section of [15], certain algorithms are suggested, including RANSAC [16], spectral matching [17], DGR [18], and 3DRegNet [19]. This section gives a full picture of all the current methods used for point cloud registration and outlier rejection.
Ref. [20] focuses on the recent developments in point cloud registration, including same-source and cross-source methods, as well as the connections between optimization-based and deep learning techniques. The authors emphasize the need for a comprehensive survey to summarize recent advances in point cloud registration, as existing reviews mainly focus on conventional point cloud registration and deep learning techniques, with limited coverage of cross-source point cloud registration. The [20] paper’s contributions include a comprehensive review of same-source point cloud registration, a review of cross-source registration, the development of a new cross-source point cloud benchmark, and the exploration of potential applications and future research directions in point cloud registration. It also discusses the challenges in cross-source point cloud registration, such as noise, outliers, density differences, partial overlap, and scale differences. It provides an evaluation of state-of-the-art registration algorithms on a cross-source dataset, highlighting the performance of optimization-based and deep learning methods. Refs. [17,21,22,23,24,25] provide a foundation for the survey and support the authors’ claims and contributions in the field of point cloud registration. Ref. [20] introduces new cross-source datasets, such as 3DMatch, KITTI, ETHdata, and 3DCSR, for comparison experiments and benchmarking.

2. Weighting Strategy of Targets in Semantic Segmentation

2.1. The Objective Weighting Method

The objective weighting method utilizes objective and measurable metrics, rather than subjective judgments, for weight assignment to ensure fairness and rationality. In our task of point cloud registration, the coordinates, color information, and category labels of each point cloud data point are totally quantifiable. Therefore, using the objective weighting method for weight assignment aligns well with the requirements of the application background. Common objective weighting methods include the standard deviation method, the entropy weight method, and the criteria importance through intercriteria correlation (CRITIC) method [26,27,28]. The standard deviation method’s basic idea is to calculate the variance Si of the data under each indicator. We obtain the weight value for the ith indicator by dividing Si by the sum of all Si values. Larger variances result in higher weights. This method is relatively simple and easy to implement, effectively distinguishing between different indicators. However, it assumes that all indicators are equally important, and a larger variance only indicates higher discrimination ability for different scenarios, not necessarily the importance of the indicator [29]. The entropy weight method, on the other hand, calculates entropy weights based on the dispersion of data for each indicator using information entropy. Afterward, we apply certain corrections to these entropy weights, leading to more objective indicator weights. This method allows for weight adjustments and exhibits adaptability, but it cannot handle strong correlations between indicators; thus, there is a risk of unreasonable weight assignments [30]. The CRITIC method comprehensively assesses the objective weights of indicators based on the comparative intensity of evaluation indicators and the conflict between indicators. It takes into account both the variability in indicator values and the correlation between indicators. Larger numerical values do not necessarily indicate greater importance; instead, the method utilizes the inherent objectivity of the data for scientific evaluation. Comparative intensity is defined as the magnitude of the differences in values between various evaluation schemes for the same indicator, expressed in the form of standard deviation. A larger standard deviation indicates greater fluctuation, implying larger differences in values between schemes, leading to higher weights. Correlation coefficients represent the conflict between indicators. If two indicators exhibit a strong positive correlation, indicating lower conflict, then the weights will be lower. The CRITIC method is used in this paper to give semantic segmentation targets weights. It takes into account the strong correlation between point cloud sets of the same target as well as positional relationships like occlusion and overlap between different targets that affect the registration process.

2.2. Definition of Alternatives and Criteria

In the classic CRITIC method, objects of different categories are defined as alternatives and performance metrics of different aspects of objects are defined as criteria [31]. This paper proposes a novel definition of alternatives and criteria, considering the practical significance of labels in semantic segmentation. The individual segmentation performance of each label influences the overall segmentation performance, and these influences positively correlate. A better local segmentation performance contributes to the improvement in overall segmentation quality. On the other hand, the labels have relative independence, allowing them to serve as different criteria for the same performance metric. Therefore, in the following sections of this paper, different performance metrics in semantic segmentation will be defined as alternatives, and different labels will be defined as criteria.
Based on the format of point cloud data in semantic segmentation, we define each point cloud as a vector,
p i = { x i , y i , z i , R i , G i , B i , L i } , 0 i n
where x, y, and z are the 3D coordinates of point cloud, R, G, and B are the color information of each point cloud. L represents the label of each point cloud, indicating which class of target it belongs to, and n is the number of all the point clouds in the space.
We can divide the labels of point clouds used in this paper into the following categories based on the Stanford Large-Scale 3D Indoor Spaces Dataset (S3DIS) classification [32], as shown in Table 1.
All the labels refer to common furniture found in indoor rooms, with the exception of the last one, “clutter”, which is added to categorize point clouds that cannot be reliably classified into a specific class.
We assume three performance metrics as alternatives for the weighting strategy, which include the importance (IM) of labels, the intersection-over-union (IoU) of labels, and the Class-Pixel-Accuracy (CPA) of labels [33,34]. We subjectively assign and quantitatively express the importance of labels in the following ways, as shown in Table 2:
We set the maximum score of importance at 1. We assign a lower importance to objects such as ceilings, floors, and walls that serve as backgrounds. We assign a higher importance to furniture like tables, sofas, and chairs, which are numerous and highly utilized. Other types of furniture, with relatively infrequent use, fall between the two in terms of importance. We set the importance of clutter objects to only 0.1 in order to exclude them from the radar detection range.
We define the IoU and CPA based on a confusion matrix, a specific matrix that visualizes algorithm performance, where each column represents predicted values, and each row represents the actual categories [35]. In the semantic segmentation results of point clouds, we evaluate the segmentation performance for each class using a binary confusion matrix.

2.3. Steps of Calculation

A decision matrix is defined as follows:
M = M 11 M 1 m M n 1 M n m
The variables n and m represent the quantity of alternatives and evaluation criteria, respectively. Mij is the value of the jth criterion for the ith alternative, 1 ≤ in, 1 ≤ jm. We specify n as 3 and m as 13 for our 3D indoor point cloud semantic segmentation task.
Because each criterion has different dimensions, normalizing the elements in the decision matrix is necessary to unify the units. This paper’s criteria anticipate higher values, leading to the adoption of a forward normalization approach:
D i j = M i j M j M j + M j ; i = 1 , , n ; j = 1 , , m
where M j = min(M1j, M2j, …, Mnj), M j + = max(M1j, M2j, …, Mnj), and Dij is the element Mij of the decision matrix after normalization by dividing its distance from the maximum value by the length of the range.
The Pearson correlation coefficient between two criteria j and k is defined as follows:
r j k = i = 1 n ( D i j D j ¯ ) ( D i k D k ¯ ) i = 1 n ( D i j D j ¯ ) 2 i = 1 n ( D i k D k ¯ ) 2
where D j ¯ and D k ¯ are mean values of the jth and kth criteria after normalization by Equation (5):
D j ¯ = 1 m i = 1 n D i j ; j = 1 , , m
D k ¯ = 1 m i = 1 n D i k ; k = 1 , , m
The standard deviation of each criterion is defined:
σ j = 1 m 1 i = 1 n ( D i j D j ¯ ) 2 ; j = 1 , , m
where σ is the standard deviation, which serves to contrast the intensity among each criterion of the same alternative.
The measure of conflict among each criterion is defined:
p j = k = 1 m ( 1 r j k ) ; j = 1 , , m
where pj is the measure of conflict between the jth criterion and the others, which indicates the correlation among the criteria. The more positive the correlation is, the smaller the value of the inconsistency will be.
The amount of information of each criterion is defined:
Q j = σ j p j
where Qj is the amount of information of the jth criterion, which equals the product of standard deviation and measure of conflict. The greater Qj is, the more information transmitted by the corresponding criterion, and the relatively higher importance it will have.
Finally, the weights of the criteria are defined as follows:
ω j = Q j j = 1 m Q j ; j = 1 , , m
where ωj indicates the weight of the jth criterion and will be applied in the subsequent ICP registration process.

2.4. Improvement of Steps

The criteria for labels in the semantic segmentation application background do not strictly adhere to a linear correlation. For instance, there is no inherent positive or negative correlation between the importance of a particular label and its CPA. The classical CRITIC method, unfortunately, fails to adequately consider this kind of fuzzy nonlinearity [36]. Therefore, we make some improvements to CRITIC based on the methodology proposed by Sharkasi and Rezakhah [37]. In the normalization procedure, we include fuzzy set theory and Hamming distance as similarity measures. Additionally, we use the distance correlation instead of the Pearson correlation to describe both the linear and nonlinear relationships between the criteria.
The following procedures allow for improvements: First, the elements of decision matrix D are converted into percentages.
M i j = M i j i = 1 n M i j
Then, the fuzzy form of elements of decision matrix D is defined as follows:
{ D i j L = M i j Δ d D i j H = M i j + Δ d
where Δd is a very small positive number for all i and j values: 0 < Dij + Δd < 1 and 0 < Dij − Δd < 1. The boundaries of fuzzy RP points are defined as:
{ L j = i = 1 n D i j L n H j = i = 1 n D i j H n
where Lj is the lower bound of the fuzzy RP points and Hj is the upper bound of the fuzzy RP points. The normalization procedure can be described as:
D i j = { 1 , [ D i j L , D i j H ] [ L j , H j ] 1 | D i j L L j | + | D i j H H j | 2 ( D i j H + D i j L ) , [ D i j L , D i j H ] [ L j , H j ]
where Dij is the element after normalization, which belongs to fuzzy set [DijL,DijH] and DjL− = min(D1jL, D2jL, …, DnjL), DjH+ = max(D1jH, D2jH, …, DnjH).
The distance correlation is defined as:
ρ j k = i = 1 n ( D i j D j ¯ ) ( D i k D k ¯ ) n i = 1 n ( D i j D j ¯ ) 2 n i = 1 n ( D i k D k ¯ ) 2 n
where D j ¯ and D k ¯ are mean values of the jth and kth criteria after normalization by Equation (16), the numerator refers to the covariance between the jth and kth criteria, cov(j,k), and the denominator refers to the square root of the product of the variances of the two criteria, var(j) and var(k).
Finally, Equation (14) is used for normalization instead of Equation (3), and Equation (15) is used for the distance correlation coefficient ρjk instead of the Pearson correlation coefficient rjk from Equation (6). In Equation (10), rjk is replaced with ρjk to calculate the measure of conflict.

3. The ICP Algorithm with Weighting Strategy

3.1. Selection of Error Metric

In the ICP algorithm, an objective function is defined to minimize the measurement error. Generally, there are two metrics for error measurement: point-to-point and point-to-plane.
e i = i = 1 N ( R p i + t k q i ) T 2
e i = i = 1 N ( R p i + t k q i ) T n i 2
Equation (16) describes the point-to-point metric, which simply finds the minimum distance between the target point and the corresponding point on the source surface, where R is the rotational matrix, pi is the source point cloud, tk is the translation matrix, and qi is the target point cloud. Equation (17) shows the point-to-plane metric. Its goal is to find the shortest path between the target point and the tangent plane of the source surface’s corresponding point, where ni is the vector that goes through the tangent plane of the source surface at pi.
As shown in Figure 1b, pi’ is a point on the destination surface and pi is the corresponding point on the source surface. The vector ei is the normal vector to the tangent plane of the source surface at pi, which takes the distance from pi’ to the plane tangent to pi as the point-to-plane error metric. The objective function is to find a suitable rigid transformation that satisfies the minimum sum of all these distances.

3.2. Derivation of Solution

The rigid transformation process of a point clouds is defined as:
q i = R ( α k , β k , γ k ) p i + t k
where pi is the ith point in the source point cloud and qi is the corresponding point after rigid transformation. R(αk,βk,γk) is the rotation matrix with rotation angle αk, βk, and γk, and tk is the corresponding translation value. k refers to the iterations of the ICP algorithm.
Based on Equation (18), the objective function can be written as:
F min ( α k + 1 , β k + 1 , γ k + 1 , t k + 1 ) = i = 1 ( n i T [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ] ) 2
Considering that ni, pi, tk+1, pi R3, and that R(α,β,γ) is a 3 × 3 matrix, the objective function can be denoted as:
F min ( α k + 1 , β k + 1 , γ k + 1 , t k + 1 ) = i = 1 ( n i T [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ] ) T ( n i T [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ] ) )   = i = 1 [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ] T n i n i T [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ]
Then, the weight coefficient is introduced into Equation (18):
F min ( α k + 1 , β k + 1 , γ k + 1 , t k + 1 ) = i = 1 [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ] T ω i n i n i T [ R ( α k + 1 , β k + 1 , γ k + 1 ) p i + t k + 1 p i ]
where ωi is the weight coefficient of the ith point whose value is calculated according to the weighting strategy in Section 3.
The traditional method for solving this nonlinear objective function is to linearize it, then use the least squares method to obtain a linearly approximate solution. These methods often have trouble obtaining exact closed-form solutions because of the nonlinearity introduced by the reference plane and the rigid transformation operator. This makes convergence times slower when working with large point cloud datasets. This paper proposes an exact closed-form solution to improve computational efficiency and enhance registration accuracy, by reducing the objective function into a matrix form with constraints and using Lagrange’s multipliers.
In weighted PP-ICP, the rigid transformation is decomposed into many small transformations, each producing a minimal rotation angle along the rotation axis, called θ, which theoretically should be an infinitesimal value that satisfies:
0 < θ 1 °
For each small transformation, the rotation angles αk, βk, and γk have small values. Therefore, the rotation matrix R(αk,βk,γk) can be simplified based on the approximation rule proposed in [8]:
cos θ 1 , sin θ 0 , θ 2 0   if   θ 0
Then, the rotation matrix R(αk,βk,γk) can be simplified to
R ( α k , β k , γ k ) [ 1 α k β k γ k α k γ k + β k γ k α k β k γ k + 1 β k γ k α k β k α k 1 ] [ 1 γ k β k γ k 1 α k β k α k 1 ]
A six-dimensional vector is defined
x = [ t x , t y , t z , α , β , γ ]
as the solution of the objective function. Given the ith point in the source point cloud
p i = [ p i 0 , p i 1 , p i 2 ]
an auxiliary matrix is defined
A i = [ 1 0 0 0 p i 2 p i 1 0 1 0 p i 2 0 p i 0 0 0 1 p i 1 p i 0 0 ]
Then, Equation (21) can be written as
F min ( x ) = i ( A i x p i 1 ) T ω i n i n i T ( A i x p i 1 )
by defining
B i = ω i n i n i T
Ignoring constant terms, Equation (29) can be expanded as
F min ( x ) = x T ( i A i T B i A i ) x + ( i 2 p i T B i A i ) x
by defining
G = i A i T B i A i
H = i 2 p i T B i A i
J = [ 0 3 × 3 0 3 × 3 0 3 × 3 I 3 × 3 ]
Finally, the objective function can be written as
F min ( x ) = x T G x + H x , s . t . x T J x = 0
where the constraint is the approximate matrix form of α2 + β2 + γ2 ≈ 0.
To solve the nonlinear minimization problem, Lagrange’s multipliers are used for derivation. The Lagrange function is defined
L ( x ) = x T G x + H x + λ x T J x
The necessary condition for optimization is derived as
L x = 2 x T G + H + 2 λ x T J = 0 T
which is equivalent to
x = ( 2 G + 2 λ J ) T H T
Equation (37) is substituted into the constraint of Equation (34) to obtain
H ( 2 G + 2 λ J ) 1 J ( 2 G + 2 λ J ) T H T = 0
where λ is the only unknown parameter. According to [38], we can find the solutions in closed form for Equation (38), a sixth-order polynomial. Finally, Equation (37) yields x.
In conclusion, we can summarize the weighted PP-ICP process in this paper as follows: Firstly, PointNet semantically segments the input point cloud data, generating labels for different types of items along with corresponding parameters: importance, IoU, and CPA. Next, we assign weights to these category labels using an objective weighting method, which employs an improved CRITIC as the weighting strategy to calculate the corresponding weight values. Finally, we use the calculated weight values as the weight parameters for the point-to-plane ICP, registering the input point cloud effectively. Figure 2 illustrates the algorithm’s process of weighted PP-ICP.

4. Experiment and Analysis

4.1. Data Preparation

The experimental datasets are sourced from public datasets, S3DIS [32], which consist of six different indoor spaces from six areas, respectively: conference room for Area_1, hallway for Area_2, office for Area_3, lobby for Area_4, storage for Area_5, and open space for Area_6. PointNet infers each label’s IoU and CPA, calculates their average values from the six areas, and summarizes them along with the label’s IM in Table 3.

4.2. Weighting Strategy Steps

Based on Table 3, the decision matrix can be described as
M = [ 0.200 0.200 0.200 0.200 0.600 0.600 0.400 1.000 1.000 0.800 0.600 0.600 0.000 0.977 0.979 0.870 0.918 0.685 0.887 0.896 0.832 0.922 0.836 0.841 0.888 0.804 0.989 0.987 0.957 0.932 0.735 0.902 0.919 0.890 0.946 0.844 0.909 0.930 0.918 ]
Equations (11) and (12) are applied to transform the decision matrix into fuzzy form, as shown in Table 4.
The small positive number Δd in Equation (12) is set to be 0.005.
The boundaries of each criterion and corresponding fuzzy RP points are calculated according to Equation (13), and the results are shown in Table 5.
The decision matrix is normalized according to Equation (14), and the results are shown in Table 6.
The standard variances of criteria are calculated according to Equation (7), and the results are shown as Table 7.
The distance correlations between different pairs of criteria are calculated according to Equation (15), and the results are shown in Table 8.
The amounts of information and weights of criteria are calculated according to Equations (8)–(10) and corresponding ranks are assigned to the criteria based on the weights. The results are shown in Table 9.
Table 9 indicates that among all the labels, table, chair, and wall hold the top three rankings in terms of weights, which aligns well with the practical requirements of point cloud registration in 3D indoor spaces. In the Stanford Large-Scale 3D Indoor Spaces Dataset, chairs, walls, and tables are indispensable objects that occupy the central positions in the indoor space and the majority of the spatial volume. Improving the accuracy of point cloud registration for these objects will correspondingly enhance the registration accuracy for the entire 3D indoor space, as registration transforms the entire indoor space into a rigid structure. On the other hand, the weight assigned to clutter is minimized to a value close to zero, effectively minimizing the negative impact of noise and incorrectly matched points that will reduce registration accuracy.

4.3. Point Cloud Registration Experiments

We use the entire S3DIS dataset in the experiments, from Area_1 to Area_6 [32]. When the S3DIS datasets are used as a whole, all the test data are the averaged results of the six indoor spaces in the datasets, as shown in Figure 3. PointNet is used for semantic segmentation of the point cloud.
We divide the registration experiment into two parts. In the first part, we conduct experiments with weighted PP-ICP under different parameters, using the corresponding experimental data to determine the parameters that enable the algorithm to perform optimally. We also conduct other experiments to study the influence of semantic segmentation quality on the registration results. The second part is a comparison of weighted PP-ICP with three other registration algorithms: ICP, the NDT, and PointNetLK. We assess the advantages and disadvantages of weighted PP-ICP relative to the other algorithms by comparing the algorithms in terms of root mean square (RMS) error, runtime, and hardware resource utilization across six test scenarios in the datasets, using the method of controlled variables. In these experiments, we conduct both single-axis and multi-axis angle tests on the entire S3DIS dataset. In the single-axis angle test, we rotate the datasets around one axis, keeping the other two axes fixed. The range of rotation angles varies from 15° to 120°, in steps of 15°. The multi-axis angle tests maintain a range of angles between 15° and 120°, with a step of 15° in between. For each type of rigid transformation, we rotate all three axes clockwise by the same angle value, for example, [30°, 30°, 30°] and [60°, 60°, 60°].
Section 3 uses subjective assignment to determine each category’s importance in the semantic segmentation results. The following tests will verify the validity of these assignments. According to Table 2, we reduce the importance of all labels above 0.5 and increase those below 0.5, resulting in a change magnitude of 0.2 for each group, which decreases to 0.1 when starting from 0.2. This process generates four groups of data regarding label importance, as shown in Table 10.
We calculate and integrate the corresponding weight values into the proposed registration algorithm based on these four sets of data. We perform validation using multi-axis angle tests, as illustrated in Figure 4 and Table 11.
Figure 4 and Table 11 demonstrate that group 1 consistently has the smallest RMS across all rotational angles. The maximum value in group 3, 3.134, is 3.11 times higher than the maximum value in group 1, which is 0.762; the maximum value in group 4, 3.674, exceeds the maximum in group 1 by 3.82 times. These results validate that the subjective assignment of label importance in Section 3 is reasonable.
Given that this paper proposes an algorithm based on weights generated from point cloud semantic segmentation results, we expect the quality of segmentation to significantly influence the algorithm’s registration performance. However, we notice differences in the point cloud segmentation results among the six areas of the datasets during the experimental process, prompting us to conduct further tests to investigate the impact of semantic segmentation on the registration algorithm in this paper.
PointNet calculates the whole-scene point accuracy for each area in the datasets to assess the quality of semantic segmentation for the entire area. Table 12 displays the whole-scene point accuracy values for the six areas.
Table 12 shows that Area_6 has the best quality of semantic segmentation, while Area_5 has the worst. Three tests for the six areas are conducted with the rotational angles of the three axes being [30°, 30°, 30°], [60°, 60°, 60°], and [90°, 90°, 90°] respectively. The visualizations of the registration results when the rotational angle is equal to 30° are shown in Figure 5, and the test data are shown in Figure 6 and Table 13.
Figure 6 and Table 13 reveal that Area_6 has the smallest RMS under all three rotation angles, whereas Area_5 has the largest. This corresponds with the results of the PointNet semantic segmentation, where the whole-scene point accuracy is the highest in Area_6 and the lowest in Area_5. Similarly, the whole-scene point accuracy in Area_2 is higher than in Area_4, and the RMS in Area_2 is also smaller than in Area_4 under all three rotation angles. Based on what we have seen in practice, we can say that the proposed algorithm in this paper has a positive relationship between how well it registers and how well it semantically segments point clouds. Correct classifications in semantic segmentation are the only way to assign the correct weights in the registration algorithm’s weight distribution process, thereby enhancing the algorithm’s accuracy.
We conduct comparison experiments after determining the importance of labels in semantic segmentation. We set the translation to [0.1, 0.2, 0.15] for all tests. We set the change in angle in the rotational axis for each small rigid transformation in the weighted PP-ICP to 1°. Figure 7 and Table 14 display the root mean square error (RMS).
Figure 7 and Table 14 demonstrate that the proposed algorithm in this paper consistently achieves the smallest RMS in all tests, while the original ICP consistently demonstrates the highest RMS. This demonstrates the significant improvement in registration accuracy achieved by the algorithm proposed in this paper. On the other hand, as the rotation angle increases, the RMS of all the algorithms increases, but the weighted PP-ICP always has the smallest increase. This means that the algorithm is very stable when rigid bodies change shape at different rotation angles.
The multi-axis angle tests maintain the angle’s range between 15° and 120°, with a 15° step. For each type of rigid transformation, we rotate all three axes clockwise by the same angle value, such as [30°, 30°, 30°] and [60°, 60°, 60°]. Figure 8 and Table 15 display the tests.
Figure 8 and Table 15 demonstrate that in the multi-axis angle tests, the weighted PP-ICP consistently exhibits the smallest RMS, with the exception of the cases at 15° and 30°. Similar to the single-axis angle tests, the RMS increases as the rotation angle increases. However, we observe that in the multi-axis angle tests, the RMS of all the algorithms surpasses that of the single-axis angle tests, suggesting an increase in registration errors as the complexity of the rigid-body transformations increases.
We evaluate the runtime and computation resources used during the multi-axis tests for the four algorithms on a personal computer with a CPU processor (AMD Ryzen 7 4800H) and 16.0 GB of RAM to validate the efficiency. Figure 9 and Table 16 display the results.
Figure 9 and Table 16 conclude that the weighted PP-ICP algorithm in this paper exhibits significantly shorter runtimes than the original ICP, NDT, and PointNetLK. However, in comparison to the NDT, the proposed algorithm does not perform better in terms of CPU or memory usage. The main reason for this is that the NDT breaks up the original point cloud into a voxel grid to obtain Gaussian distributions through feature extraction. It then changes these distributions to the reference point cloud’s coordinate system [4]. Compared to ICP, which directly matches the original point cloud, the NDT significantly reduces computational complexity, thereby enhancing the hardware efficiency of the algorithm.

5. Discussion

We decompose a complete rigid-body transformation into a superposition of many tiny rigid-body transformations. For instance, a 120-degree rigid-body transformation is essentially a combination of numerous short-time, small-angle rigid-body transformations. By breaking down complex transformations into smaller, simpler components, we can better understand and analyze the overall motion of the rigid body. This decomposition method allows us to study the effects of each individual transformation and how they contribute to the overall movement of the object. By studying these small-angle transformations, we can gain insights into the mechanics and dynamics of rigid-body motion. Within each of these tiny rigid-body transformations, the angle is extremely small, enabling the application of approximation theory. This decomposition allows us to break down complex transformations into simpler, more manageable steps that can be analyzed and understood more easily. By applying approximation theory to each small angle transformation, we can accurately model and predict the behavior of the overall rigid-body transformation. This approach is particularly useful in robotics, computer graphics, and other fields where precise control and understanding of rigid-body movements are crucial. This small angle allows the algorithm to quickly iterate to the optimal solution. During the subsequent experimental validation phase, we discover that the algorithm’s running time increases with the size of the transformation angle, as it superimposes more tiny, rigid-body transformations. However, this decomposition fully utilizes the benefits of the new method for solving the iterative problem suggested in Section 3. Each small transformation exhibits high accuracy, which enhances the accuracy of the final superimposed changes. The original paper provides a theoretical derivation, and we set the amount of angular change in the rotation axis to 1° for each iteration of the algorithm in the subsequent experiments. This incremental approach allows us to achieve precise results with each iteration, ultimately leading to a highly accurate final transformation. By setting the rotation axis change to 1° for each iteration, we are able to control the granularity of the transformations and ensure that the algorithm runs efficiently. This method proves to be highly effective in solving the iterative problem and provides a solid foundation for future experiments and applications.

6. Conclusions

Many millimeter-wave radar applications, including autonomous driving, robotic navigation, and augmented reality, widely use point cloud registration. During the registration process, accuracy and effectiveness are two key factors that will have a direct impact on performance. This paper proposes a novel algorithm for point cloud registration. First, PointNet uses different category labels in the 3D indoor spaces to semantically segment the point cloud for registration. We introduce a statistical weighting strategy that assigns varying weights to the point cloud corresponding to the category labels. Then, during the registration process, we propose a new method to calculate the closed form of ICP iterative optimal solutions in three-dimensional spaces. This method addresses the issue in traditional registration algorithms where the linear approximation of the objective function reduces accuracy. We successfully calculate the closed-form solutions by applying the weights of the point cloud to the objective function solving process. We validate the accuracy and efficiency of the proposed algorithm on point cloud datasets of 3D indoor spaces, and the registration experiments reveal significant improvements in both accuracy and efficiency after incorporating the weighting strategy into the ICP algorithm. The technique may swiftly iterate to the ideal solution by breaking hard transformations into numerous tiny transformations, each with an infinitesimal angle. Furthermore, the approach to finding closed-form solutions presented in Section 4 may significantly improve the registration’s accuracy.
The dataset tests show that the RMS of different types of rigid transformation become much smaller compared to ICP, the NDT, and PointNetLK. On the other hand, compared to the original ICP, the algorithm’s runtime and usage of both CPU and memory improve. Overall, this paper’s proposed weighting-strategy-based point cloud registration algorithm demonstrates improved accuracy and increased efficiency in large-scale 3D indoor spaces.
While the novel algorithm in this paper, based on a weighting strategy, significantly improves the accuracy and efficiency of point cloud registration in 3D indoor spaces, it still falls short in terms of hardware resource utilization when compared to other registration algorithms like the NDT. Future work will focus on further optimizing the ICP iterative process to reduce computation costs, as well as exploring the integration of other more efficient registration algorithms with the proposed weighting strategy.

Author Contributions

Methodology, W.L., H.Z. and S.S.; Software, W.C. and S.S.; Formal analysis, W.L., H.Z. and X.L.; Writing—original draft, W.L.; Writing—review & editing, H.Z., W.C., X.L. and S.S.; Visualization, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Heilongjiang Province’s Key Research and Development Project: ’open competition mechanism to select the best candidates’, grant number 2023ZXJ03C02; and National Science Foundation of China, grant number 52075117.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Charles, R.Q.; Su, H.; Kaichun, M.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  2. Arun, K.S.; Huang, T.S.; Blostein, S.D. Least-squares fitting of two 3-D point sets. IEEE Trans. Pattern Anal. Mach. Intell. 1987, 5, 698–700. [Google Scholar] [CrossRef] [PubMed]
  3. Besl, P.J.; McKay, N.D. Method for registration of 3-D shapes. Sens. Fusion IV Control Paradig. Data Struct. 1992, 1611, 586–606. [Google Scholar] [CrossRef]
  4. Rusinkiewicz, S.; Levoy, M. Efficient variants of the ICP algorithm. In Proceedings of the Third International Conference on 3-D Digital Imaging and Modeling, Quebec City, QC, Canada, 28 May–1 June 2001. [Google Scholar]
  5. Magnusson, M.; Andreasson, H.; Nuchter, A.; Lilienthal, A.J. Appearance-based loop detection from 3D laser data using the normal distributions transform. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009. [Google Scholar]
  6. Aoki, Y.; Goforth, H.; Srivatsan, R.A.; Lucey, S. Pointnetlk: Robust & efficient point cloud registration using pointnet. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  7. Chang, C.H.; Chou, C.N.; Chang, E.Y. CLKN: Cascaded Lucas-Kanade Networks for Image Alignment. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  8. Low, K.L. Linear Least-Squares Optimization for Point-to-Plane Icp Surface Registration; University of North Carolina: Chapel Hill, NC, USA, 2004. [Google Scholar]
  9. Serafin, J.; Grisetti, G. NICP: Dense normal based point cloud registration. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015. [Google Scholar]
  10. Yang, J.; Li, H.; Campbell, D.; Jia, Y. Go-ICP: A globally optimal solution to 3D ICP point-set registration. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 2241–2254. [Google Scholar] [CrossRef] [PubMed]
  11. Deschaud, J.E. IMLS-SLAM: Scan-to-model matching based on 3D data. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018. [Google Scholar]
  12. Das, A.; Waslander, S.L. Scan registration with multi-scale k-means normal distributions transform. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012. [Google Scholar]
  13. Lu, W.; Wan, G.; Zhou, Y.; Fu, X.; Yuan, P.; Song, S. Deepvcp: An end-to-end deep neural network for point cloud registration. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  14. Li, T.; Shi, Y.; Sun, X.; Wang, J.; Yin, B. PCGAN: Prediction-Compensation Generative Adversarial Network for Meshes. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4667–4679. [Google Scholar] [CrossRef]
  15. Bai, X.; Luo, Z.; Zhou, L.; Chen, H.; Li, L.; Hu, Z.; Fu, H.; Tai, C.-L. Pointdsc: Robust point cloud registration using deep spatial consistency. In Proceedings of the IEEE/CVF Conference on Computer Visio and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15859–15869. [Google Scholar]
  16. Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.-L. D3Feat: Jointlearning of dense detection and description of 3D local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  17. Zhang, Z.; Dai, Y.; Sun, J. Deep learning based point cloud registration: An overview. Virtual Real. Intell. Hardw. 2020, 2, 222–246. [Google Scholar] [CrossRef]
  18. Pais, G.D.; Ramalingam, S.; Govindu, V.M.; Nascimento, J.C.; Chellappa, R.; Miraldo, P. 3DRegNet: A deep neural network for 3D point registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  19. Chen, P.; Li, X. Revised spectral matching algorithm for scenes with mutually inconsistent local transformations. IET Image Processing. 2015, 9, 916–922. [Google Scholar] [CrossRef]
  20. Huang, X.; Mei, G.; Zhang, J.; Abbas, R. A comprehensive survey on point cloud registration. arXiv 2021, arXiv:2103.02690. [Google Scholar]
  21. Pomerleau, F.; Colas, F.; Siegwart, R. A review of point cloud registration algorithms for mobile robotics. Found. Trends® Robot. 2015, 4, 1–104. [Google Scholar] [CrossRef]
  22. Cheng, L.; Chen, S.; Liu, X.; Xu, H.; Wu, Y.; Li, M.; Chen, Y. Registration of laser scanning point clouds: A review. Sensors 2018, 18, 1641. [Google Scholar] [CrossRef] [PubMed]
  23. Saiti, E.; Theoharis, T. An application independent review of multimodal 3d registration methods. Comput. Graph. 2020, 91, 153–178. [Google Scholar] [CrossRef]
  24. Zhou, Q.-Y.; Park, J.; Koltun, V. Fast global registra tion. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2016; pp. 766–782. [Google Scholar]
  25. Huang, X.; Fan, L.; Wu, Q.; Zhang, J.; Yuan, C. Fast registration for cross-source point clouds by using weak regional affinity and pixel-wise refinement. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 1552–1557. [Google Scholar]
  26. Xin, S.; Wang, Y.; Lv, W. Standard deviation control chart based on weighted standard deviation method. In Proceedings of the 33rd Chinese Control Conference, Nanjing, China, 28–30 July 2014. [Google Scholar]
  27. Liu, Z.; Zhang, Z.; Zhang, X.; Lin, X.; Zhao, H.; Shen, Z. A Path Analysis of Electric Heating Technology Based on Entropy Weight Method. In Proceedings of the 2021 IEEE 5th Conference on Energy Internet and Energy System Integration (EI2), Taiyuan, China, 22–24 October 2021. [Google Scholar]
  28. Diakoulaki, D.; Mavrotas, G.; Papayannakis, L. Determining objective weights in multiple criteria problems: The critic method. Comput. Oper. Res. 1995, 22, 763–770. [Google Scholar] [CrossRef]
  29. Xu, Y.; Cai, Z. Standard deviation method for determining the weights of group multiple attribute decision making under uncertain linguistic environment. In Proceedings of the 2008 7th World Congress on Intelligent Control and Automation, Chongqing, China, 25–27 June 2008. [Google Scholar]
  30. Wang, J.; Hao, Y.; Yang, C. An entropy weight-based method for path loss predictions for terrestrial services in the VHF and UHF bands. Radio Sci. 2023, 58, 1–11. [Google Scholar] [CrossRef]
  31. Wang, Z.; Yang, X.; Hu, H.; Lou, Y. Actor-critic method-based search strategy for high precision peg-in-hole tasks. In Proceedings of the 2019 IEEE International Conference on Real-Time Computing and Robotics (RCAR), Irkutsk, Russia, 4–9 August 2019. [Google Scholar]
  32. Armeni, I.; Sener, O.; Zamir, A.R.; Jiang, H.; Brilakis, I.; Fischer, M.; Savarese, S. 3D semantic parsing of large-scale indoor spaces. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar]
  33. Zhou, D.; Fang, J.; Song, X.; Guan, C.; Yin, J.; Dai, Y.; Yang, R. Iou loss for 2d/3d object detection. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Quebec City, QC, Canada, 16–19 September 2019. [Google Scholar]
  34. Bressan, P.O.; Junior, J.M.; Martins, J.A.C.; de Melo, M.J.; Gonçalves, D.N.; Freitas, D.M.; Ramos, A.P.M.; Furuya, M.T.G.; Osco, L.P.; de Andrade Silva, J.; et al. Semantic segmentation with labeling uncertainty and class imbalance applied to vegetation maping. Int. J. Appl. Earth Obs. Geoinf. 2022, 108, 102690. [Google Scholar]
  35. Marom, N.D.; Rokach, L.; Shmilovici, A. Using the confusion matrix for improving ensemble classifiers. In Proceedings of the 2010 IEEE 26-th Convention of Electrical and Electronics Engineers in Israel, Eilat, Israel, 17–20 November 2010. [Google Scholar]
  36. Baidya, J.; Garg, H.; Saha, A.; Mishra, A.R.; Rani, P.; Dutta, D. Selection of third party reverses logistic providers: An approach of BCF-CRITIC-MULTIMOORA using Archimedean power aggregation operators. Complex Intell. Syst. 2021, 7, 2503–2530. [Google Scholar] [CrossRef]
  37. Sharkasi, N.; Rezakhah, S. A modified CRITIC with a reference point based on fuzzy logic and hamming distance. Knowl. Based Syst. 2022, 255, 109768. [Google Scholar] [CrossRef]
  38. Censi, A. An ICP variant using a point-to-line metric. In Proceedings of the 2008 IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, 19–23 May 2008. [Google Scholar]
Figure 1. Error metrics: (a) point-to-point; (b) point-to-plane.
Figure 1. Error metrics: (a) point-to-point; (b) point-to-plane.
Applsci 14 05240 g001
Figure 2. Process of weighted PP-ICP.
Figure 2. Process of weighted PP-ICP.
Applsci 14 05240 g002
Figure 3. Point cloud semantic segmentation of S3DIS: (a) Area_1 (conference room); (b) Area_2 (hallway); (c) Area_3 (office); (d) Area_4 (lobby); (e) Area_5 (storage); (f) Area_6 (open space).
Figure 3. Point cloud semantic segmentation of S3DIS: (a) Area_1 (conference room); (b) Area_2 (hallway); (c) Area_3 (office); (d) Area_4 (lobby); (e) Area_5 (storage); (f) Area_6 (open space).
Applsci 14 05240 g003aApplsci 14 05240 g003b
Figure 4. RMS of weighted PP-ICP for different groups of label importance.
Figure 4. RMS of weighted PP-ICP for different groups of label importance.
Applsci 14 05240 g004
Figure 5. Registration results of S3DIS datasets: (a) Area_1; (b) Area_2; (c) Area_3; (d) Area_4; (e) Area_5; (f) Area_6. Rotational angle = [30°, 30°, 30°].
Figure 5. Registration results of S3DIS datasets: (a) Area_1; (b) Area_2; (c) Area_3; (d) Area_4; (e) Area_5; (f) Area_6. Rotational angle = [30°, 30°, 30°].
Applsci 14 05240 g005
Figure 6. RMS for six areas.
Figure 6. RMS for six areas.
Applsci 14 05240 g006
Figure 7. RMS of single-axis angle tests: (a) x-axis; (b) y-axis; (c) z-axis.
Figure 7. RMS of single-axis angle tests: (a) x-axis; (b) y-axis; (c) z-axis.
Applsci 14 05240 g007
Figure 8. RMS of multi-axis angle tests.
Figure 8. RMS of multi-axis angle tests.
Applsci 14 05240 g008
Figure 9. Runtime and computation resource costs during multi-axis tests: (a) runtime; (b) CPU usage; (c) memory usage.
Figure 9. Runtime and computation resource costs during multi-axis tests: (a) runtime; (b) CPU usage; (c) memory usage.
Applsci 14 05240 g009
Table 1. The labels of point clouds in 3D indoor spaces.
Table 1. The labels of point clouds in 3D indoor spaces.
ClassCeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
No.0123456789101112
Table 2. The importance of labels in 3D indoor spaces.
Table 2. The importance of labels in 3D indoor spaces.
ClassCeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
IM0.20.20.20.20.60.60.4110.80.60.60.1
Table 3. The average values of alternatives and criteria for six areas.
Table 3. The average values of alternatives and criteria for six areas.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
IM0.20.20.20.20.60.60.4110.80.60.60.1
IoU0.9580.9770.8430.8180.5140.8250.7580.7860.8590.7570.7990.8080.728
CPA0.9770.9870.9340.8930.5870.8930.8240.8620.9010.7860.8870.8880.848
Table 4. Decision matrix in fuzzy form.
Table 4. Decision matrix in fuzzy form.
Criteria IMIoUCPA
Ceiling D i j L 0.08870.44370.4526
D i j H 0.09870.45370.4626
Floor D i j L 0.08740.44650.4511
D i j H 0.09740.45650.4611
Wall D i j L 0.09620.42140.4674
D i j H 0.10620.43140.4774
Beam D i j L 0.09970.42300.4623
D i j H 0.10970.43300.4723
Column D i j L 0.34770.29720.3401
D i j H 0.35770.30720.3501
Window D i j L 0.25380.35090.3802
D i j H 0.26380.36090.3902
Door D i j L 0.19680.37740.4107
D i j H 0.20680.38740.4207
Table D i j L 0.37260.29180.3205
D i j H 0.38260.30180.3305
Chair D i j L 0.35730.30620.3214
D i j H 0.36730.31620.3314
Sofa D i j L 0.33640.31810.3305
D i j H 0.34640.32810.3405
Case D i j L 0.25750.34450.3830
D i j H 0.26750.35450.3930
Board D i j L 0.25630.34690.3818
D i j H 0.26630.35690.3918
Clutter D i j L 0.05470.42940.5010
D i j H 0.06470.43940.5110
Table 5. Boundaries of each criterion and corresponding fuzzy RP points.
Table 5. Boundaries of each criterion and corresponding fuzzy RP points.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
Lj0.3280.3280.3280.3280.3280.3280.3280.3280.3280.3280.3280.3280.328
Hj0.3380.3380.3380.3380.3380.3380.3380.3380.3380.3380.3380.3380.338
DjL0.0890.0870.0960.1000.2970.2540.1970.2920.3060.3180.2580.2560.055
DjH+0.4630.4610.4770.4720.3580.3900.4210.3830.3670.3460.3930.3920.511
Table 6. Normalized decision matrix.
Table 6. Normalized decision matrix.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
IM0.9110.9100.9120.9150.9990.9900.9710.9960.9980.7590.9890.9760.875
IoU0.9570.9560.9640.9650.9980.9970.9890.9970.9990.8620.9980.9970.954
CPA0.094 0.087 0.086 0.079 0.268 0.252 0.176 0.316 0.309 0.318 0.174 0.218 0.045
Table 7. Standard variances of criteria.
Table 7. Standard variances of criteria.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
σ j 0.0290.0240.0300.0010.00080.00450.0130.0350.0390.0310.0070.0030.0001
Table 8. Distance correlations between different pairs of criteria.
Table 8. Distance correlations between different pairs of criteria.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
Ceiling1.0000.1990.5290.1310.9650.4660.8220.1800.2490.4730.6730.0920.213
Floor0.1991.0000.2630.7220.3230.4800.1240.1100.8570.0290.4330.5180.304
Wall0.5290.2631.0000.4280.4060.4890.3420.4220.3460.5100.7590.0080.652
Beam0.1310.7220.4281.0000.3040.1750.4080.3910.1280.7010.4520.9470.558
Column0.9650.3230.4060.3041.0000.7280.2870.7760.2550.9180.6870.5110.826
Window0.4660.480.4890.1750.7281.0000.4920.8360.7290.0500.3560.1140.005
Door0.8220.1240.3420.4080.2870.4921.0000.0880.8480.9790.867−0.0950.539
Table0.1800.1100.4220.3910.7760.8360.0881.0000.0860.9720.0100.6010.591
Chair0.2490.8570.3460.1280.2550.7290.8480.0861.0000.5810.4500.7120.556
Sofa0.4730.0290.5100.7010.9180.0500.9790.9720.5811.0000.6470.5310.737
Case0.6730.4330.7590.4520.6870.3560.8670.0100.4500.6471.0000.4370.404
Board0.0920.5180.0080.9470.5110.1140.0950.6010.7120.5310.4371.0000.482
Clutter0.2130.3040.6520.5580.8260.0050.5390.5910.5560.7370.4040.4821.000
Table 9. Information amounts, weights, and ranks of criteria.
Table 9. Information amounts, weights, and ranks of criteria.
CeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
Qj0.072 0.077 0.112 0.008 0.004 0.011 0.027 0.147 0.116 0.098 0.011 0.017 0.001
ωj0.103 0.110 0.160 0.011 0.006 0.016 0.039 0.210 0.165 0.140 0.015 0.024 0.001
Ranks65312119712410813
Table 10. Groups of importance of labels in 3D indoor spaces.
Table 10. Groups of importance of labels in 3D indoor spaces.
GroupCeilingFloorWallBeamColumnWindowDoorTableChairSofaCaseBoardClutter
10.20.20.20.20.60.60.41.01.00.80.60.60.1
20.40.40.40.40.40.40.60.80.80.60.40.40.3
30.60.60.60.60.20.20.80.60.60.40.20.20.5
40.80.80.80.80.10.11.00.40.40.20.10.10.7
Table 11. RMS of weighted PP-ICP for different groups of label importance.
Table 11. RMS of weighted PP-ICP for different groups of label importance.
Angles (°)Group 1Group 2Group 3Group 4
150.3330.7151.9593.674
300.4790.8592.0743.627
450.4470.852.0763.628
600.4241.2272.2183.367
750.6631.2972.4593.289
900.6461.6623.1342.665
1050.7621.0542.3473.001
1200.7281.6152.0172.503
Table 12. Whole-scene point accuracy for six areas.
Table 12. Whole-scene point accuracy for six areas.
Area_1Area_2Area_3Area_4Area_5Area_6
0.9460.9510.9350.9230.9020.974
Table 13. RMS for six areas.
Table 13. RMS for six areas.
AnglesArea_1Area_2Area_3Area_4Area_5Area_6
30°0.5110.3690.5410.3840.6020.197
60°0.5610.5200.5040.6560.7960.305
90°0.7260.4150.8810.4330.9040.397
Table 14. RMS of single-axis angle tests for the algorithms.
Table 14. RMS of single-axis angle tests for the algorithms.
Angles (°)AxisICPNDTPointNetLKWeighted PP-ICP
15x0.1120.0640.0320.015
y0.1110.0420.0260.015
z0.1050.0150.0380.001
30x0.2020.1590.1070.066
y0.2140.1180.1030.069
z0.1790.0770.1110.037
45x0.2750.2410.1710.101
y0.2710.2250.1840.129
z0.2530.1280.1850.081
60x0.4120.3160.2370.134
y0.3920.2850.3140.177
z0.3290.1890.2560.126
75x0.4510.4210.3220.197
y0.4570.3760.3210.225
z0.4250.2570.2910.149
90x0.5660.5190.3980.211
y0.5730.4340.3520.262
z0.4950.3210.4040.176
105x0.6410.6210.4370.265
y0.6580.5380.5580.319
z0.5870.3730.3930.212
120x0.6920.6530.5410.273
y0.7080.5770.5680.364
z0.6290.4120.4530.276
Table 15. RMS of multi-axis angle tests for the algorithms.
Table 15. RMS of multi-axis angle tests for the algorithms.
Angles (°)ICPNDTPointNetLKWeighted PP-ICP
150.3580.4230.2420.375
300.4730.5120.3590.434
450.5890.5970.4930.491
600.6450.6770.5760.557
750.8320.7330.6980.651
900.9860.8290.7910.626
1051.1290.8940.9540.717
1201.1770.9731.0880.787
Table 16. Runtime and computation resource costs during multi-axis tests.
Table 16. Runtime and computation resource costs during multi-axis tests.
Indicators ICPNDTPointNetLKWeighted PP-ICP
Time (s)15°94.56192.65677.07250.233
30°157.958192.108141.939108.381
45°254.401283.051214.551154.845
60°322.314374.898289.913201.934
75°383.113459.502353.857247.719
90°445.865541.271427.142311.205
105°543.583623.247497.156356.182
120°608.738717.924563.291407.393
CPU Usage (%)15°76.9799.43369.55764.061
30°79.59512.16667.12365.274
45°82.98816.43770.85867.763
60°79.43121.71479.16667.589
75°90.21126.92481.87171.298
90°94.43925.02272.53979.271
105°87.99134.42183.14382.918
120°99.99631.68885.50381.985
Memory Usage (%)15°13.6160.35712.0517.574
30°17.8896.62315.9779.417
45°22.2525.62819.07711.785
60°27.02512.49817.44813.237
75°33.71910.49120.01117.593
90°37.19915.37113.17419.088
105°44.99819.99921.45621.618
120°49.52320.21526.03419.893
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lv, W.; Zhang, H.; Chen, W.; Li, X.; Sang, S. A Point Cloud Registration Algorithm Based on Weighting Strategy for 3D Indoor Spaces. Appl. Sci. 2024, 14, 5240. https://doi.org/10.3390/app14125240

AMA Style

Lv W, Zhang H, Chen W, Li X, Sang S. A Point Cloud Registration Algorithm Based on Weighting Strategy for 3D Indoor Spaces. Applied Sciences. 2024; 14(12):5240. https://doi.org/10.3390/app14125240

Chicago/Turabian Style

Lv, Wenshan, Haifeng Zhang, Weiren Chen, Xiaoming Li, and Shengtian Sang. 2024. "A Point Cloud Registration Algorithm Based on Weighting Strategy for 3D Indoor Spaces" Applied Sciences 14, no. 12: 5240. https://doi.org/10.3390/app14125240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop