Next Article in Journal
Generating Daily Land Surface Temperature Downscaling Data Based on Sentinel-3 Images
Next Article in Special Issue
Dual-Branch Attention-Assisted CNN for Hyperspectral Image Classification
Previous Article in Journal
Estimating Rainfall from Surveillance Audio Based on Parallel Network with Multi-Scale Fusion and Attention Mechanism
Previous Article in Special Issue
A Lightweight Multi-Level Information Network for Multispectral and Hyperspectral Image Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration

1
College of Computer and Information Science, Fujian Agriculture and Forestry University, Fuzhou 350002, China
2
Digital Fujian Research Institute of Big Data for Agriculture and Forestry, Fujian Agriculture and Forestry University, Fuzhou 350002, China
3
Key Laboratory of Smart Agriculture and Forestry, Fujian Agriculture and Forestry University, Fuzhou 350002, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(22), 5751; https://doi.org/10.3390/rs14225751
Submission received: 12 August 2022 / Revised: 3 November 2022 / Accepted: 8 November 2022 / Published: 14 November 2022
(This article belongs to the Special Issue Deep Learning for the Analysis of Multi-/Hyperspectral Images)

Abstract

:
Seeking reliable correspondences between two scenes is crucial for solving feature-based point cloud registration tasks. In this paper, we propose a novel outlier rejection network, called Channel-Spatial Contextual Enhancement Network (CSCE-Net), to obtain rich contextual information on correspondences, which can effectively remove outliers and improve the accuracy of point cloud registration. To be specific, we design a novel channel-spatial contextual (CSC) block, which is mainly composed of the Channel-Spatial Attention (CSA) layer and the Nonlocal Channel-Spatial Attention (Nonlocal CSA) layer. The CSC block is able to obtain more reliable contextual information, in which the CSA layer can selectively aggregate the mutual information between the channel and spatial dimensions. The Nonlocal CSA layer can compute feature similarity and spatial consistency for each correspondence, and the CSA layer and Nonlocal CSA layer can support each other. In addition, to improve the distinguishing ability between inliers and outliers, we present an advanced seed selection mechanism to select more dependable initial correspondences. Extensive experiments demonstrate that CSCE-Net outperforms state-of-the-art methods for outlier rejection and pose estimation tasks on public datasets with varying 3D local descriptors. In addition, the network parameters of CSCE-Net are reduced from 1.05M to 0.56M compared to the recently learning-based outlier rejection method PointDSC.

Graphical Abstract

1. Introduction

Point cloud registration is widely used in 3D reconstruction [1], object pose estimation [2], simultaneous localization and mapping, [3] and other fields [4,5,6]. The feature-based point cloud registration pipelines commonly start from keypoint detection [7,8,9], feature point description [10,11,12], followed by robust alignment by outlier rejection [13,14,15]. Although feature point detection and 3D local features have developed rapidly, there are still many outliers in the initial correspondences generated by the feature-based matching, especially when the overlap of the two point clouds is very small. In this paper, we mainly design an outlier rejection network, which is a critical step in the robust point cloud registration pipeline, to alleviate the above issue.
Recently, 3D outlier rejection tasks based on deep learning, such as 3DRegNet [13] and DGR [14], formulated the outlier rejection as a classification problem of inlier/outlier. These networks embed deep features of input correspondences and utilize the resulting features to predict the probability that each correspondence is an inlier to remove the outlier. For feature embedding, they solely rely on general descriptors such as pointwise multilayer perceptron (MLP) and sparse convolution to capture contextual feature information of each correspondence. They all ignore the important spatial consistency properties that the inlier should follow in geometric properties in the 3D domain. Instead, PointDSC [15] considers the length of spatial consistency in the feature embedding of each correspondence. The PointDSC combines the length consistency and the spatial feature similarity to capture contextual features about each input correspondence. However, the above networks still cannot focus on essential channels contextual for each correspondence. In addition, they also lack crucial spatial information that would complement the channel features and effectively highlight areas of important spatial feature information extracted along the channel axis. Therefore, each correspondence has poor contextual feature representation in the above networks. In particular, ignoring the channel-spatial global context information would enormously limit the ability of the network to distinguish each correspondence, especially when the point clouds of pairs are less overlapping. As a result, these networks retain more outliers and, thus, reduce the point cloud alignment accuracy. Observing Figure 1a, we find that the outlier rejection network PointDSC retains many outliers due to inadequate extraction of contextual features, which results in huge point cloud alignment errors and consequent registration failures.
In this paper, we propose a practical channel-spatial contextual (CSC) block to solve the above issues. CSC block can capture each correspondence’s channel-spatial global context information by computing complementary channel and spatial attention. We also import the spatial consistency and the channel-spatial attention into the nonlocal network to compute the feature similarity for each pair of correspondences in spatial dimension based on the channel and spatial feature fusion. Therefore, the CSC block can separately calculate the channel-spatial features of each correspondence and the correlation between the correspondences, which enables the network to extract sufficient contextual information and process each correspondence distinguishably. Furthermore, we propose an advanced Seed Selection mechanism that more accurately calculates each correspondence’s initial confidence to select the seeds with well-distributed and highly confident, which contributes to the construction of a more reliable consensus set for each seed. Integrating the newly proposed CSC block and advanced Seed Selection mechanism with the Second Order Spatial Compatibility (SC 2 ) measure [16] and Spectral Matching (SM) [17], we propose a novel outlier rejection network Channel-Spatial Contextual Enhancement Network (CSCE-Net). Specifically, CSCE-Net introduces a CSC block to capture rich global context information and enhance the presentation capability of important channel-spatial map information. In addition, CSCE-Net designs an advanced seeding mechanism to select high-confidence correspondences for the SC 2 measure and SM to improve the inlier/outlier discrimination. In addition, we iteratively used CSCE-Net twice, with the inlier probability obtained from the previous network serving as an additional input to the following network to improve the performance of the final network. Observing Figure 1b, it is not difficult to see that our CSCE-Net is able to retain more inliers in two low-overlap scenes, dramatically reducing alignment errors when performing positional estimation compared to Figure 1a. The quantitative and visual results provide excellent evidence of the effectiveness of our CSCE-Net in seeking reliable correspondence during point cloud registration.
The main contributions of this work are in three aspects:
  • We propose a CSC block in which the CSA layer and Nonlocal CSA layer are backbones to capture sufficient contextual information and enhance the representative ability of inliers and important channels and spatial.
  • We construct an advanced seed selection mechanism to more accurately calculate seeds with high confidence and well-distributed and facilitate the construction of more reliable consensus sets for each seed.
  • Our CSCE-Net outperforms state-of-the-art methods for outlier rejection and pose estimation tasks on two challenging public datasets by combining different 3D local descriptors and better balancing the number of parameters in the network.
The rest of this article is structured as follows: We describe the related literature in Section 2. In Section 3, we introduce the structure information of CSCE-Net, CSC block, and Seed Selection in detail. We exhibit the experimental details and comparative results in Section 4. In the end, we give the relevant conclusions in Section 5.

2. Related Works

In feature-based point cloud registration, firstly, a putative correspondence set has been built by some 3D local features, such as fast point feature histograms (FPFH) [18] and fully convolutional geometric features (FCGF) [10]. Secondly, it is necessary to remove outliers, as the putative correspondence set usually has a large number of incorrect matches. In this section, we briefly introduce some traditional and learning-based outlier rejection methods in Section 2.1 and Section 2.2, respectively.

2.1. Traditional Outlier Rejection

Random Sample Consensus (RANSAC) [19] is still the more popular method of outlier rejection. It iteratively uses the smallest subset obtained by random sampling and employs a least-squares fitting that is not robust to outliers in the model fitting step. Many of its various variants [20,21,22] have introduced new sampling strategies and local optimizations to speed up estimation or improve robustness. LO-RANSAC [20] makes the parametric model calculated for the outlier-free sample consistent with all interior points by applying local optimization to the solution estimated for the random sample. Since most existing registration methods require a set of assumed correspondences obtained by extracting invariant descriptors, SDRSAC [21] presents a new stochastic algorithm for robust point cloud alignment without correspondences. GESAC [22] improves on classical RANSAC by generating a larger subset in the sampling phase and introducing a shape-annealing robust estimate in the model fitting step. However, their main drawback, as with RANSAC, continues to be slow convergence and low accuracy in the case of a relatively large outlier ratio. More recently, Graph-cut RANSAC [23] runs a graph-cutting algorithm in a local optimization step to perform efficient outlier separation. Magsac [24] introduces a σ -consensus to construct a without inlier-outlier threshold method for RANSAC. As a result, they can improve the efficiency of outliers rejection further.
In addition, spectral matching (SM) [17] is a famous traditional algorithm that relies heavily on 3D spatial consistency to find inlier correspondences. It uses length consistency to construct a compatibility graph and then finds the main clusters of the graph to obtain an inlier set through eigenanalysis. However, as explained in [25], it cannot effectively deal with high outlier ratios, in which case the main internal clusters become less obvious. Moreover, FGR [26] achieves a fast global alignment algorithm for locally overlapping 3D surfaces by manipulating the candidate matches on the covered surface. TEASER [27] reformulates the registration problem using a truncated least squares (TLS) cost and solving the 3D rotation transformation through a general graph-theoretic framework. FGR and TEASER can tolerate outliers using robust loss functions (e.g., GemanMcClure). Quatro [28] utilizes a global alignment method based on degenerate robust decoupling to solve the problem of a large number of anomalous correspondences. Most traditional outlier rejection methods have been summarized in [25,29,30,31,32].
Although widely used in computer vision, traditional outlier rejection methods still suffer from some fundamental drawbacks. For example, as the ratio of outliers in the initial matching set increases dramatically, the above-mentioned traditional algorithms struggle to balance accuracy and efficiency. Hence comes the emergence of deep learning-based outlier rejection.

2.2. Learning-Based Outlier Rejection

In learning-based approaches, some feature learning methods [8,9,10,12] have used convolutional neural networks (CNN) to acquire more reliable keypoints and descriptors. Although these methods have better results than manual methods [18], they do not solve the problem of having a large number of outliers in the initial correspondence set. Therefore, learning-based outlier removal methods are proposed as a post-processing step to solving this problem.
A learning-based outlier rejection approach is introduced for the first time in the 2D image matching area, [33,34,35,36,37,38] where outlier rejection is represented as an inlier/outlier classification problem. CN-Net [33] is the first method to use an end-to-end learning-based approach to label each correspondence as an inlier or outlier value. Meanwhile, the fundamental matrix for recovering the camera pose of two matched images is estimated using a weighted eight-point algorithm. NM-Net [34] redefines neighbors to find reliable correspondence. OANet [35] introduces an order-aware filtering block to cluster correspondences crucial global context. MS2DG-Net [36] uses sparse semantic similarity instead of Euclidean distance to generate dynamic contextual features.
Recent attempts [13,14,15,39,40] have also introduced deep learning networks to perform 3D correspondence pruning. The 3DRegNet [13] redefines the CN-Net [33] into a 3D form and devises a regression module to solve the rigid transformation. DGR [14] recommends the use of a learning-based feature called the full convolutional geometric feature (FCGF) [10] to perform the alignment and employs a 6D convolutional network to predict the likelihood of each correspondence. DetarNet [39] introduces a decoupled solution for translation and rotation, resulting in better performance for both. Using Hough voting in 6D transformed parameter space, DHVR [40] describes a robust and efficient framework for paired registration of realistic 3D scans. However, when performing feature embedding of correspondences and pruning, these networks ignore the important property of spatial consistency that 3D point clouds should follow in the 3D domain. In contrast, the recent PointDSC [15] has developed a non-local module and neural spectrum matching based on the length spatial consistency to accelerate model generation.
However, these networks still do not capture enough contextual information for network learning because their feature embedding networks ignore each correspondence’s channel and spatial features, thus, limiting the feature representation capability. In this paper, our CSCE-Net uses complementary spatial channel attention combined with length spatial consistency to adequately extract the contextual features of each correspondence to improve the performance of outlier rejection and hence the accuracy of the final alignment.

3. Method

In this section, we first introduce the problem formulation in Section 3.1. Then, the network framework is introduced in Section 3.2. Finally, we describe the proposed CSC block and the Seed Selection in Section 3.3 and Section 3.4, respectively.

3.1. Problem Formulation

In this work, we are given two sets of keypoints U R 3 and V R 3 from a pair of partially overlapping 3D point clouds and each keypoint comes with an associated local descriptor. After that, the putative correspondence set S of the network inputs are generated by performing a nearest neighbor search using local descriptors. Each correspondence s i S is denoted as s i = ( u i , v i ) ∈ R 6 , where u i U , v i V are the space coordinates of a pair of 3D keypoints. Our final goal is to find an outlier/inlier label value for s i , i.e., w i = 0 or 1, respectively, and then utilize these inliers between two point sets to recover an optimal 3D rigid transformation R ^ , t ^ .
Specifically, we first embed the correspondence of two point sets into the high-dimensional feature. Next, we employ these features to find the high-confidence correspondences as seeds. The corresponding consensus set (i.e., S ) of each seed is obtained by the SC 2 measure. Then we compute the 3D rigid transformation { R , t } for each seed set by the least-squares fitting [41]:
R , t = argmin R , t i S α i R u i + t v i 2 ,
where α is the weight of inlier probability calculated from the NSM module. Equation (1) is solved by SVD [41].
Further, we find the optimal 3D rigid transformation { R ^ , t ^ } by the number of corresponding relations for each { R , t } satisfying a given threshold:
R ^ , t ^ = argmax R , t i S [ R u i + t v i < μ ] ,
where μ denotes the inlier threshold and [·] is the Iverson bracket. Finally, the outlier/inlier label w R S is obtained by w i = [ R ^ u i + t ^ v i < μ ] . We then recompute the final 3D rotation transformation { R ^ , t ^ } by the least squares using all the retained inliers.
Loss Function. Following previous work in PointDSC [15], we optimize the neural network utilizing a hybrid loss function (i.e., a node-wise loss and an edge-wise loss) as follows:
L = β L n o d e + L e d g e ,
where β is a hyper-parameter that balances the two losses.
L n o d e denotes the node-wise loss that is used to supervise each correspondence individually. We employ binary cross-entropy loss as node-wise supervision to learn the initial confidence:
L n o d e = BCE ( θ , w ˜ ) ,
where θ is the initial confidence of the prediction, and w ˜ represents the ground-truth outlier/inlier labels constructed by
w ˜ i = [ R ˜ u i + t ˜ v i < μ ] ,
where R ˜ and t ˜ are the ground truth rotation and translation matrices, and μ is the inlier threshold.
L e d g e is the edge-wise loss, which is applied to supervise the pairwise relationship between correspondences as a complement to node-wise supervision.
L e d g e = 1 1 S 2 i j ( δ i j δ ˜ i j ) 2 ,
where δ i j is the relevance value estimated based on feature similarity and δ ˜ i j =[ s i and s j are inliers] is the ground-truth relevance value, which is defined as:
δ i j = [ 1 1 ϵ f 2 f ¯ i f ¯ j 2 ] ,
where f ¯ i and f ¯ j are the L 2 -normalized eigenvectors, and ϵ f is a parameter of feature difference sensitivity.

3.2. Network Framework

The framework of our CSCE-Net is shown in Figure 2. The CSCE-Net consists of four main parts: our proposed CSC block, the presented seed selection, the Second Order Spatial Compatibility Measure, and Spectral Matching. Our CSCE-Net starts with N initial correspondences as input and maps the features from 6 to 128 dimensions using a shared perceptron. Next, it adequately extracts the global context information for each correspondence by utilizing 6 CSC blocks. Then, the seed selection mechanism exploits the high-dimensional features obtained from the CSC blocks to calculate the initial confidence of each correspondence and selects these high-confidence and well-distributed correspondences as seeds ( Ns ) using the Non-Maximum Suppression method. Seeds selected for rich contextual information are more likely to be represented as inliers, which contributes to the construction of a more reliable consensus set. In addition, to better distinguish between insertions and outliers, we inserted a second-order spatial compatibility (SC 2 ) measure proposed by SC 2 -PCR [16], which constructs a new global measure of similarity between two correspondences by counting the number of correspondences that are compatible with both correspondences simultaneously. The role of the SC 2 measure is to construct a consensus set for each seed. Finally, CSCE-Net adopts the spectral matching [16,17] to compute the weight vector α of inlier probabilities and the 3D rigid transform { R , t } for each consensus set by least-squares fitting.

3.3. CSC Block

The CSC block mainly utilizes perceptrons and complementary channel-spatial attention to selectively aggregate information in the channel and spatial dimensions, capture the complex global context of feature maps, and obtain feature maps with solid and expressive capabilities. As shown in Figure 2, the newly proposed CSC block consists of four key parts, i.e., the two identical Context Norm layers, the CSA layer, and the Nonlocal CSA layer. The Context Norm layer contains a Context Normalization layer to get the global context, a Batch Normalization layer with ReLU activation function to speed up network training, and a Multi-Layer Perceptrons layer containing 128 neurons for network learning. We use the Context Norm layer to handle disordered and irregular correspondences.
As shown in Figure 3, the CSA layer reinforces the channel-spatial features by serially concatenating the 2D maps produced by the Channel attention (CA) module and the Spatial attention (SA) module, respectively. The lack of channel-spatial features in feature embeddings greatly limits the ability of the network to reject outliers. Therefore, we construct a novel channel-spatial attention layer to improve the representation ability and emphasize significant characteristics along these two main dimensions: channel and spatial. For aggregating spatial information, the average pooling method is generally used at present [42]. Moreover, Woo et al. [43] believe that max-pooling infers more delicate channel-level and spatial-level attention, combining average pooling and max pooling to retain more feature information. However, max-pooling only preserves important local feature cues. Thus, we apply the global standard deviation pooling operation to channel and spatial attention, which can take global and local feature cues into account, enhance feature representation capabilities, and strengthen the correlation between features of each correspondence. Specifically, to obtain more robust feature maps, we concatenate the 2D attention map F c R C × 1 × 1 and F s R 1 × N × 1 , which are output by CA module and SA module, respectively. Our experiments (see Section 4.4) conclude that the optimal connection method of the CA and SA module is sequential connection. Therefore, the CSA layer is summarized as follows:
F = F c F F = F s F ,
where ⊙ represents element-wise multiplication and F is the final refinement output of the CSA layer.
CA module: The channel attention module is shown in Figure 4. We first aggregate the spatial information of feature maps by using the standard deviation pooling and average pooling operation to generate two distinct spatial context descriptors: M std c and M avg c , which represent std-pooled features and average-pooled features, respectively. Then, the above two features are forwarded to the shared network to generate our channel attention map F c R C × 1 × 1 , where C is the number of channels. The shared network consists of a multilayer perceptron (MLP) with a hidden layer. The activation size of the hidden layer is set as R C / d × 1 × 1 , where d is the reduction ratio. After applying the shared network to both descriptors, we use element-wise summation to combine the output feature vectors. The channel attention is calculated as follows:
F c = ρ ( MLP ( StdPool ( F ) ) + MLP ( AvgPool ( F ) ) ) = ρ ( MLP ( M std c ) + MLP ( M avg c ) ) ,
where ρ represents the sigmoid function. MLP stands for the shared network.
SA module: As shown in Figure 5, we first generate two 2D maps M std s R 1 × N × 1 and M avg s R 1 × N × 1 by using two pooling operations to aggregate the channel information of one feature map. Each feature represents the std-pooled and average-pooled features of the entire channel. Then we concatenate the above two features together. After that, we apply a standard convolutional layer to generate a 2D spatial attention map. Finally, the sigmoid function is used to normalize the attention map. The spatial attention is calculated as follows:
F s = ρ ( O 1 × 1 ( [ StdPool ( F ) ; AvgPool ( F ) ] ) ) = ρ ( O 1 × 1 ( [ M std s ; M avg s ] ) ) ,
where O 1 × 1 denotes a convolution operation with a filter size of 1 × 1 and ρ represents the sigmoid function.
The nonlocal CSA layer is used to update features inspired by PointDSC [15], which mainly introduces a spatial consistency to complement the feature similarity of the nonlocal network [44]. To enhance the channel spatial features, we present the CSA layer to supplement the features of each correspondence after spatial topology and feature similarity computation. Therefore, the nonlocal CSA layer can not only obtain the spatial consistency and feature similarity information of each correspondence but also enhance the representation ability of the channel-spatial.

3.4. Seed Selection

Finding a dominant inlier cluster is difficult in the next SM layer, which cannot clearly distinguish both inlier and outlier. In this instance, utilizing the output of SM via weighted least-squares fitting [41] to transformation estimation may result in a suboptimal solution, since many outliers are still not explicitly eliminated. Therefore, we propose a new seed selection mechanism applied to apply spectral matching locally. In PointDSC [15], the seed selection mechanism is designed by two identical shared perceptrons and ReLu activation functions to compute the initial confidence for each correspondence. However, the random and straightforward structure ignores the global context encoding before calculating the confidence for each correspondence. That causes seed selection to find a suboptimal confidence distribution. Therefore, our new seed selection incorporates two contextual normalizations (shown in Figure 6) that encode each correspondence’s contextual feature maps by means and variances. The new seed selection mechanism can accurately calculate the context of each correspondence before calculating the confidence so that the confidence distribution is optimal. Therefore, it finds well-distributed and more reliable correspondences as seeds and looks for consistent correspondences around them in the feature space. After then, the new seed selection makes each subset have a higher inlier ratio than the set of input correspondences, which contributes SC 2 to construct a more reliable consensus set for each seed.
As shown in Figure 6, we use an MLP to select seeds by calculating the initial confidence for each correspondence, which uses the features learned by the CSC block. Then we utilize Non-Maximum Suppression [45] on these confidences to find well-distributed seeds. These selected seeds will be formed into multiple corresponding consensus sets by SC 2 measure and input into SM.

4. Experiments

In this section, we first present datasets and experimental setup in Section 4.1. Next, we evaluate our CSCE-Net on indoor scenes using different descriptors in Section 4.2. Further, we investigate the generalization ability of CSCE-Net on the outdoor scenes in Section 4.3. Finally, we introduce ablation studies in Section 4.4.

4.1. Datasets and Experimental Setup

Indoor scenes. We follow the same evaluation scheme as 3DMatch [46] to prepare the training and testing data in indoor scenes. For each pair of point clouds, we first down-sample the point clouds using a 5 cm voxel grid. The local feature descriptors were then extracted and matched to form the presumed correspondence. For all 3DMatch scans, we give two non-contiguous scene segments coarsely aligned using Random Sample Consensus (RANSAC). We then remove them from the dataset if RANSAC fails or if the overlap between the two segments is less than 30%. The procedure yielded 2186 pairs of partially overlapping point cloud segments in 8 scenes for training and 1623 for testing. Following [15], to prove the generalizability of CSCE-Net across descriptors, we employed FPFH [18] (handcrafted descriptors) and FCGF [10] (learned descriptors) as feature descriptors, respectively.
Outdoor scenes. We use the KITTI dataset [47] to test its effectiveness for outdoor scenes. KITTI contains a training set of 11 sequences, following [10,14,15] to divide it into train/val/test sets as follows: Sequences 0 to 5 are used for training, sequences 7 to 8 for validation, and sequences 8 to 10 for testing, obtaining 1358 pairs of partially overlapping point clouds for training and 555 pairs for testing. Then we downsample the above point cloud with 30 cm voxels and construct the initial input correspondence by extracting feature descriptors (i.e., FPFH and FCGF).
Evaluation criteria. We evaluate the performance of CSCE-Net on different datasets by the tasks of outlier rejection and scene pose estimation. Firstly, for the outpoint rejection task, we followed [14] to report three evaluation criteria: Inlier Precision ( IP ) , Inlier Recall ( IR ) , and F 1 score ( F 1 ) . Secondly, following [15], we also report three evaluation criteria for the scene pose estimation task: Rotation Error ( RE ) , Translation Error ( TE ) , and Registration Recall ( RR ) . The averages RE and TE are calculated only on successfully registered pairs. For RR, an alignment result is deemed successful if RE and TE are less than the given thresholds. The threshold is set to (15 , 30 cm) for indoor scenes and (5 , 60 cm) for outdoor scenes.
Implementation details. In all experiments, we constructed N × 6 initial correspondences by the nearest neighbor search of the feature descriptors FPFH and FCGF. Referring to PointDSC [15], we randomly selected 1000 correspondences for each pair of point clouds during the training process and constructed a batch input of size 16. The inlier threshold μ is set to 0.1 cm for indoor scenes and 0.6 cm for outdoor scenes. We set the hyper-parameter β = 3 and make ϵ f learned by the network. Using the ADAM optimizer, we optimized the network at an exponential decay factor of 0.99 and an initial learning rate of 0.0001 and trained the network with 50 epochs. All experiments are performed on PyTorch and RTX2080 graphics cards.

4.2. Evaluation of Indoor Scenes

We first report the results of the 3DMatch dataset in Table 1. We compare our method with 6 baselines: SM [17], RANSAC [19], GC RANSAC [23] and TEASER [27], DGR [14], and PointDSC [15]. For PointDSC, we use the pre-trained model provided by the method for testing. The results for other baseline methods are from [15].
Combined with FPFH. As shown in Table 1, compared to the learning-based baseline, CSCE-Net outperforms the second-best method PointDSC by more than 4% in F 1 , demonstrating the effectiveness of our outlier rejection method. In addition, the F 1 of DGR are as low as 17.35%. The reason is precise: each correspondence lacks the channel-spatial contextual features in feature embedding and the indispensable spatial consistency between correspondences in the 3D domain. Furthermore, as shown in Table 2, thanks to the excellent outlier rejection effect of CSCE-Net, the RR of the two-point clouds also achieves the best results (83.61%) when the two errors of RE and TE are minimized. This result is also 5.05% higher than the second-best PointDSC, demonstrating the competitiveness of CSCE-Net in point cloud registration. In addition, the traditional RANSAC method can achieve competitive results after a sufficient number of iterations. However, our method is about 40 times faster than RANSAC-100k while also achieving higher F 1 and RR . Figure 7 shows the outlier rejection visualization results of our CSCE-Net and the state-of-the-art PointDSC on 3DMatch.
Combined with FCGF. We further evaluate all methods for outlier rejection and pose estimation using the learned descriptor FCGF. As shown in Table 1 and Table 2, our CSCE-Net still outperforms the second-best PointDSC with a F 1 of 82.61% and RR of 93.47%. We also note that the IP of TEASER and the IR of GC-RANSAC-100k are higher than the respective values for CSCE-Net. At the same time, it is easy to find the IP and IR of CSCE-Net are also higher than those of GC-RANSAC-100k and TEASER, respectively. Therefore, we believe that CSCE-Net is able to balance IP and IR better while achieving higher F 1 . In addition, it is also not difficult to find that most of the methods combined with the more robust descriptor FCGF have better outlier rejection and pose estimation performance than FPFH.

4.3. Evaluation of Outdoor Scenes

To evaluate the generalization of CSCE-Net to unseen domains and new datasets, we re-test the outdoor lidar dataset by using a model trained on 3DMatch and a model retrained on KITTI, respectively. We choose SM [17], RANSAC [19], DGR [14], and PointDSC [15] as baseline methods.
We carefully report two training model results for CSCE-Net under two descriptors: one to test the pre-trained (no extra mark) model obtained on 3DMatch and the other to test the trained model from scratch (marked as “re-trained”) on KITTI. As shown in Table 3, first, when combined with the FPFH descriptor, the F 1 of CSCE-Net is approximately 7% higher than that of PointDSC, and the RR is approximately 6.5% higher in the training model of 3DMatch, which fully demonstrates its competitiveness and strong generalization ability to unseen datasets. At the same time, we note that the F 1 of DGR is only 4.51% because its outlier rejection and pose estimation generalize too weakly on outdoor datasets. When re-trained, the F 1 and RR of CSCE-Net is as high as 93.58% and 99.10%, respectively, outperforming all baseline methods. In addition, the F 1 and RR of the CSCE-Net are optimal when combined with the FCGF descriptors. These conclusions also show that our CSCE-Net is equally competitive for outdoor scenes.

4.4. Ablation Studies

In this section, we experimentally demonstrate the effectiveness of our CSCE-Net, our proposed CSC block structure design choice, and the advanced seed selection mechanism. We used the pre-trained model of 3DMatch combined with the FCGF descriptor to test the 3DMatch and KITTI datasets for these ablation studies. We first consider how to combine the channel and spatial attention modules in the CSA layer (Table 4). We then test the role of the channel and spatial attention and the position of the CSA layer in the CSC block (Table 5). Subsequently, to demonstrate the effectiveness of CSCE-Net, we progressively add the proposed and adopted modules to the baseline and report the results (Table 6). Finally, the parameter comparison between our CSCE-Net and the recent PointDSC (Table 7). The test results represent the F 1 and RR obtained by the pre-trained model in 3DMatch combined with the FPFH descriptor in 3DMatch and KITTI.
Arrangement of the channel and spatial attention. In this experiment, we verify the optimal arrangement of the channel and spatial attention of the CSA layer in the CSC block in three orders: channel-spatial, spatial-channel, and channel-spatial parallel. Since the channel and spatial modules have different functions, the order in which they are arranged may affect the entirety of the performance. In short, channel attention is applied globally, while spatial attention works locally. Therefore, the two types of attention can be applied complementary to construct a 3D attention map. From the different arrangements in Table 4, we find that the attention maps generated by channel-spatial attention in sequence are the best. Therefore, the structure of channel-spatial attention in the CSA layer in the ablation experiment below is a sequential arrangement of channel and spatial.
CSC block design. As shown in Table 5, we test the performance of different combinations of outlier rejection and positional estimation tasks within the CSC block on the 3DMatch and KITTI datasets. The first row of the table is PointDSC, which is the most recent state-of-the-art learning-based outlier rejection method. Therefore, we use PointDSC as a baseline for comparison. We can see that all of our iterative combinations perform better than the baseline. Specifically, the channel attention module (CSA1 + CSA2 + CA) achieved RR improvements of 1.69% and 2.07% over baseline in the 3DMatch and KITTI scenes, respectively. The spatial attention module (CSA1 + CSA2 + SA) improved the RR in the 3DMatch and KITTI scenes by 1.46% and 1.61% over baseline, respectively. These two terms indicate that the channel attention module and the spatial attention module are able to capture rich feature map contextual information and, thus, improve the performance of the pose estimation task for the point cloud. For using only CSA layer1 (CSA1 + CA + SA), we obtained a significant improvement over baseline performance, with RR improvements of 2.2% and 2.54% over baseline in the 3DMatch and KITTI scenes, respectively. This demonstrates that this complementary channel-spatial attention we propose outperforms individual channel and spatial attention. For using only CSA layer2 (CSA2 + CA + SA), we also improved 1.32% and 1.31% in RR over baseline on 3DMatch and KITTI. These two terms show that adding CSA layers at different positions of the CSC block is effective. We ended up with a 3.69%/3.01% and 4.0%/3.63% improvement in F1/RR on the 3DMatch and KITTI datasets, respectively, compared to the baseline after inserting CSA layers at different locations in the CSC block.
Final network design. To demonstrate the effect of our CSC block and the advanced improved seed selection mechanism, we conduct ablation experiments to add the CSC block and the enhanced new seed selection mechanism for the baseline. We first add the CSC block as the feature embedding part. By extracting abundant contextual features for each correspondence, the separation of inliers and outliers is facilitated, thus, improving the accuracy of the point cloud registration. As shown in Table 6, the registration recall rate guided by the CSC block was 3.01% higher than PointDSC when combined with FPFH and 3.63% higher when combined with FCGF. We further adopted a new seed selection mechanism for selecting more reliable correspondences as seeds. When a seed acts as an inlier, the consensus set of seeds formed later is more likely to form inlier clusters, thus, improving the accuracy of transformations between point cloud pairs. Finally, to calculate the exact seed consensus set, we inserted the SC 2 measure to optimize the network.
Computational cost. As shown in Table 7, we have considered the network parameters while improving the network performance. The parameters required by our CSCE-Net network (0.56M) are almost half of those required by the current state-of-the-art outlier rejection network PointDSC (1.05M). If we reduce the feature embedding block of PointDSC to use only six layers (i.e., PointDSC(6)), then PointDSC(6) corresponding network parameters are directly reduced to half (0.53M), but then the network performance will also be significantly reduced. Thus, our network achieves a balance between performance and the number of parameters required.

5. Conclusions

In this paper, we propose an efficient CSC block and an advanced Seed Selection mechanism as the backbone of the newly proposed CSCE-Net to find more inliers in the initial correspondence and recover 3D rigid transformation between two point clouds. In particular, our proposed CSC block can collect abundant contextual information for each pair of correspondences and enhance the feature representation through some novel channel-spatial attention operations. The advanced Seed Selection utilizes the features obtained from the CSC block to improve the ratio of inliers. Results of outlier removal and pose estimation tasks in challenging indoor and outdoor scenes to show that our CSCE-Net achieves impressive performance improvements compared to state-of-the-art competing networks by combining different 3D local descriptors. In addition, the network parameters of our proposed CSCE-Net are lower than those of the state-of-the-art outlier rejection network PointDSC.

Author Contributions

Conceptualization, J.W.; Methodology, J.W. and C.Y.; Software, J.W.; Writing—original draft, J.W.; Writing—review & editing, C.Y., L.W. and R.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62171130 and 61802064, in part by the Fujian Province Health Education Joint Research Project under Grant 2019WJ28, in part by the Natural Science Fund of Fujian Province under Grant 2019J01402.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Deschaud, J.-E. Imls-slam: Scan-to-model matching based on 3d data. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation, Brisbane, Australia, 21–25 May 2018; pp. 2480–2485. [Google Scholar]
  2. Wong, J.M.; Kee, V.; Le, T.; Wagner, S.; Mariottini, G.-L.; Schneider, A.; Hamilton, L.; Chipalkatty, R.; Hebert, M.; Johnson, D.M.; et al. Segicp: Integrated deep semantic segmentation and pose estimation. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Vancouver, BC, Canada, 24–28 September 2017; pp. 5784–5789. [Google Scholar]
  3. Bailey, T.; Durrant-Whyte, H. Simultaneous localization and mapping (slam): Part ii. IEEE Robot. Autom. Mag. 2006, 13, 108–117. [Google Scholar] [CrossRef] [Green Version]
  4. Corso, N.; Zakhor, A. A Indoor localization algorithms for an ambulatory human operated 3D mobile mapping system. Remote Sens. 2013, 5, 6611–6646. [Google Scholar] [CrossRef] [Green Version]
  5. Fan, A.; Ma, J.; Jiang, X.; Ling, H. Efficient Deterministic Search with Robust Loss Functions for Geometric Model Fitting. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 8212–8229. [Google Scholar] [CrossRef] [PubMed]
  6. Xu, H.; Ma, J.; Yuan, J.; Le, Z.; Liu, W. RFNet: Unsupervised Network for Mutually Reinforcing Multi-modal Image Registration and Fusion. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA, 19–24 June 2022; pp. 19679–19688. [Google Scholar]
  7. Yew, Z.J.; Lee, G.H. 3dfeat-net: Weakly supervised local 3d features for point cloud registration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 607–623. [Google Scholar]
  8. Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.-L. D3feat: Joint learning of dense detection and description of 3d local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6359–6367. [Google Scholar]
  9. Huang, S.; Gojcic, Z.; Usvyatsov, M.; Wieser, A.; Schindler, K. Predator: Registration of 3d point clouds with low overlap. In Proceedings of the IEEE/CVF International Conference on Computer Vision and Patter Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 4267–4276. [Google Scholar]
  10. Choy, C.; Park, J.; Koltun, V. Fully convolutional geometric features. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 8958–8966. [Google Scholar]
  11. Ao, S.; Hu, Q.; Yang, B.; Markham, A.; Guo, Y. Spinnet: Learning a general surface descriptor for 3d point cloud registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision and Patter Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 11753–11762. [Google Scholar]
  12. Arnold, E.; Mozaffari, S.; Dianati, M. Fast and robust registration of partially overlapping point clouds. IEEE Robot. Autom. Lett. 2021, 7, 1502–1509. [Google Scholar] [CrossRef]
  13. Pais, G.D.; Ramalingam, S.; Govindu, V.M.; Nascimento, J.C.; Chellappa, R.; Miraldo, P. 3dregnet: A deep neural network for 3d point registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7193–7203. [Google Scholar]
  14. Choy, C.; Dong, W.; Koltun, V. Deep global registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2514–2523. [Google Scholar]
  15. Bai, X.; Luo, Z.; Zhou, L.; Chen, H.; Li, L.; Hu, Z.; Fu, H.; Tai, C.-L. Pointdsc: Robust point cloud registration using deep spatial consistency. In Proceedings of the IEEE/CVF International Conference on Computer Vision and Patter Recognition, Montreal, QC, Canada, 10–17 October 2021; pp. 15859–15869. [Google Scholar]
  16. Chen, Z.; Sun, K.; Yang, F.; Tao, W. Sc2-pcr: A second order spatial compatibility for efficient and robust point cloud registration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA, 19–24 June 2022; pp. 13221–13231. [Google Scholar]
  17. Leordeanu, M.; Hebert, M. A spectral technique for correspondence problems using pairwise constraints. In Proceedings of the IEEE International Conference on Computer Vision, San Diego, CA, USA, 20–26 June 2005; pp. 1482–1489. [Google Scholar]
  18. Rusu, R.B.; Blodow, N.; Beetz, M. Fast point feature histograms (fpfh) for 3d registration. In Proceedings of the 2009 IEEE International Conference on Robotics and Automation, IEEE, Kobe, Japan, 12–17 May 2009; pp. 3212–3217. [Google Scholar]
  19. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  20. Chum, O.; Matas, J.; Kittler, J. Locally optimized ransac. In Proceedings of the Joint Pattern Recognition Symposium, Madison, WI, USA, 17 June 2003; pp. 236–243. [Google Scholar]
  21. Le, H.M.; Do, T.-T.; Hoang, T.; Cheung, N.-M. Sdrsac: Semidefinite-based randomized approach for robust point cloud registration without correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 124–133. [Google Scholar]
  22. Li, J.; Hu, Q.; Ai, M. Gesac: Robust graph enhanced sample consensus for point cloud registration. ISPRS J. Photogramm. Remote Sens. 2020, 167, 363–374. [Google Scholar] [CrossRef]
  23. Barath, D.; Matas, J. Graph-cut ransac. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 6733–6741. [Google Scholar]
  24. Barath, D.; Matas, J.; Noskova, J. MAGSAC: Marginalizing sample consensus. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 10197–10205. [Google Scholar]
  25. Yang, J.; Xian, K.; Wang, P.; Zhang, Y. A performance evaluation of correspondence grouping methods for 3d rigid data matching. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1859–1874. [Google Scholar] [CrossRef] [PubMed]
  26. Zhou, Q.-Y.; Park, J.; Koltun, V. Fast global registration. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 766–782. [Google Scholar]
  27. Yang, H.; Shi, J.; Carlone, L. Teaser: Fast and certifiable point cloud registration. IEEE Trans. Robot. 2020, 37, 314–333. [Google Scholar] [CrossRef]
  28. Lim, H.; Yeon, S.; Ryu, S.; Lee, Y.; Kim, Y.; Yun, J.; Jung, E.; Lee, D.; Myung, H. A single correspondence is enough: Robust global registration to avoid degeneracy in urban environments. In Proceedings of the International Conference on Robotics and Automation, Philadelphia, PA, USA, 23–27 May 2022; pp. 8010–8017. [Google Scholar]
  29. Liu, Y.; Li, Y.; Dai, L.; Yang, C.; Wei, L.; Lai, T.; Chen, R. Robust feature matching via advanced neighborhood topology consensus. Neurocomputing 2021, 421, 273–284. [Google Scholar] [CrossRef]
  30. Ma, J.; Fan, A.; Jiang, X.; Xiao, G. Feature Matching via Motion-Consistency Driven Probabilistic Graphical Model. Int. J. Comput. Vis. 2020, 130, 2249–2264. [Google Scholar] [CrossRef]
  31. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image Matching from Handcrafted to Deep Features: A Survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  32. Ma, J.; Zhao, J.; Jiang, J.; Zhou, H.; Guo, X. Locality preserving matching. Int. J. Comput. Vis. 2019, 512–531. [Google Scholar] [CrossRef]
  33. Yi, K.M.; Trulls, E.; Ono, Y.; Lepetit, V.; Salzmann, M.; Fua, P. Learning to find good correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 2666–2674. [Google Scholar]
  34. Zhao, C.; Cao, Z.; Li, C.; Li, X.; Yang, J. Nm-net: Mining reliable neighbors for robust feature correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 215–224. [Google Scholar]
  35. Zhang, J.; Sun, D.; Luo, Z.; Yao, A.; Zhou, L.; Shen, T.; Chen, Y.; Quan, L.; Liao, H. Learning two-view correspondences and geometry using order-aware network. In Proceedings of the International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 5844–5853. [Google Scholar]
  36. Dai, L.; Liu, Y.; Ma, J.; Wei, L.; Lai, T.; Yang, C.; Chen, R. MS2DG-Net: Progressive Correspondence Learning via Multiple Sparse Semantics Dynamic Graph. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LO, USA, 19–24 June 2022; pp. 8973–8982. [Google Scholar]
  37. Liu, Y.; Zhao, B.N.; Zhao, S.; Zhang, L. Progressive Motion Coherence for Remote Sensing Image Matching. IEEE Trans. Geosci. Remote Sens. 2022, 5631113. [Google Scholar] [CrossRef]
  38. Liu, X.; Xiao, G.; Li, Z.; Chen, R. Point2cn: Progressive two-view correspondence learning via information fusion. Signal Process. 2021, 189, 108304. [Google Scholar] [CrossRef]
  39. Chen, Z.; Yang, F.; Tao, W. Detarnet: Decoupling translation and rotation by siamese network for point cloud registration. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; pp. 401–409. [Google Scholar]
  40. Lee, J.; Kim, S.; Cho, M.; Park, J. Deep hough voting for robust global registration. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 15994–16003. [Google Scholar]
  41. Besl, P.J.; McKay, N.D. Method for registration of 3-d shapes. In Sensor Fusion IV: Control Paradigms and Data Structures; SPIE: Bellingham, WA, USA, 1992; Volume 1611, pp. 586–606. [Google Scholar]
  42. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  43. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 3–19. [Google Scholar]
  44. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UH, USA, 18–22 June 2018; pp. 7794–7803. [Google Scholar]
  45. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  46. Zeng, A.; Song, S.; Nießner, M.; Fisher, M.; Xiao, J.; Funkhouser, T. 3dmatch: Learning local geometric descriptors from rgb-d reconstructions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1802–1811. [Google Scholar]
  47. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
Figure 1. Visualization results of outlier rejection (top row) and pose estimation (bottom row) for pairwise point clouds in indoor scenes. IP/IR and RE/TE indicate inlier precision/inlier recall and rotation error/translation error, respectively. ( a ) The popular network PointDSC [15] with weak contextual features retains many outliers and obtains significant alignment errors in pairwise point clouds with slight overlaps resulting in registration failures. ( b ) CSCE-Net extracts richer contextual features and retains more inliers in pairwise point clouds that overlap less, significantly reducing alignment errors and achieving accurate point cloud alignment.
Figure 1. Visualization results of outlier rejection (top row) and pose estimation (bottom row) for pairwise point clouds in indoor scenes. IP/IR and RE/TE indicate inlier precision/inlier recall and rotation error/translation error, respectively. ( a ) The popular network PointDSC [15] with weak contextual features retains many outliers and obtains significant alignment errors in pairwise point clouds with slight overlaps resulting in registration failures. ( b ) CSCE-Net extracts richer contextual features and retains more inliers in pairwise point clouds that overlap less, significantly reducing alignment errors and achieving accurate point cloud alignment.
Remotesensing 14 05751 g001
Figure 2. The framework diagram of our CSCE-Net. It contains several CSC block blocks, a Seed Selection mechanism, a Second Order Spatial Compatibility Measure, and Spectral Matching.
Figure 2. The framework diagram of our CSCE-Net. It contains several CSC block blocks, a Seed Selection mechanism, a Second Order Spatial Compatibility Measure, and Spectral Matching.
Remotesensing 14 05751 g002
Figure 3. Channel-Spatial Attention (CSA) layer.
Figure 3. Channel-Spatial Attention (CSA) layer.
Remotesensing 14 05751 g003
Figure 4. Channel Attention (CA) module.
Figure 4. Channel Attention (CA) module.
Remotesensing 14 05751 g004
Figure 5. Spatial Attention (SA) module.
Figure 5. Spatial Attention (SA) module.
Remotesensing 14 05751 g005
Figure 6. Seed Selection. It mainly consists of two similar sequences Context Norm layer, Shared Perceptron layer, and ReLU layer. Finally, it goes through a shared perceptron to reduce the dimensionality output.
Figure 6. Seed Selection. It mainly consists of two similar sequences Context Norm layer, Shared Perceptron layer, and ReLU layer. Finally, it goes through a shared perceptron to reduce the dimensionality output.
Remotesensing 14 05751 g006
Figure 7. Visualization results for PointDSC (1, 3 rows) and CSCE-Net (2, 4 rows). These six scenes are the result of the 3DMatch dataset. The left side of each scene represents PointDSC and the right side represents CSCE-Net. The green lines indicate the identified correct matches (inliers) of the network, while the red lines indicate the preserved false matches (outliers) of the network. Inlier Precision ( I P ) and Inlier Recall ( I R ) are reported at the lower-left corner of each pair of point clouds. Best viewed with magnification and color.
Figure 7. Visualization results for PointDSC (1, 3 rows) and CSCE-Net (2, 4 rows). These six scenes are the result of the 3DMatch dataset. The left side of each scene represents PointDSC and the right side represents CSCE-Net. The green lines indicate the identified correct matches (inliers) of the network, while the red lines indicate the preserved false matches (outliers) of the network. Inlier Precision ( I P ) and Inlier Recall ( I R ) are reported at the lower-left corner of each pair of point clouds. Best viewed with magnification and color.
Remotesensing 14 05751 g007
Table 1. Quantitative results of outlier rejection on the 3DMatch datasets.
Table 1. Quantitative results of outlier rejection on the 3DMatch datasets.
Method3DMatch
IP (%↑)IR (%↑)F1 (%↑)Time (s)
FPFH [18]SM [17]47.9670.6950.700.03
TEASER [27]73.0162.6366.930.03
GC-RANSAC-100 k [23]48.5569.3856.780.62
RANSAC-100k68.1867.4067.475.24
DGR [14]28.8012.4217.352.49
PointDSC [15]68.6371.6369.890.08
CSCE-Net73.1177.4774.750.13
FCGF [10]SM [17]81.4438.3648.210.03
TEASER [27]82.4368.0873.960.11
GC-RANSAC-100 k [23]64.4693.3975.690.47
RANSAC-100 k78.3885.3081.435.50
DGR [14]67.4778.9472.761.36
PointDSC [15]79.0786.4882.310.08
CSCE-Net79.1687.0282.610.13
Table 2. Quantitative results of pose estimation on the 3DMatch datasets.
Table 2. Quantitative results of pose estimation on the 3DMatch datasets.
Method3DMatch
RE (°↓)TE (cm↓)RR (%↑)Time (s)
FPFH [18]SM [17]2.948.1555.880.03
TEASER [27]2.487.3175.480.03
GC-RANSAC-100 k [23]2.336.8767.650.62
RANSAC-100k3.5510.0473.575.24
DGR [14]3.7810.8069.132.49
PointDSC [15]2.096.5978.560.08
CSCE-Net2.116.7383.610.13
FCGF [10]SM [17]2.297.0786.570.03
TEASER [27]2.738.6685.770.11
GC-RANSAC-100 k [23]2.337.1192.050.47
RANSAC-100 k2.497.5491.505.50
DGR [14]2.407.4891.301.36
PointDSC [15]2.056.5493.220.08
CSCE-Net2.066.5393.470.13
Table 3. Quantitative results of outlier rejection and pose estimation on the KITTI datasets.
Table 3. Quantitative results of outlier rejection and pose estimation on the KITTI datasets.
MethodKITTI
F1 (%↑)RETE (cm↓)RR (%↑)Time (s)
FPFH [18]SM [17]56.370.4712.1579.640.18
RANSAC-100k73.131.2225.8889.3713.7
DGR [14]4.511.6734.7473.690.86
PointDSC [15]85.590.357.1792.250.23
CSCE-Net92.840.337.0598.740.61
PointDSC re-trained88.510.357.1698.200.23
CSCE-Net re-trained93.580.327.0399.100.61
FCGF [10]SM [17]22.840.5019.7396.760.10
RANSAC-100 k85.420.3822.6098.3813.4
DGR [14]73.600.4323.2895.140.86
PointDSC [15]85.290.3320.9997.480.31
CSCE-Net85.810.3220.8997.840.61
PointDSC re-trained85.370.3320.9498.200.31
CSCE-Net re-trained86.060.3120.8998.660.61
Table 4. Combining methods of the channel and spatial attention. The `+’ indicates a serial connection here and the `&’ indicates a parallel connection. The test results are combined with the FPFH descriptor.
Table 4. Combining methods of the channel and spatial attention. The `+’ indicates a serial connection here and the `&’ indicates a parallel connection. The test results are combined with the FPFH descriptor.
DescriptionFPFH
F1 (%↑)RR (%↑)
3DMatchchannel & spatial in parallel72.8281.78
spatial + channel73.7682.65
channel + spatial74.7583.61
KITTIchannel & spatial in parallel90.4196.48
spatial + channel91.9497.84
channel + spatial92.8498.74
Table 5. Ablation study of CSC blocks on 3DMatch and KITTI datasets. Baseline: PointDSC [15]. CSA1: using the CSA layer. CSA2: using the CSA layer in the Nonlocal CSA layer. CA: using the channel attention module in CSA layers. SA: using the spatial attention module in CSA layers. The test results are combined with the FPFH descriptor.
Table 5. Ablation study of CSC blocks on 3DMatch and KITTI datasets. Baseline: PointDSC [15]. CSA1: using the CSA layer. CSA2: using the CSA layer in the Nonlocal CSA layer. CA: using the channel attention module in CSA layers. SA: using the spatial attention module in CSA layers. The test results are combined with the FPFH descriptor.
BaselineCSA1CSA2CASAFPFH
F1 (%↑)RR (%↑)
3DMatch 69.8978.56
72.1380.25
71.9180.02
72.8280.76
71.8679.88
73.5881.57
KITTI 85.5992.25
88.2294.32
87.9593.86
88.6994.79
87.6093.56
89.5995.88
Table 6. Ablation study of the CSCE-Net blocks on 3DMatch and KITTI datasets. Baseline: PointDSC [15]. CSC: Channel-Spatial Contextual block. Seed: The advanced Seed Selection mechanism. SC 2 : Second Order Spatial Compatibility (SC 2 ) measure. The test results are combined with the FPFH descriptor.
Table 6. Ablation study of the CSCE-Net blocks on 3DMatch and KITTI datasets. Baseline: PointDSC [15]. CSC: Channel-Spatial Contextual block. Seed: The advanced Seed Selection mechanism. SC 2 : Second Order Spatial Compatibility (SC 2 ) measure. The test results are combined with the FPFH descriptor.
BaselineCSCSeedSC 2 FPFH
F1 (%↑)RR (%↑)
3DMatch 69.8978.56
73.5881.57
74.0282.12
74.7583.61
KITTI 85.5992.25
89.5995.88
91.9897.86
92.8498.74
Table 7. Compare the network parameters (Param) of PointDSC and our CSCE-Net. PointDSC (6) indicates that the PointDSC feature embedding is reduced from 12 layers to 6 layers. The test results are combined with the 3DMatch dataset.
Table 7. Compare the network parameters (Param) of PointDSC and our CSCE-Net. PointDSC (6) indicates that the PointDSC feature embedding is reduced from 12 layers to 6 layers. The test results are combined with the 3DMatch dataset.
MethodFPFHFCGFParam (M)
F1/RRF1/RR
PointDSC [15]69.89/78.5682.31/93.221.05
PointDSC (6)67.25/76.9680.09/92.170.53
CSCE-Net74.75/83.6182.61/93.470.56
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Yang, C.; Wei, L.; Chen, R. CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration. Remote Sens. 2022, 14, 5751. https://doi.org/10.3390/rs14225751

AMA Style

Wang J, Yang C, Wei L, Chen R. CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration. Remote Sensing. 2022; 14(22):5751. https://doi.org/10.3390/rs14225751

Chicago/Turabian Style

Wang, Jingtao, Changcai Yang, Lifang Wei, and Riqing Chen. 2022. "CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration" Remote Sensing 14, no. 22: 5751. https://doi.org/10.3390/rs14225751

APA Style

Wang, J., Yang, C., Wei, L., & Chen, R. (2022). CSCE-Net: Channel-Spatial Contextual Enhancement Network for Robust Point Cloud Registration. Remote Sensing, 14(22), 5751. https://doi.org/10.3390/rs14225751

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop