Next Article in Journal
An Improved Fading Factor-Based Adaptive Robust Filtering Algorithm for SINS/GNSS Integration with Dynamic Disturbance Suppression
Previous Article in Journal
Analyzing the Sources and Variations of Nighttime Lights in Hong Kong from VIIRS Monthly Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Multi-Modal Point Cloud Completion Algorithm Guided by Image Rotation Attention

1
College of Electronic Science and Technology, National University of Defense Technology, Changsha 410005, China
2
College of Meteorology and Oceanography, National University of Defense Technology, Changsha 410005, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(8), 1448; https://doi.org/10.3390/rs17081448
Submission received: 13 February 2025 / Revised: 6 April 2025 / Accepted: 16 April 2025 / Published: 18 April 2025
(This article belongs to the Topic Computer Vision and Image Processing, 2nd Edition)

Abstract

:
This paper proposes a novel point cloud multi-scale completion algorithm guided by image rotation attention mechanisms to address the challenges of severe information loss and suboptimal fusion effects in multi-modal feature extraction and integration during point cloud shape completion. The proposed network employs an encoder–decoder structure, integrating a Rotating Channel Attention (RCA) module for enhanced image feature extraction and a multi-scale feature extraction method for point clouds to improve both local and global feature information. The network also utilizes multi-level self-attention mechanisms to achieve effective multi-modal feature fusion. The decoder employs a multi-branch completion method, guided by Chamfer distance loss, to accomplish the point cloud completion task. Extensive experiments on the ShapeNet-ViPC and ModelNet40ViPC datasets demonstrate the effectiveness of the proposed algorithm. Compared to eight related algorithms, the proposed method achieves superior performance in terms of completion accuracy and efficiency. Specifically, compared to the state-of-the-art XMFnet, the average Chamfer distance (CD) value is reduced by 11.71%. The algorithm also shows significant improvements in visual comparisons, with more distinct structural details and a more uniform density distribution in the completed point clouds. The ablation studies further validate the effectiveness of the RCA module and the multi-scale module, highlighting their complementary nature in enhancing point cloud completion accuracy. Future work will focus on improving the network’s performance and exploring its application in more complex 3D vision tasks.

1. Introduction

In recent years, the pursuit of accurate and efficient methods for 3D scene understanding has gained significant attention in various fields, including autonomous vehicles, robotics, and augmented reality. A key technology in 3D vision, Light Detection and Ranging (LiDAR), provides precise measurements and preserves the original geometric information of objects in 3D space [1,2]. LiDAR point clouds, which are collections of 3D points representing the surfaces of objects, have been widely used in numerous applications. For instance, in autonomous vehicles, LiDAR point clouds are used for obstacle detection and path planning. In robotics, they are employed for environment mapping and object manipulation. In augmented reality, point clouds help in the integration of virtual elements with the real world. Additionally, in industrial inspection and cultural heritage preservation, point clouds are used for detailed object scanning and analysis.
Despite the numerous applications, the challenge of working with incomplete LiDAR point cloud data has significantly impeded the progress in 3D scene understanding. Point cloud data are often sparse or incomplete due to various factors such as object occlusion, surface reflection, and sensor limitations. For example, in outdoor environments, objects may be partially obscured by other objects, leading to missing data in the point cloud. In indoor settings, reflective surfaces can cause LiDAR sensors to miss certain areas. Moreover, the limited scanning range and resolution of LiDAR sensors can result in sparse point cloud data, especially in large-scale scenes. These issues undermine the performance of subsequent 3D vision tasks, including object classification, detection, and segmentation, which are fundamental to many applications.
To address these challenges, point cloud completion has become a crucial task in 3D vision. The goal of point cloud completion is to reconstruct the missing parts of a point cloud, thereby providing a complete and accurate representation of the object or scene. Traditional point cloud completion methods often rely on geometric assumptions or manual feature engineering, which are limited in their ability to capture complex structures and details. For example, some methods use primitive shapes such as spheres and cylinders to fill in missing regions, but these approaches may fail to capture the intricate details of real-world objects. With the advancement of deep learning, researchers have developed various deep learning-based point cloud completion methods. These methods typically use autoencoder structures to learn the underlying features of point clouds and generate complete point clouds. For instance, PointNet [3] and PointNet++ [4] are pioneering works that use multilayer perceptrons (MLPs) and hierarchical architectures to process point clouds. However, these methods often struggle to balance local and global feature information, leading to suboptimal completion results. Moreover, they primarily focus on single-modal point cloud data and do not leverage the complementary information from other modalities, such as images. In this context, multi-modal point cloud completion methods have emerged as a promising direction. These methods integrate data from different sources, such as LiDAR and images, to enhance the completion process. Images provide rich texture and structural information that can guide the completion of sparse or incomplete point clouds. For example, the ViPC network [5] and CSDN network [6] are early attempts to combine point cloud and image data for completion tasks. However, these methods still face challenges in effectively fusing multi-modal information and maintaining the structural integrity of the completed point clouds.
The motivation behind our research is to develop a more effective point cloud completion algorithm that can better utilize the complementary strengths of point clouds and images. Our proposed method focuses on enhancing feature extraction from both modalities and improving the fusion of multi-modal features. Specifically, we introduce a LiDAR point cloud multi-scale completion algorithm guided by image rotation attention mechanisms. This algorithm employs an encoder–decoder structure, where the image feature extractor utilizes rotation attention mechanisms to enhance image feature extraction, and the point cloud feature extractor employs multi-scale methods to improve the global and local information of point clouds. The network also uses multi-level self-attention mechanisms to achieve effective multi-modal feature fusion.
Our research aims to address the limitations of existing point cloud completion methods by proposing a novel network architecture that leverages the strengths of both point cloud and image data. Through extensive experiments, we demonstrate the effectiveness of our approach, showing superior performance compared to state-of-the-art methods in terms of completion accuracy and efficiency. We believe that our work provides a significant advancement in the field of 3D vision and offers new possibilities for applications requiring accurate and reliable point cloud data. The code for our method is publicly available at https://github.com/CaineGu/CMFN, accessed on 2 February 2025.

2. Related Works

In the realm of three-dimensional (3D) vision technology, the completion of point cloud data has been approached through various methodologies, which can be broadly categorized into single-modal and multi-modal point cloud completion methods. Single-modal methods utilize only point cloud data for shape completion, while multi-modal methods integrate data from different sources, such as LiDAR and images, to enhance the completion process.
Single-modal point cloud completion methods utilize only object point cloud data for shape completion. Qi Charles et al.’s PointNet [3], as a multilayer perceptron (MLP) method for points, directly inspired the network design of many subsequent scholars. PointNet++ [4] and TopNet [7] incorporate hierarchical architectures to consider the local structural features of point clouds. To mitigate the structural loss brought by MLP, AtlasNet [8] and MSN [9] reconstruct complete outputs by evaluating a set of surface parameters, from which a complete point cloud can be generated. The encoder–decoder architecture pioneered by PCN [10] requires no structural assumptions or annotation information about the underlying shape. PoinTr [11] adapts the Transformer module for 3D completion tasks by creating a geometry-aware module to model local geometric relationships. PointAttN [12] enhances point cloud completion by employing attention mechanisms to capture detailed point relationships and improve feature extraction. Additionally, SnowflakeNet [13] generates child points by learning the transition of specific regions to gradually split parent points, enabling the network to predict highly detailed shape geometries. Yang et al. [14] proposed FoldingNet, which uses a grid deformation strategy to reconstruct point clouds, demonstrating the potential of auto-encoder architectures for point cloud processing. Pan [15] introduced ECG, which employs graph convolution to achieve edge-aware point cloud completion, focusing on preserving sharp edges and fine details. Pan et al. [16] further developed a variational relational point completion network, which enhances the robustness of 3D classification by incorporating relational reasoning into the completion process.
In single-modal point cloud completion works, the limited available geometric information results in high uncertainty in inferring missing regions of point clouds. Furthermore, the low scanning resolution of 3D data typically leads to sparsely captured data, making it difficult to determine whether the missing regions in the point cloud are due to incompleteness or inherent sparsity. Therefore, leveraging the complementary information between multi-modalities to enhance point cloud completion effects is essential. The cross-modal multi-scale point cloud feature fusion shape completion network (CMFN) utilizes different data information fusion methods to achieve the enhancement of point cloud data information. The unordered and unstructured nature of point clouds prevents them from directly extracting features using structural characteristics like images, so using image information to represent point cloud information during the point cloud completion process can improve the completion accuracy. In image-guided point cloud shape completion methods, extracting dual-modality information and effectively fusing it is key to point cloud completion work, which this paper focuses on in terms of feature extraction for both modalities.
Multi-modal point cloud completion methods address these limitations by leveraging the complementary information from different modalities. Zhang [5] proposed the ViPC network, which inputs incomplete point clouds and object images into the network, generating a coarse point cloud based on the object image information before fusing it with the incomplete point cloud. Building on this, Zhu [6] proposed the CSDN network, which differs from the ViPC network in that it uses an encoder–decoder structure to first recover the overall but coarse shape of the point cloud before refining local details. Emanuele [17] proposed XMFnet, which outperforms its predecessors in point cloud completion performance, characterized by the introduction of cross-attention mechanisms and self-attention mechanisms to supervise the structural generation of point clouds. Although the aforementioned methods have achieved better completion results using multi-modal information in point cloud completion tasks, they also have some issues in implementation: the ViPC network consumes a significant amount of memory and computational resources to generate coarse point clouds from image information, with suboptimal results; the CSDN network has excessive computational demands, and the underutilization of multi-modal information leads to imprecise completion details; and the XMFnet network’s insufficient feature extraction capabilities for point cloud and image data result in non-refined local structures in the decoded point cloud. Multi-modal point cloud completion methods, despite their potential, face several challenges as documented in recent studies. The integration of data from various sensors, such as LiDAR and cameras, requires effective fusion techniques to leverage the complementary information. However, the heterogeneity between modalities and the difficulty in aligning features from different sensors have been noted as significant obstacles. Additionally, the computational complexity and the need for high memory capacity to process and fuse multi-modal data have been highlighted as limitations. The utilization of multi-modal information, leading to imprecise completion details, has also been reported. Furthermore, the insufficient feature extraction capabilities of some networks result in non-refined local structures in the decoded point clouds. These issues underscore the need for advanced algorithms that can effectively take advantage of multi-modal data to enhance point cloud completion tasks.
Despite these advancements, multi-modal methods face challenges such as the heterogeneity between modalities, computational complexity, and the need for high memory capacity [18]. Additionally, the underutilization of multi-modal information can lead to imprecise completion details [19]. To address these shortcomings, our proposed algorithm employs a LiDAR point cloud multi-scale completion algorithm guided by image rotation attention mechanisms, focusing on enhancing feature extraction from both point clouds and images.

3. Methodology

Our proposed network, designed to address the challenges of balancing local and global feature information in point clouds and mitigating the loss of structural information in images, leverages an encoder–decoder architecture to facilitate the completion of LiDAR point clouds with the guidance of image data. The network is divided into four main components: the point cloud multi-scale feature extractor, the image feature extractor, the cross-modal feature fusion (CMF) module, and the decoder. To address the shortcomings of current multi-modal point cloud completion algorithms, this paper proposes a LiDAR point cloud multi-scale completion algorithm based on the image rotation attention mechanism. The specific contributions are (1) the introduction of the Rotating Channel Attention (RCA) module, which enhances the extraction capability of image features by adding a Rotating Channel Attention mechanism to existing image feature extractors and (2) the optimization of the point cloud feature extractor, which employs a multi-scale method for hierarchical feature extraction, both locally and globally, to enhance the capability of point cloud feature extraction. Each component plays a critical role in the pipeline, from feature extraction to the final point cloud completion task.

3.1. Overall Framework of Network

Figure 1 shows the overall framework of the cross-modal multi-scale feature fusion network. The input to the point cloud feature extractor is two point clouds of different scales ( P in 1 and P in 2 ), which are then fed into their respective feature extractors to obtain multi-scale feature vectors ( F p 1 and F p 2 ). The image feature extractor extracts features from image data, resulting in the feature vector F I . The following two subsections of this paper will focus on the image feature extractor and the point cloud multi-scale feature extractor.
The cross-modal feature fusion (CMF) utilizes cross-attention and self-attention mechanisms [20] to integrate feature information from both modalities. As shown in Figure 1, the image features F I are fused with the point cloud features F p 1 to obtain the fused features F p 2 . The features F 2 are then concatenated with F 1 in parallel and fed into a multilayer perceptron (MLP) layer to obtain the final fused features F f . This paper employs multi-head attention mechanisms, first distinguishing between the two different modality features. The paper uses F p 1 and F p 2 as the query matrix Q and the key matrix K, respectively, with V representing the value matrix for each. The paper applies three rounds of cross-attention and self-attention mechanisms, allowing the image features and point cloud features to be gradually fused into the feature vector F f , which is then fed into the decoder.
The decoder (D) is an important component of the network, aimed at accurately estimating the specific locations of the missing parts of the point cloud while maintaining both local and global structures. The paper employs a joint feature embedding method to complete the full point cloud. First, the original input data P in 1 with N points undergoes a second round of farthest point sampling to obtain the sampling result P in 2 = F P S [ N / 2 ] ( P i n 1 ) . The decoder uses k branches to generate the point cloud, with each branch using multiple MLP layers to generate points, which are then combined into a new point cloud and finally concatenated with the farthest point sampling point cloud to form the final completed point cloud. The CMFN network proposed in this paper uses the Chamfer distance (CD) loss function [21] to evaluate the similarity between the completed point cloud P cp and the real point cloud P gt , guiding the network training.

3.2. Point Cloud Multi-Scale Extractor

The point cloud multi-scale feature extractor (MSFE) is a vital component of our proposed network, which is designed to effectively handle point clouds at different scales. This extractor takes as input two point clouds of varying resolutions, P in 1 and P in 2 , where the number of points is defined as P in = [ N , N / 4 ] = [ 2048 , 512 ] . Here, N represents the total number of points in the original point cloud, while the second scale point cloud P in 2 is derived through farthest point sampling (FPS) [7], which is a downsampling technique that selects points from the original point cloud to preserve its spatial distribution. The FPS algorithm begins by randomly selecting an initial point from the point cloud. Subsequently, it iteratively selects the point that is farthest from the already chosen points. This approach ensures that the selected points are well distributed across the point cloud, maintaining the overall shape and structure.
This method effectively captures the geometric characteristics of the original point cloud while reducing its size, facilitating more efficient processing in subsequent stages. The input data of different scales are processed through the Feature Embedding Layer (FEL), resulting in point cloud feature vectors F p 1 and F p 2 of the same dimensionality. The FEL employs Dynamic Graph Convolution (DGCNN) [22] to extract features from the point cloud data.
Employing multi-scale feature extraction is to capture both local and global features of the point cloud effectively. The original point cloud P in 1 retains fine details such as edges and surfaces, while the downsampled point cloud P in 2 captures broader structural characteristics. By processing the point cloud at these two different scales, we can extract features that represent varying levels of detail. This multi-scale approach ensures a comprehensive representation of the point cloud, which is crucial for accurate point cloud completion. The combination of features extracted from both scales enhances the network’s ability to reconstruct missing parts of the point cloud, leading to improved performance in subsequent tasks.

3.3. Image Feature Extractor

To better utilize the texture and structural information of images to assist in the completion of incomplete point clouds with higher accuracy, this paper proposes an image feature extractor based on the Rotating Channel Attention (RCA) mechanism.
As shown in Figure 2 the Rotating Channel Attention (RCA) consists of three parallel branches. The first and second branches involve rotation and interaction between the channel dimension C and the spatial dimensions H and W (rotating 90° clockwise around the z-axis and y-axis, respectively), achieving information exchange between different dimensions. The last branch constructs spatial attention using the spatial dimensions H and W (rotating 90° clockwise around the x-axis), focusing on local information. The outputs of the three branches are then aggregated using an averaging method to produce the final output.
The calculation process is as follows:
y 1 = Conv [ R x ( I in ) ] I in = W 0 I in y 2 = Conv [ R y ( I in ) ] I in = W 1 I in y 3 = Conv [ R z ( I in ) ] I in = W 2 I in
y = Aggr { y 1 , y 2 , y 3 }
where R x ( · ) , R y ( · ) , and R z ( · ) represent the rotation operations around the x-axis, y-axis, and z-axis, respectively, and Conv ( · ) denotes the convolution operation. The resulting coefficient matrices after rotation and convolution are represented by W 0 , W 1 , and W 2 . The Aggr ( · ) operation denotes the aggregation of these matrices to generate the final result y.
The image input I in is processed through the RCA operation before being fed into the ResNet18 [23] network, ultimately yielding the image feature F I . The RCA module shares information between different channels of the image, enriching the image information in each channel and further enhancing the cross-modal feature fusion effect.

3.4. Cross-Model Feature Fusion (CMF)

The Cross-Model Feature Fusion (CMF) module integrates features from two modalities using cross-attention and self-attention mechanisms. As shown in Figure 3, multi-scale point cloud data P i n 1 , P i n 2 , and image data I i n are processed by their respective feature extractors to obtain point cloud features F P 1 , F P 2 , and image features F I . In the CMF module, F I is fused with F P 1 and F P 2 sequentially to obtain fused features F 1 and F 2 . These are then concatenated and passed through an MLP layer to produce the final fused feature F f .
Multi-head attention is used to enhance feature representation and network robustness. The image features F I serve as the key (K) and value (V) matrices, while the point cloud features F P 1 and F P 2 act as the query Q matrices. Taking the F P 1 branch as an example,
( Q , K , V ) = ( F P 1 , F I , F I ) × ( W q , W k , W v )
Q , K R N × d a , V R N × d v
Here, W Q , W K , and W V are learnable linear weights. The attention weight coefficient matrix a ˜ i j is computed using Q and K, then normalized via softmax:
W ˜ = a ˜ i j = Q · K T
a ¯ = softmax ( a ˜ i j ) = exp ( a ˜ i j ) k exp ( a ˜ k j )
W = a i j = a ¯ i j k a ¯ i k
F 1 = W · V
Using image features as K and V matrices, and point cloud features as Q matrices allows the rich texture and structural information from images to guide the completion of incomplete point clouds. Multiple cross-attention and self-attention operations enable gradual fusion of image and point cloud features.
The fused features F 1 and F 2 from two scales are concatenated and fed into an MLP layer to obtain the final encoded feature F f :
F f = MLP ( Cat ( F 1 , F 2 ) )
C a t ( · ) represents concatenation of two-feature matrices. F 1 and F 2 have dimensions [B,1,N,F], concatenated to [B,2,N,F], and mapped to F f with dimension [B,N,F], containing rich global and local feature information.

3.5. Point Cloud Similarity Evaluation Metrics

The network model evaluation metrics employed in this study are the Chamfer distance (CD) and the F-Score [4]. The CD loss function is utilized within the text to guide the network training. The similarity evaluation criterion between the completed point cloud and the real point cloud is based on the Chamfer distance, which is defined as follows:
L CD Y , Y ^ = 1 2 N y Y min y ^ Y ^ y y ^ + y ^ Y ^ min y Y y ^ y
where Y represents the real point cloud, y represents a point in the real point cloud, Y ^ represents the completed point cloud, y ^ represents a point in the completed point cloud, and N represents the number of points in the point cloud. When training with the CD loss function, the goal of the optimization algorithm is to minimize the CD loss, thereby allowing the completed point cloud to better match the real point cloud.
The F-Score is a shape-matching-based point cloud similarity evaluation metric used to detect any shape deviations between two point clouds. The calculation steps for the F-Score are as follows:
  • For each point p real in the real point cloud, find the closest point p completed in the completed point cloud;
  • If the Euclidean distance between p real and its closest point p completed is less than a threshold (in experiments, the threshold is set to 0.001 times the diameter of the point cloud), then point p real is considered to be successfully matched; otherwise, it is considered a false match;
  • Calculate the number of true matches (TP), false matches (FP), and the number of points not successfully matched (FN);
  • Based on the definitions of precision and recall, compute the F-Score:
Precision = TP TP + FP , Recall = TP TP + FN
F - Score = 2 × Precision × Recall Precision + Recall
where the precision rate represents the proportion of correctly matched points, and the recall rate represents the proportion of points that are correctly matched. The F-Score takes into account both the precision and recall rates, providing a comprehensive measure of the similarity between two point clouds.
In point cloud completion tasks, Chamfer distance (CD) and F-Score are commonly used evaluation metrics that assess model performance from different angles. CD primarily measures the geometric distance between the generated and real point clouds, focusing on geometric accuracy and detail preservation. It calculates the average of the nearest neighbor distances between points in the two point clouds. A smaller CD value indicates better geometric matching and richer details. F-Score, on the other hand, evaluates the overall distribution and integrity of the completed point cloud by considering both precision and recall. It assesses the percentage of correctly reconstructed points or surfaces within a certain distance threshold. A higher F-Score means the generated point cloud has a more reasonable distribution, with fewer outliers or holes.
In practice, these two metrics are often used together because they provide complementary information about the model’s performance, allowing for a comprehensive assessment of the effectiveness of point cloud completion algorithms.

4. Results and Discussion

To test the effectiveness of the network algorithm, all experiments in this paper are conducted on the public dataset ShapeNet-ViPC and ModelNet40ViPC, which is established from ModelNet40 [24]. Additionally, this paper’s algorithm is compared with recent relevant deep learning methods for point cloud completion, and the effectiveness of each module within the network is verified through ablation experiments.

4.1. Datasets and Experimental Configuration

The network algorithm proposed in this paper is trained and tested on the public dataset ShapeNet-ViPC and ModelNet40ViPC to evaluate its effectiveness. The ShapeNet-ViPC dataset comprises 13 object categories with a total of 38,328 objects, each featuring 24 incomplete point clouds with different occlusion perspectives and 24 corresponding real point clouds. For the 24 incomplete point cloud perspectives, there are 24 images in PNG format. The real point clouds contain 3500 points each, the incomplete point clouds contain 2048 points each, and the image pixel size is 24 × 24. Following convention, 80% of the data is used for training and 20% for testing. The ModelNet40ViPC dataset is based on the CAD models from ModelNet40 [24] and includes 12,311 objects from 40 different categories. Each object in this dataset contains 32 views of incomplete point clouds.
Experimental Conditions: The experiments in this paper are conducted on the Ubuntu operating system using the PyTorch 1.10.2 framework with the corresponding version of CUDA, and the GPU device used is the NVIDIA V100 in AutoDL Compute which is a company dedicated to offering powerful and efficient computing resources. Experimental parameter settings are shown in Table 1.
The input point cloud data is given by P in = N , N 4 = 2048 , 512 . After passing through the point cloud multi-scale feature extractor, the features are extracted with dimensions of [ 128 , 256 , 128 ] . The image feature extractor uses a ResNet18 network embedded with the RCA module, with image input I in = [ 128 , 3 , 224 × 224 ] , and the output is F I = [ 128 , 256 , 14 × 14 ] . In the cross-modal feature fusion module (CMF), the multi-head is set to 4, with inputs F scale 1 and F I , producing fused features F fused 1 and F fused 2 at different scales, both with dimensions of [ 128 , 256 , 128 ] . In the decoder, the original point cloud undergoes farthest point sampling to obtain the sampled points P fps . The decoder has k branches, each producing 128 points, and the final completed point cloud Y ^ = 2048 points is formed by combining the points from all branches with the sampled points.

4.2. Experiment on ShapeNetViPC

This paper utilizes the Chamfer distance (CD) loss to guide the network training and test the effectiveness of the network. As previously mentioned, the principle of calculating the CD loss function is such that a smaller CD distance indicates greater similarity between two point clouds. Table 2 presents a comparison of the completion results between our CMFN algorithm and other completion algorithms. The bolded numbers indicate the best completion results for that category.
From Table 2, it can be observed that on the ShapeNet-ViPC dataset, multi-modal point cloud completion methods outperform single-modal point cloud completion methods, and our proposed CMFN network has achieved state-of-the-art (SOTA) performance among both single-modal and multi-modal algorithms. The completion effects for objects in eight categories show a certain improvement over the latest multi-modal completion algorithm, XMFnet, especially in the category of lamps where the CD evaluation metric decreased by 41.38%, and the average metric across the eight categories decreased by 11.71%.
Table 3 presents a comparison of the results from nine point cloud completion algorithms on the ShapeNet-ViPC dataset, using the F-Score as the evaluation metric. The bolded numbers indicate the best completion results for each category.
It can be observed from Table 3 that the CMFN network proposed in this paper achieved optimal performance in seven categories according to the F-Score metric, with an average improvement of 3.1% over the XMFnet network, and a significant enhancement of 10.98% in the category of lamps. It is noteworthy that the performance of our algorithm in terms of the F-Score metric is slightly inferior to that of the CD metric, since our network is trained and tested using the CD loss function, which tends to perform poorly in point cloud uniformity distribution. Consequently, networks trained with the CD loss function may exhibit a decline in the evaluation of point cloud completion effectiveness when assessed with the shape-matching similarity metric of F-Score. Nonetheless, our algorithm still demonstrates an improvement in this metric.

4.3. Experiment on ModelNet40ViPC

The ModelNet40ViPC dataset [25] is based on the CAD models from ModelNet40 [24] and includes 12,311 objects from 40 different categories. The 40 categories are airplane, bathtub, bed, bench, bookshelf, bottle, car, chair, bowl, cone, cup, curtain, desk, door, dresser, flower_pot, glass_box, guitar, keyboard, lamp, laptop, mantle, monitor, night_stand, person, piano, plant, radio, range_hood, sink, sofa, stairs, stool, table, tent, toilet, tv_stand, vase, wardrobe, and xbox.
The point cloud incompleteness is primarily due to self-occlusion from the viewpoint, which is consistent with real-world point cloud sampling. The 32 different viewpoints are uniformly set on the spherical surface around the object model, shown as Figure 4. The number of points in the incomplete point clouds is variable. To verify the robustness of the network algorithm, the point cloud incompleteness rate is set between 60% and 90%. Image data are consistent with point cloud data, with 32 images rendered from the same 32 viewpoints, with a pixel size of [224 × 224] and in PNG format shown as Figure 5. The real point clouds in this dataset are obtained through Poisson sampling from the CAD models in ModelNet40, with point cloud sizes of 3500 and 2045, as shown in Figure 6. During training, the network uses point clouds with 3500 points, which are downsampled to 2048 points using farthest point sampling, focusing on the distribution of each category’s point cloud data. During testing, the real point clouds use 2048 points to more accurately test the completion performance of each algorithm. The ModelNet40ViPC dataset contains a total of 393,952 incomplete point cloud and image samples, and 24,622 real point cloud samples. In this paper, 50% of the samples are used for training and 50% for testing, using a unified training method to verify the completion performance of the network algorithm.
Table 4 presents the CD results for various algorithms on the ModelNet40ViPC dataset. The proposed CMFN network achieved the best performance in most categories, demonstrating superior completion accuracy compared to other methods. For instance, in the laptop category, CMFN achieved a CD value of 1.288, which is significantly lower than ViPC (10.552), CSDN (2.960), and XFMnet (1.531). This indicates that CMFN is more effective in reconstructing the fine details and overall structure of the laptop, resulting in a more accurate and complete point cloud. In the table category, CMFN achieved a CD value of 1.840, outperforming ViPC (9.296), CSDN (3.419), and XFMnet (2.107). This suggests that CMFN is better at capturing the structural details of the table, such as the tabletop and legs, leading to a more precise reconstruction. In the chair category, CMFN achieved a CD value of 2.580, which is better than ViPC (9.742), CSDN (4.792), and XFMnet (3.004). This demonstrates CMFN’s ability to accurately reconstruct the chair’s structure, including the seat, backrest, and legs. Overall, the CMFN network achieved the lowest average CD value of 2.669, compared to ViPC (9.990), CSDN (5.2528), and XFMnet (3.080). This indicates that the proposed network is more effective in reconstructing the missing parts of point clouds, resulting in a more accurate and complete representation of the objects.
Table 5 presents the F-Score results for various algorithms on the ModelNet40ViPC dataset. The CMFN network achieved the highest F-Score in most categories, demonstrating superior performance in terms of point cloud density and distribution. For example, in the laptop category, CMFN achieved an F-Score of 0.929, which is higher than ViPC (0.407), CSDN (0.578), and XMFnet (0.696). This indicates that CMFN not only reconstructs the missing parts accurately but also maintains a uniform density distribution in the completed point cloud, resulting in a more realistic and detailed representation of the laptop. In the table category, CMFN achieved an F-Score of 0.835, outperforming ViPC (0.467), CSDN (0.593), and XFMnet (0.710). This suggests that CMFN is better at preserving the density and distribution of points in the table’s structure, leading to a more accurate and complete reconstruction. In the chair category, CMFN achieved an F-Score of 0.879, which is better than ViPC (0.426), CSDN (0.432), and XMFnet (0.706). This demonstrates CMFN’s ability to maintain a uniform density distribution in the completed point cloud, resulting in a more accurate and detailed representation of the chair. Overall, the CMFN network achieved the highest average F-Score of 0.929, compared to ViPC (0.551), CSDN (0.518), and XFMnet (0.901). This indicates that the proposed network not only reconstructs the missing parts accurately but also maintains a uniform density distribution in the completed point clouds, leading to a more realistic and detailed representation of the objects.
The CMFN network’s superior performance is attributed to its multi-scale feature extraction, which processes point clouds at different scales (2048 and 512 points) to capture both local and global features, ensuring a comprehensive representation. The use of farthest point sampling (FPS) preserves the spatial distribution and overall structure of the point cloud. The RCA module enhances image feature extraction by sharing information between channels, enriching the feature representation and providing more accurate guidance for completion. The multi-level self-attention mechanisms in the CMF module gradually fuse image and point cloud features, leveraging complementary information from both modalities. The decoder’s multi-branch completion method, combined with Chamfer distance loss during training, ensures accurate reconstruction by minimizing the distance between the completed and real point clouds. Additionally, the network demonstrates robustness and generalization across various categories, effectively handling different types of incomplete point clouds and maintaining local and global structures during reconstruction.

4.4. Visualization

The visual comparison between the network algorithm proposed in this paper, CMFN, and the benchmark algorithm XMFnet is shown in Figure 7. The figure presents the incomplete point clouds for six categories, the completion results from the XMFnet network, the completion results from the CMFN network, and the visualization of the real point clouds. By comparing the visual results from both networks with the real point clouds, it can be observed that the completion effect of the CMFN network is generally superior to that of the XMFnet network, with more distinct structural details and a more uniform density distribution in the point clouds. In the airplane category completion, the CMFN network provides more precise details compared to XMFnet, with the lower engine contour details being more similar to the real point cloud. The XMFnet network failed to complete the shape of the lower wing of the airplane, while the CMFN network captured the details of the lower wing, and the comparison of the nose and tail completion also reflects the advantages of the network proposed in this paper. In the table lamp category, the CMFN network’s completion result is closer to the real point cloud, whereas the XMFnet network’s completion result has more outliers in the lampshade and fails to complete the details of the lamp column and base contour. The CMFN network not only completes the structural details of the missing parts but also outperforms the former in terms of uniform point cloud density distribution. Additionally, in the table category, it can be observed that the network proposed in this paper significantly outperforms the XMFnet network in completing the contour of the missing tabletop area, which is more similar to the real point cloud contour, and the point cloud density distribution on the tabletop is more uniform.

4.5. Ablation

The ablation studies were conducted to validate the effectiveness of the multi-scale module and the Rotating Channel Attention (RCA) module in the CMFN network. The experiments were performed on the ShapeNet-ViPC dataset, and the results were compared with the benchmark network XMFnet. The network models were configured as follows: complete CMFN model including both the RCA module and the multi-scale module, model with only RCA module including only the RCA module for image feature extraction, model with only multi-scale module including only the multi-scale module for point cloud feature extraction, and XMFnet, which is the benchmark network for comparison.
The experimental results are compared in Table 6. It is worth noting that while validating the effectiveness of the multi-scale module, this paper also verified the impact of different point cloud numbers at the second scale on the experimental results.
We set the first scale point cloud input to 1024, which helps maintain the original point cloud structural information during feature extraction. The second scale point cloud input is set to 512. After multiple graph convolutions, this resolution better captures the global features of the point cloud. Ablation experiments show that the second scale input of 512 outperforms 1024, proving that 512 is more effective in extracting global feature information.
The RCA module enhances image feature extraction by sharing information between channels, which significantly improves the network’s ability to capture structural information from images. This is evident in the results for the lamp category, where the RCA module achieved a CD value of 1.490, compared to 1.810 for XMFnet. The improvement can be attributed to the RCA module’s ability to effectively fuse multi-channel information, enriching the feature representation of the image data. This enhanced feature representation provides more accurate guidance for the point cloud completion task, leading to better performance.
The multi-scale module, using DGCNN with KNN clustering, effectively extracts both local and global point cloud features. The first scale, set to 2048 points, preserves the original point cloud’s structural information, while the second scale, set to 512 points, captures global features after multiple graph convolutions. The ablation study shows that the 512-point configuration outperforms the 1024-point configuration, confirming its higher global feature extraction capability. For example, in the lamp category, the multi-scale module with 512 points achieved a CD value of 1.327, compared to 1.452 for 1024 points. This indicates that the 512-point configuration is more effective in capturing global features, which are crucial for accurate point cloud completion.
The complete CMFN model, combining both the RCA module and the multi-scale module, achieved the best performance across all categories. For instance, in the lamp category, the CMFN model achieved a CD value of 1.061, which was significantly better than the RCA module alone (1.490), the multi-scale module alone (1.327), and XMFnet (1.810). Similarly, in the watercraft category, the CMFN model achieved a CD value of 0.788, compared to 0.823 for the RCA module, 0.828 for the multi-scale module, and 0.945 for XMFnet. In the cabinet category, the CMFN model achieved a CD value of 1.796, compared to 1.936 for the RCA module, 1.877 for the multi-scale module, and 1.980 for XMFnet.
The combined effect of the RCA and multi-scale modules can be attributed to their complementary nature. The RCA module enhances image feature extraction, providing rich structural information that guides the point cloud completion process. The multi-scale module, on the other hand, ensures that both local and global features of the point cloud are effectively captured and utilized. This synergy between the two modules leads to a more comprehensive and accurate representation of the point cloud, resulting in superior completion performance. The ablation studies clearly demonstrate the effectiveness of both the RCA module and the multi-scale module in the CMFN network. The RCA module enhances image feature extraction, while the multi-scale module improves point cloud feature extraction. The combination of these two modules in the complete CMFN model results in the best performance, indicating that they complement each other and significantly enhance point cloud completion accuracy.

5. Conclusions

This paper presents a novel point cloud shape completion network that enhances data feature extraction capabilities by employing a multi-modal feature fusion approach within an encoder–decoder structure. An RCA (Rotating Channel Attention) module is integrated into the image feature extraction process to improve the extraction of image features, leveraging the textural and structural information of images to assist in the extraction of point cloud features. In the point cloud feature extraction process, a multi-scale method is utilized to enhance the extraction capabilities, not only increasing the inter-point relatedness within each sample’s point cloud but also integrating the correlation between local and global feature information, thereby generating point clouds with more accurate structural information.
This paper employs a multilayer cross-attention mechanism to gradually fuse image features and point cloud features, achieving a multi-modal information feature-level fusion effect. The proposed algorithm compares favorably with both single-modal and multi-modal point cloud completion algorithms, achieving the best results. Specifically, when compared with the most recent multi-modal point cloud completion algorithm XMFnet, the average CD value was reduced by 11.71%, with a 41.38% reduction in the CD value for the lamp category. It is noteworthy that the proposed algorithm exhibited issues of local point cloud density inconsistency and a higher number of outliers during point cloud completion. Analysis indicates that the inconsistency in completion point cloud density is due to the use of the CD loss function to guide network training, while the excess of outliers may be attributed to insufficient local structural detail constraints during the point cloud generation process. Therefore, designing a more effective point cloud similarity evaluation method and a decoder with multi-modal feature constraints will be one of the future research directions in point cloud completion.

Author Contributions

Conceptualization, S.G. and J.W.; methodology, S.G. and K.X.; software, S.G. and B.H.; validation, B.H. and Y.M.; formal analysis, S.G.; resources, J.W.; data curation, B.H.; writing—original draft preparation, B.H.; writing—review and editing, S.G., B.H., and Y.M.; visualization, B.H.; supervision, J.W.; funding acquisition, K.X. All authors have read and agreed to the published version of the manuscript.

Funding

The National Natural Science Foundation of China, Project Name: Research on High-Quality Perception of Targets in Dynamically Complex Scenes; Grant Number: U24B20138.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Guo, Y.; Wang, H.; Hu, Q.; Liu, H.; Liu, L.; Bennamoun, M. Deep Learning for 3D Point Clouds: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4338–4364. [Google Scholar] [CrossRef] [PubMed]
  2. Mitra, N.J.; Pauly, M.; Wand, M.; Ceylan, D. Symmetry in 3D Geometry: Extraction and Applications. Comput. Graph. Forum 2013, 32, 1–23. [Google Scholar] [CrossRef]
  3. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Washington, DC, USA, 2017; pp. 77–85. [Google Scholar] [CrossRef]
  4. Qi, C.R.; Yi, L.; Su, H.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; pp. 5099–5108. [Google Scholar]
  5. Zhang, X.; Feng, Y.; Li, S.; Zou, C.; Wan, H.; Zhao, X.; Guo, Y.; Gao, Y. View-Guided Point Cloud Completion. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2021, Virtual, 19–25 June 2021; Computer Vision Foundation/IEEE: Piscataway, NJ, USA, 2021; pp. 15890–15899. [Google Scholar] [CrossRef]
  6. Zhu, Z.; Nan, L.; Xie, H.; Chen, H.; Wang, J.; Wei, M.; Qin, J. CSDN: Cross-Modal Shape-Transfer Dual-Refinement Network for Point Cloud Completion. IEEE Trans. Vis. Comput. Graph. 2024, 30, 3545–3563. [Google Scholar] [CrossRef] [PubMed]
  7. Tchapmi, L.P.; Kosaraju, V.; Rezatofighi, H.; Reid, I.D.; Savarese, S. TopNet: Structural Point Cloud Decoder. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; Computer Vision Foundation/IEEE: Piscataway, NJ, USA, 2019; pp. 383–392. [Google Scholar] [CrossRef]
  8. Schnabel, R.; Degener, P.; Klein, R. Completion and Reconstruction with Primitive Shapes. Comput. Graph. Forum 2009, 28, 503–512. [Google Scholar] [CrossRef]
  9. Liu, M.; Sheng, L.; Yang, S.; Shao, J.; Hu, S. Morphing and Sampling Network for Dense Point Cloud Completion. In Proceedings of the The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, 7–12 February 2020; AAAI Press: Washington, DC, USA, 2020; pp. 11596–11603. [Google Scholar] [CrossRef]
  10. Yuan, W.; Khot, T.; Held, D.; Mertz, C.; Hebert, M. PCN: Point Completion Network. In Proceedings of the 2018 International Conference on 3D Vision, 3DV 2018, Verona, Italy, 5–8 September 2018; IEEE Computer Society: Washington, DC, USA, 2018; pp. 728–737. [Google Scholar] [CrossRef]
  11. Yu, X.; Rao, Y.; Wang, Z.; Liu, Z.; Lu, J.; Zhou, J. PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 12478–12487. [Google Scholar] [CrossRef]
  12. Wang, J.; Cui, Y.; Guo, D.; Li, J.; Liu, Q.; Shen, C. PointAttN: You Only Need Attention for Point Cloud Completion. In Proceedings of the Thirty-Eighth AAAI Conference on Artificial Intelligence, AAAI 2024, Thirty-Sixth Conference on Innovative Applications of Artificial Intelligence, IAAI 2024, Fourteenth Symposium on Educational Advances in Artificial Intelligence, EAAI 2014, Vancouver, BC, Canada, 20–27 February 2024; Wooldridge, M.J., Dy, J.G., Natarajan, S., Eds.; AAAI Press: Washington, DC, USA, 2024; pp. 5472–5480. [Google Scholar] [CrossRef]
  13. Xiang, P.; Wen, X.; Liu, Y.; Cao, Y.; Wan, P.; Zheng, W.; Han, Z. SnowflakeNet: Point Cloud Completion by Snowflake Point Deconvolution with Skip-Transformer. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5479–5489. [Google Scholar] [CrossRef]
  14. Yang, Y.; Feng, C.; Shen, Y.; Tian, D. FoldingNet: Point Cloud Auto-Encoder via Deep Grid Deformation. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; Computer Vision Foundation/IEEE Computer Society: Piscataway, NJ, USA, 2018; pp. 206–215. [Google Scholar] [CrossRef]
  15. Pan, L. ECG: Edge-aware Point Cloud Completion with Graph Convolution. IEEE Robot. Autom. Lett. 2020, 5, 4392–4398. [Google Scholar] [CrossRef]
  16. Pan, L.; Chen, X.; Cai, Z.; Zhang, J.; Zhao, H.; Yi, S.; Liu, Z. Variational Relational Point Completion Network for Robust 3D Classification. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 11340–11351. [Google Scholar] [CrossRef] [PubMed]
  17. Aiello, E.; Valsesia, D.; Magli, E. Cross-modal Learning for Image-Guided Point Cloud Shape Completion. In Proceedings of the Advances in Neural Information Processing Systems 35: Annual Conference on Neural Information Processing Systems 2022, NeurIPS 2022, New Orleans, LA, USA, 28 November–9 December 2022; Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A., Eds.; [Google Scholar]
  18. Gharineiat, Z.; Kurdi, F.T.; Campbell, G. Review of Automatic Processing of Topography and Surface Feature Identification LiDAR Data Using Machine Learning Techniques. Remote. Sens. 2022, 14, 4685. [Google Scholar] [CrossRef]
  19. Michalowska, M.; Rapinski, J. A Review of Tree Species Classification Based on Airborne LiDAR Data and Applied Classifiers. Remote. Sens. 2021, 13, 353. [Google Scholar] [CrossRef]
  20. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; pp. 5998–6008. [Google Scholar]
  21. Fan, H.; Su, H.; Guibas, L.J. A Point Set Generation Network for 3D Object Reconstruction from a Single Image. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; IEEE Computer Society: Piscataway, NJ, USA, 2017; pp. 2463–2471. [Google Scholar] [CrossRef]
  22. Wang, Y.; Sun, Y.; Liu, Z.; Sarma, S.E.; Bronstein, M.M.; Solomon, J.M. Dynamic Graph CNN for Learning on Point Clouds. ACM Trans. Graph. 2019, 38, 146:1–146:12. [Google Scholar] [CrossRef]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity Mappings in Deep Residual Networks. In Proceedings of the Computer Vision—ECCV 2016—14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part IV. Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science. Springer: Berlin/Heidelberg, Germany, 2016; Volume 9908, pp. 630–645. [Google Scholar] [CrossRef]
  24. Wu, Z.; Song, S.; Khosla, A.; Yu, F.; Zhang, L.; Tang, X.; Xiao, J. 3D ShapeNets: A deep representation for volumetric shapes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7–12 June 2015; IEEE Computer Society: Piscataway, NJ, USA, 2015; pp. 1912–1920. [Google Scholar] [CrossRef]
  25. Liu, X.; Hou, B.; Wang, H.; Xu, K.; Wan, J.; Guo, Y. DuInNet: Dual-Modality Feature Interaction for Point Cloud Completion. arXiv 2024, arXiv:2407.07374. [Google Scholar]
Figure 1. Overall frame diagram of cross-modal multi-scale feature fusion.
Figure 1. Overall frame diagram of cross-modal multi-scale feature fusion.
Remotesensing 17 01448 g001
Figure 2. Schematic diagram of rotating channel attention.
Figure 2. Schematic diagram of rotating channel attention.
Remotesensing 17 01448 g002
Figure 3. Cross-Model Feature Fusion network architecture.
Figure 3. Cross-Model Feature Fusion network architecture.
Remotesensing 17 01448 g003
Figure 4. The 32 different viewpoints used to complete the point cloud.
Figure 4. The 32 different viewpoints used to complete the point cloud.
Remotesensing 17 01448 g004
Figure 5. Image data sampled from 32 different viewpoints.
Figure 5. Image data sampled from 32 different viewpoints.
Remotesensing 17 01448 g005
Figure 6. Incomplete point clouds sampled with Poisson sampling from CAD models.
Figure 6. Incomplete point clouds sampled with Poisson sampling from CAD models.
Remotesensing 17 01448 g006
Figure 7. Visualization results comparison chart of CMFN with different menthods.
Figure 7. Visualization results comparison chart of CMFN with different menthods.
Remotesensing 17 01448 g007
Table 1. Experimental parameter settings.
Table 1. Experimental parameter settings.
ParameterValue/Description
OptimizerAdaptive Moment Estimation (Adam) [23]
Initial Learning Rate0.001
Learning Rate DecayDecayed to 1/10 of its original value at the 25th and 125th epochs, respectively
Batch Size128
Total Epochs200
Table 2. Using CD evaluation index to compare the effect on ShapeNet-VIPC dataset.
Table 2. Using CD evaluation index to compare the effect on ShapeNet-VIPC dataset.
MethodsAvgAirplaneCabinetCarChairLampSofaTableWatercraft
Single-modalAtlasNet [8]6.0625.0326.4144.8688.1617.1826.0236.5614.261
FoldingNet [14]6.2715.2426.9585.3078.8236.5046.3687.0803.882
PCN [10]5.6194.2466.4094.8407.4416.3315.6686.5083.510
TopNet [7]4.9763.7105.6294.5306.3915.5475.2815.3813.350
ECG [15]4.9572.9526.7215.2435.8674.6026.8134.3323.127
PoinTr [11]8.3824.75310.4728.6829.3927.75410.9337.7847.291
PointAttN [12]6.6343.28210.7756.1327.1415.9219.7276.1643.592
VRC-Net [16]4.5982.8136.1084.9325.3424.1036.6143.9532.925
Multi-modalViPC [5]3.3081.7604.5583.1832.4762.8674.4814.9902.197
CSDN [6]1.6531.8733.2451.9431.8852.0963.4174.0092.236
XMFnet [17]1.4430.5721.9801.7541.4031.8101.7021.3860.945
Ours1.2740.5611.7961.6861.3761.0611.5821.3420.788
Table 3. Using F 1 @ 0.001 index to compare the effect on ShapeNet-VIPC dataset.
Table 3. Using F 1 @ 0.001 index to compare the effect on ShapeNet-VIPC dataset.
MethodsAvgAirplaneCabinetCarChairLampSofaTableWatercraft
Single-modalAtlasNet [8]0.4100.5090.3040.3790.3260.4260.3180.4690.551
FoldingNet [14]0.3310.4320.2370.3000.2040.3600.2490.3510.518
PCN [10]0.4070.5780.2700.3310.3230.4560.2930.4310.577
TopNet [7]0.4670.5930.3580.4050.3880.4910.3610.5280.615
ECG [15]0.7040.8800.5420.7130.6710.6890.5340.7920.810
PoinTr [11]0.6350.4620.8730.7030.6930.6550.5920.7870.792
PointAttN [12]0.7320.8920.5920.7320.7710.5440.5340.7960.820
VRC-Net [16]0.7640.9020.6210.7530.7220.8230.6540.8100.832
Multi-modalViPC [5]0.5910.8030.4510.5120.5290.7060.4340.5940.730
CSDN [6]0.6720.5630.5240.4430.4850.8260.4170.5230.836
XMFnet [17]0.7960.9610.6620.6910.8090.7920.7230.8300.901
Ours0.8220.9680.6960.7100.8120.8790.7480.8350.929
Table 4. Comparison of experimental results on the ModelNet40ViPC dataset ( C D × 10 3 ).
Table 4. Comparison of experimental results on the ModelNet40ViPC dataset ( C D × 10 3 ).
Methods
ModelsViPC [5]CSDN [6]XFMnet [17]Ours (RCA)Ours (Multi-Scale 512)Ours (Multi-Scale 1024)Ours (RCA+M-S)
glass_box10.8334.6142.7512.7683.2063.3632.770
range_hood11.4645.0793.0863.1533.2903.5302.799
laptop10.5522.9601.5311.5791.5051.6401.288
table9.2963.4192.1072.1031.9742.2381.840
bed7.6843.8502.7142.7462.5522.7112.387
chair9.7424.7923.0043.0362.7783.0232.580
bookshelf8.9484.4943.1963.2353.1963.2572.974
piano10.0265.7663.7633.7603.7304.0113.404
sink12.0495.7273.5123.5113.1443.4353.038
airplane4.2741.7581.0111.0271.0471.1410.894
dresser11.0664.8452.8692.9172.9653.1102.592
sofa7.6804.1022.8142.8532.8042.9382.560
bottle6.4832.4481.5601.5911.5111.5411.414
monitor7.2753.7562.4892.5372.6112.7082.310
tv_stand10.4935.4873.8043.6203.5883.7373.501
toilet11.2025.9423.8263.8843.8484.0663.554
stool12.7556.8314.4643.5193.1793.4043.267
xbox9.9824.2062.5032.5082.5112.5212.180
door6.7772.0311.0381.0771.1141.1810.988
night_stand12.1365.5803.6663.6933.4493.6513.230
bench9.0444.3532.5492.6112.2322.4482.094
vase11.4456.6753.9063.7023.5833.6743.419
tent12.5856.7193.9713.6793.3473.6793.163
desk12.5176.3214.0034.0183.7104.0013.507
car7.5894.0443.0513.1313.8753.0182.741
radio11.2335.8172.9833.0542.9013.0412.581
stairs9.8988.7044.6254.3643.5643.6443.462
guitar3.5211.4160.5060.5020.4550.4860.403
mantel11.9064.2922.3272.3542.6852.7082.219
cup13.5047.3485.2885.1054.6924.8574.708
plant9.1028.0765.9295.3154.7944.8064.721
curtain7.0492.8991.2831.3181.0781.1900.983
lamp19.82512.6224.9865.0733.3133.5463.931
flower_pot11.2097.6225.3015.0784.5754.7534.544
cone10.6553.8062.1972.1472.1492.3591.954
keyboard4.1741.7810.8240.8500.8430.8760.760
bathtub10.7985.0103.3713.2473.1213.5153.019
wardrobe9.8314.1812.5432.6032.6222.7812.279
bowl16.71515.1165.3045.0422.2085.0404.676
person6.2675.6232.5412.4682.1242.1792.050
all9.9905.25283.0803.0192.7812.9952.669
Table 5. Comparison of experimental results on the ModelNet40 dataset ( F 1 @ 0.001 ).
Table 5. Comparison of experimental results on the ModelNet40 dataset ( F 1 @ 0.001 ).
Methods
ModelsViPCCSDNXFMnetOurs (RCA)Ours (Multi-Scale 512)Ours (Multi-Scale 1024)Ours (RCA+M-S)
glass_box0.2070.4730.5070.5540.5160.5080.561
range_hood0.2650.4920.5780.5790.5630.5590.595
laptop0.4160.6620.7110.7170.7130.7010.743
table0.4690.6970.7590.7540.7610.7460.779
bed0.3240.5270.5880.5840.5880.5780.607
chair0.4330.6030.6590.6590.6630.6550.679
bookshelf0.3250.5280.5640.5800.5680.5640.597
piano0.2880.4760.5260.5440.5370.5270.560
sink0.3450.5790.6440.6400.6520.6320.667
airplane0.6300.7760.8880.8860.8720.8600.896
dresser0.2230.4780.5750.5720.5440.5260.582
sofa0.3150.4880.5540.5490.5450.5410.569
bottle0.4550.7390.8220.8200.8160.7930.839
monitor0.3910.5730.6370.6330.6170.6130.649
tv_stand0.2760.4840.5470.5440.5420.5310.563
toilet0.2760.4530.5040.5140.5030.4980.523
stool0.4130.5930.6320.6400.6760.6670.678
xbox0.2630.5380.6300.6240.6030.5870.645
door0.5420.7810.8030.8470.8300.8200.858
night_stand0.2420.4640.5430.5410.5390.5240.558
bench0.4710.6540.7110.7070.7210.7100.737
vase0.2910.5070.5850.5830.5860.5720.604
tent0.2870.5270.5930.5870.6050.5870.626
desk0.3440.5370.5880.5890.6020.5880.619
car0.3180.4990.5430.5440.5440.5330.562
radio0.3020.5310.6130.6190.6170.6020.645
stairs0.4520.5670.6340.6310.6640.6630.678
guitar0.7410.8870.9550.9500.9600.9540.965
mantel0.2850.5350.6440.6420.6200.6090.653
cup0.2170.4200.4740.4710.4840.4730.494
plant0.3650.4530.5160.5160.5310.5340.538
curtain0.5330.7590.8380.8340.8440.8300.860
lamp0.3830.5750.6670.6670.7010.6910.702
flower_pot0.2700.4130.4760.4810.4820.4760.493
cone0.3640.6780.7150.7750.7570.7730.798
keyboard0.6350.7810.8570.8700.8650.8570.889
bathtub0.2960.5230.5770.5730.5780.5640.598
wardrobe0.2700.5360.6400.6310.6080.5860.652
bowl0.2090.4530.5230.5190.5240.5170.536
person0.5210.6290.7070.7070.7200.7170.733
all0.3660.5710.6380.6410.6420.6320.663
Table 6. Comparison of ablation results.
Table 6. Comparison of ablation results.
RCAMulti-ScaleXMFnetCMFN
5121024RCA+Multi-Scale
Lamp1.4901.3271.4521.8101.061
Watercraft0.8230.8280.8380.9450.788
Cabinet1.9361.8771.9151.9801.796
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gu, S.; Xu, K.; Wan, J.; Hou, B.; Ma, Y. Research on Multi-Modal Point Cloud Completion Algorithm Guided by Image Rotation Attention. Remote Sens. 2025, 17, 1448. https://doi.org/10.3390/rs17081448

AMA Style

Gu S, Xu K, Wan J, Hou B, Ma Y. Research on Multi-Modal Point Cloud Completion Algorithm Guided by Image Rotation Attention. Remote Sensing. 2025; 17(8):1448. https://doi.org/10.3390/rs17081448

Chicago/Turabian Style

Gu, Shangtai, Ke Xu, Jianwei Wan, Baolin Hou, and Yanxin Ma. 2025. "Research on Multi-Modal Point Cloud Completion Algorithm Guided by Image Rotation Attention" Remote Sensing 17, no. 8: 1448. https://doi.org/10.3390/rs17081448

APA Style

Gu, S., Xu, K., Wan, J., Hou, B., & Ma, Y. (2025). Research on Multi-Modal Point Cloud Completion Algorithm Guided by Image Rotation Attention. Remote Sensing, 17(8), 1448. https://doi.org/10.3390/rs17081448

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop