Next Article in Journal
Real-World Urban Light Emission Functions and Quantitative Comparison with Spacecraft Measurements
Next Article in Special Issue
A Global Structure and Adaptive Weight Aware ICP Algorithm for Image Registration
Previous Article in Journal
Spatio-Temporal Variability Analysis of Vegetation Dynamics in China from 2000 to 2022 Based on Leaf Area Index: A Multi-Temporal Image Classification Perspective
Previous Article in Special Issue
PBFormer: Point and Bi-Spatiotemporal Transformer for Pointwise Change Detection of 3D Urban Point Clouds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Object-Level Building-Matching Method across 2D Images and 3D Point Clouds Based on the Signed Distance Descriptor (SDD)

1
Key Laboratory of Advanced Marine Communication and Information Technology, Ministry of Industry and Information, Harbin 150009, China
2
College of Information and Communication Engineering, Harbin Engineering University, Harbin 150009, China
3
Harbin Aerospace Star Data System Science and Technology Co., Ltd., Harbin 150028, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(12), 2974; https://doi.org/10.3390/rs15122974
Submission received: 4 May 2023 / Revised: 31 May 2023 / Accepted: 5 June 2023 / Published: 7 June 2023

Abstract

:
In this work, a novel object-level building-matching method using cross-dimensional data, including 2D images and 3D point clouds, is proposed. The core of this method is a newly proposed plug-and-play Joint Descriptor Extraction Module (JDEM) that is used to extract descriptors containing buildings’ three-dimensional shape information from object-level remote sensing data of different dimensions for matching. The descriptor is named Signed Distance Descriptor (SDD). Due to differences in the inherent properties of different dimensional data, it is challenging to match buildings’ 2D images and 3D point clouds on the object level. In addition, features extracted from the same building in images taken at different angles are usually not exactly identical, which will also affect the accuracy of cross-dimensional matching. Therefore, the question of how to extract accurate, effective, and robust joint descriptors is key to cross-dimensional matching. Our JDEM maps different dimensions of data to the same 3D descriptor SDD space through the 3D geometric invariance of buildings. In addition, Multi-View Adaptive Loss (MAL), proposed in this paper, aims to improve the adaptability of the image encoder module to images with different angles and enhance the robustness of the joint descriptors. Moreover, a cross-dimensional object-level data set was created to verify the effectiveness of our method. The data set contains multi-angle optical images, point clouds, and the corresponding 3D models of more than 400 buildings. A large number of experimental results show that our object-level cross-dimensional matching method achieves state-of-the-art outcomes.

1. Introduction

Building matching (BM) is the task of determining the corresponding data of buildings in query data from a database with given geographic labels and other information, which can be applied to the real-time positioning of unmanned aerial vehicles (UAV) [1,2,3], visual pose estimation [4,5,6], and 3D reconstruction [7] in remote sensing. The most common data type used for BM is images. Image-based retrieval methods for BM (e.g., [8,9,10,11]) attempt to identify similar database images that depict the same landmarks as the query image. Typically, the retrieved images are ranked according to a given similarity metric (e.g., the L1 norm distance between Bag-of-Words vectors [12], L2 norm distance between the compact representation and vector [13,14]) to obtain the BM result. In localization tasks, the position of the best-matching database image is usually treated as the position of the query image, or the positions of the top N images are fused to obtain the position of the query image [15,16,17]. With the rapid development of deep learning, various end-to-end deep methods have been studied, significantly improving the accuracy of BM.
Although the above methods have led to achievements, it is well-understood that their BM accuracy is largely dependent on the quality and quantity of images in the database. Specifically, if the images of the corresponding area in the database have significant differences from the matching image in terms of shooting angles and lighting conditions, accurate BM cannot be achieved in the given area. The images in the database are downloaded in bulk from the network. On the one hand, the number of images of tourist attractions or landmarks in the network is much higher than that of ordinary buildings, which leads to a highly uneven distribution of BM accuracy. When the building image to be matched appears infrequently in the network, it is not easy to obtain accurate matching results. On the other hand, for large-scale maps [18], this method requires a large number of images in the database as a support, which requires high computing power and storage resources. When the target carrier of the localization task is an unmanned aerial vehicle that has lost network control signals, it will be difficult to complete BM without the support of ground data resources and computing resources. Compared with two-dimensional image data, three-dimensional point cloud data composed of coordinates do not have the problem of the shooting angle and light conditions and do not require the storage a large amount of image data in the database. Therefore, compared with the image-matching positioning method, the image-point-cloud-matching method requires less computing power and storage resources at the terminal. Secondly, the image-point-cloud-matching method is not affected by the quality and quantity of images in the database, reducing the possibility of erroneous matching caused by the problem of data imbalance in the database. Therefore, in order to overcome the serious dependence of the localization task on the image database and reduce the demand for computing and storage resources for BM-based localization tasks, the pixel-level cross-dimensional matching method for 2D optical images and 3D point clouds provides a promising strategy, as shown in Figure 1a.
The current cross-dimensional data pixel-level matching methods can be divided into three categories. The typical process of the first category is to first recover the 3D structure of the scene [19,20], usually reconstructed from images taken at equal intervals using motion structure (SfM) [21,22] and multi-view systems (MVS) [23] as the input. Each 3D point is triangulated using multiple 2D local features (such as SIFT [24]) and associated with the corresponding image descriptor. Then, the pixel-level cross-dimensional correspondence between the local feature descriptor in the query image and the 3D point descriptor is found [15]. These pixel-level descriptors are homogeneous, because the points inherit the descriptors of the corresponding pixels in the reconstructed 3D scene [25]. The second category of methods identifies associations between different-dimensional data by mapping descriptors from different domains to a shared latent space. However, they only construct block-by-block descriptors, which typically lead to coarse-grained matching results. Instead, 3DTNet [26] takes 2D and 3D local patches as the input and extracts 2D features from 2D patches using unique 3D descriptors that help it to learn 3D patches. However, 3DTNet is only used for 3D matching. The network uses 2D features as auxiliary information to render 3D features more discriminative. Recently, learning descriptors that allow for direct matching and retrieval between 2D and 3D local patches have been proposed with 2D3DMatch-Net [27] and LCD [28]. Other research has established a connection between 2D and 3D features for specific applications, such as object pose estimation. Additionally, some methods achieve cross-dimensional matching through registration, such as that described in [29], by converting registration problems into classification and inverse camera projection optimization problems using relative rigid transformation for cross-dimensional matching. The authors of [30] proposed a method that uses a two-stage approach to align two inputs of data in a virtual reference coordinate system (virtual alignment) and then compare and align the data to complete the matching. However, even if accurate descriptors can be extracted from 2D images and 3D point clouds, the above two categories of methods still cannot establish accurate pixel-level cross-dimensional matching relationships. There are two reasons for this phenomenon. Firstly, due to the sparsity of point clouds, local point descriptors can be mapped to many pixel descriptors in 2D images, increasing matching ambiguity. To address this problem, Liu et al. [30] proposed a large-scale camera positioning method based on 2D–3D data matching, adding a disambiguation module that uses global contextual information to solve the problem of matching ambiguity. Secondly, as 2D images are usually a 2D mapping of the appearance of the scene [31], while 3D point clouds encode the structure of the scene, there are significant differences between the attributes of 2D images and 3D point clouds, and the descriptor loss functions of existing 2D or 3D local feature descriptions [32,33,34] cannot achieve accurate convergence in cross-dimensional matching tasks. Therefore, it is important to perform object-level cross-dimensional data matching for more effective solutions to the above problems.
In the task of matching building images and point clouds, object-level cross-dimensional data-matching methods match the building objects in the images and the point clouds as the core, as shown in Figure 1b. Compared with pixel-level matching methods, object-level matching methods extract joint descriptors from the data containing global information of the targets and map them to the descriptor space for matching. The global feature extraction method, which focuses on the overall target effectively, alleviates the problem of fuzzy matching caused by the sparsity of point clouds, and the descriptor space solves the problem of attribute differences in cross-dimensional data, reducing the dependence of traditional image-based positioning methods on large amounts of image data. This concept provides new ideas for various fields, such as the precise self-positioning of drones in the state of control signal loss, urban management, smart city construction, high-quality building shape reconstruction, and so on. The core of this method lies in the choice of the descriptor space. By choosing a better descriptor space, one can obtain more accurate cross-dimensional matching results. The authors of [35] proposed a joint global embedding method for 3D shapes and images to solve retrieval tasks. By directly binding handcrafted 3D descriptors to learned image descriptors, cross-dimensional descriptors for object-level retrieval tasks were generated. The authors of [36] proposed a deep-learning-based cross-dimensional object-level descriptor space occupancy probability descriptor (SOPD) which uses the occupancy probability of each unit space of the object as the cross-dimensional descriptor.
In this study, we designed a plug-and-play Joint Descriptor Extraction Module (JDEM) to extract our proposed new joint descriptors, called Symbolic Distance Descriptors (SDD). The SDD utilize the 3D geometric invariance of building objects, using features that contain their 3D structural information as descriptors. This not only overcomes the inherent differences in attributes between 2D and 3D data but also reduces the fuzzy matching caused by the similarity between pixel-level descriptors. In addition, we propose Multi-View Adaptive Loss (MAL) to improve the adaptability of the image encoder module to images taken from different angles and to enhance the robustness of the joint descriptors. To achieve cross-dimensional BM, a corresponding database is required. Although Yan et al. [36] proposed a 2D–3D object-level building-matching data set, only its 3D point clouds were obtained from the real world. We constructed a fully real-world 2D–3D cross-dimensional object-level building-matching data set called 2D-3D-CDOBM. The data set contains multi-angle optical images, point clouds, and the corresponding 3D models for over 400 buildings. We conducted extensive experiments on our data set to verify that our proposed descriptors can accurately perform cross-dimensional object-level matching tasks.
The important contributions of this paper are as follows:
  • A novel cross-dimensional object-level BM method based on the SDD is proposed. The method compensates for the modal differences between 2D optical images and 3D point clouds in object-level building data.
  • A plug-and-play JDEM is proposed. JDEM utilizes the three-dimensional geometric invariance of objects in the real physical world to obtain the same domain features by mapping the different dimensional data of objects to a three-dimensional space.
  • MAL is proposed to reduce the distance of descriptors extracted from images of the same object from different angles in the SDD space. The loss function improves the adaptability of the image encoder module to images from different angles.
  • A cross-dimensional object-level building-matching data set composed of real data, named 2D-3D-CDOBM, is proposed to verify the effectiveness of our method. It includes optical images of buildings taken from different angles, LiDAR point clouds of buildings scanned in different years, and hundreds of corresponding 3D models of buildings. This data set is available to download at “https://github.com/HEU-super-generalized-remote-sensing/Cross-dimensional-Object-level-Matching-Data-Set.git (accessed on 4 June 2023)”.

2. Materials and Methods

This section describes the details of the proposed descriptor, SDD (Section 2.1), the algorithm structure of the proposed cross-dimensional matching method (Section 2.2), the used loss function (Section 2.3), the proposed data used for the training, verification, and testing of the algorithm (Section 2.4), evaluation metrics (Section 2.5), and the training platform and parameter settings (Section 2.6).

2.1. SDD

Under ideal conditions, the imaging mechanism of 2D images can be simplified to a pinhole imaging model in which the captured object is projected onto the photosensitive element through the pinhole. Therefore, the images can be regarded as the projection of the 3D world scene to the 2D space. The image data format is h × v, where h is the number of horizontal pixels, and v is the number of vertical pixels. However, in the projection process, the loss of information is likely to occur due to the positional relationship, such as the incomplete structure of the occluded object in the image caused by the occlusion relationship.
While 3D point clouds are usually obtained via LiDAR scanning, by emitting a laser beam towards an object, LiDAR receives the laser radiation reflected by the scene and produces a continuous analog signal. At last, LiDAR restores this to a point cloud of the object scene [37,38]. The data form of the point cloud is shown in (1):
p i = ( x , y , z ) P = { p 1 , p 2 , p 3 , p 4 , }
where x, y, and z are the relative position coordinates of the points pi in the point cloud in the world’s physical coordinate system. LiDAR is an active sensor, and the number of point clouds generated by it is positively correlated with the LiDAR scanning time. The point density at the top of the building is higher than that on the sides of the building. Therefore, the matching of the 2D images and the 3D point clouds requires effort.
However, neither 2D images nor 3D point clouds are naturally generated. They are both derived from the 3D physical world via mapping or projection through corresponding sensors. Although the data formats are different, both are embodiments of the 3D physical world object containing the corresponding 3D information, that is, the 3D geometric invariance. According to this characteristic, by extracting 3D geometric descriptors from 2D images and 3D point clouds, one can repair their missing information and map descriptors extracted from different dimensional data to the same descriptor space. Therefore, the key to the cross-dimensional matching method is descriptor extraction.
Our method uses the SDD as a cross-domain descriptor extracted from different dimensional data. The SDD is a descriptor obtained via sampling of the Symbol Distance Feature (SDF) [39] at the threshold τ. SDF is a three-dimensional feature that describes the distance from any point in the normalized metric space to the model boundary, as shown in Figure 2b. If the sampling point is within the boundary, its SDF feature for the current sampling point is set to negative. Conversely, if the point is outside the boundary, its SDF feature is set to positive. The farther the sampling point is from the boundary, the greater the modulus of SDF will be. Since SDF is a description of distance, the threshold τ is usually set to zero.
Similar to the SDD, [36] the Spatial Occupancy Probability Feature (SOPF) is used to describe the three-dimensional structure of the building, that is, the probability of the current sampling point being inside the model. Therefore, the SOPF of the sampling points outside the model surface is set to 0, and that inside is set to 1 (100%), as shown in Figure 2c. Similarly, the SOPD also uses the threshold τ1 to sample the SOPF, and the threshold τ1 is usually set to 0.5 (50%). Compared with classification problems, deep learning networks have a better adaptability to regression problems. Therefore, compared with the SOPD, the network has a more significant learning ability and better learning effect for the SDD, which is also reflected in the experimental section (Section 3).

2.2. Cross-Dimensional Matching Method

There are many mature image feature extraction methods, such as VGG [40] and ResNet [41], and many mature point cloud feature extraction methods, such as PointNet [42] and PointNet++ [43]. In deep learning, the above encoder completes the front-end feature extraction function. The mapping relationship between the data and feature vectors of different modes is established. At this time, the formal unification task for different modal data features is completed. However, although these features contain information on different modal data, they cannot fully reflect the three-dimensional geometric invariance of the object. Therefore, we propose a plug-and-play JDEM module which can extract the SDD for cross-modal data matching after inserting the JDEM module. The classification results can be obtained through a linear layer after extracting the SDD.
A diagram of the structure of the cross-dimensional data matching method is shown in Figure 3. Overall, our cross-dimensional matching method consists of two parts: the front-end traditional feature encoders and JDEM. The feature vector extracted by the front-end encoder and the random sampling point P are used as the input for the JDEM, and the SDD feature value satisfying the threshold requirement in the random sampling point P is used as an output to realize the mapping from a one-dimensional feature vector to a three-dimensional space descriptor. In the training process, it is necessary to input the sampling point P randomly. On the one hand, this can reduce the demand for memory; on the other hand, it can improve the network’s generalization ability. In the training iteration process, the sampling point P is obtained by sampling the normalized coordinate points corresponding to the ground truth of the building SDF. The method compares the SDD value corresponding to the sampling point P obtained through the network’s prediction with the ground truth and calculates the loss function to guide the network’s training. It is worth noting that the SDD is a three-dimensional descriptor that can be visualized to facilitate the observation of the network training process and results.
Specifically, in order to verify the optimal encoder combination for our cross-dimensional matching method, we used the ResNet and PointNet series as the front-end image and point cloud encoder, respectively. ResNet18/34/50/101/152 refers to different network depths. The difference between PointNet and PointNet-k is whether or not the Knn module is added. After adding the Knn module, the point cloud will be searched for the nearest k points in each point space before entering the encoder. Then, these k + 1 points are merged and input into the encoder to improve the regional generalization of the encoder. Inspired by OccNet [44], we designed several different JDEM modules. The whole encoder combination is shown in Table 1, and the structures of the different JDEM modules are shown in Figure 4.

2.3. Loss Function

The fusion of multiple loss functions is used to calculate our loss function. The overall objective function L is a multi-task objective function divided into multistage weight loss and MAL. Multistage weight loss was proposed by the authors of [45].

2.3.1. Multistage Weight Loss

Multistage weight loss is a function that calculates the difference between the SDD extracted from the image or point cloud and the ground truth. Ground truth is the SDF obtained by calculating the distance between the point and the model boundary in which the point is sampled from the unit space, and the model is the normalized 3D model of the buildings.
Compared with the typical method for calculating feature loss, which treats each position in the metric space uniformly, as for the L1 norm showed in (2), multistage weight loss focuses more on the selection of the boundary points on the model surface and gives greater attention to the points near the model surface. In contrast, the points far from the model surface are appropriately ignored, as shown in (3). In this way, the network’s learning ability for building structures can be enhanced, and the SDD extraction effect can be optimized. At the same time, although points far away from the model’s surface receive less attention than others, they are less challenging to learn; thus, there is no negative impact on the results.
L ( F β ( I ) , G T ) = n = 1 N | F β ( I n ) G T n |
where In is the image, Fβ is the SDF feature extraction network, GT is ground truth, N is the number of images.
L ( F β ( I ) , G T ) = 1 B b = 1 B n = 1 N w ( G T n b ) × | F β ( I n ) G T n | b
where In is the image; Fβ is the SDF feature extraction network; GT is ground truth; N is the number of images; B is the number of sampling points; w(GTnb) is the weight corresponding to the current sampling point, determined by GTnb; and GTnb is the SDF ground truth of the current sampling point, that is, the distance between the current point and the surface of the model. w(GTnb) is shown in (4):
w ( G T n b ) = { 2 , G T n b 0 . 02   4 , 0 . 02 < G T n b < 0 . 02 1 , G T n b 0 . 02
where GTnb is the SDF ground truth of the current sampling point.
Through the new feature loss function, the difference between the predicted value and the ground truth of the critical points can be amplified so that the feature extraction network can extract more accurate features through a more authentic relationship.

2.3.2. MAL

The same building is usually displayed in multiple remote-sensing images taken from different angles. Due to the difference in perspective, the buildings’ features extracted from different images are usually different. Therefore, the concept of MAL is proposed to reduce the feature distance of the images of the same building taken from different angles in the feature domain. MAL can improve the adaptability of the network to images of the same building taken from different angles and enhance the robustness of the features. Specifically, the image data, the building SDF ground truth, and the building ID are input into the network simultaneously during training. The distance of the SDD extracted from the image of the same building ID in the feature space is calculated, which alleviates the problem of cross-dimensional object-level matching errors caused by perspective differences. The calculation method of MAL is shown in (5) and (6):
L c l s ( I a , I b ) = 1 N n = 1 N ω e p o c h × exp ( q × D ( I a , I b ) + m )
D ( I a , I b ) = n = 1 N ( F β ( I a ) n F β ( I b ) n ) 2
where Ia and Ib are image data of the same building ID, N is the number of the building ID, w and q are weights, m is offset, D is the SDD feature distance of images taken from different angles, and Fβ is the SDD feature extraction network.
SDD is the mapping of the 3D structure of the building in the real physical world, but there are many building structures that are similar or even the same. Therefore, MAL only focuses on the distance of the SDDs extracted from the images of the same building ID and does not pay attention to others. The formula shown in (7) is used to load the data so as to improve the probability of different-angle images of the same building appearing in the same batch and not affecting the network’s learning ability:
d a t a _ i n d e x = { a , a + n a , a n a , , b , b + n b , b n b } , a { 0 , S }
where S is the number of remaining samples in the current epoch, a and b are two random numbers from 0 to S, and na and nb are the numbers of images of the current building taken from different angles in the current epoch.
Since the role of MAL is to shorten the distance of images of the same building taken from different angles in the descriptor space, MAL is not used at the beginning of the training. After the original loss reaches the best fit, MAL is added for training until the new best fit point is reached.

2.4. Proposed 2D-3D-CDOBM Data Set

In this section, the method for obtaining our data set is introduced first. The software named Labelme (v4.5.13) [46] is used to label the buildings in the airborne optical image on the object level, cut the object image according to the label, and resize it to 224 × 224 pixels, as shown in Figure 5a. The software named CloudCompare (v2.11.3) [47] is used to cut the LiDAR point clouds in the object-level space, as shown in Figure 5c. The software named Meshlab (v3.3) [48] is used to segment the 3D models corresponding to the optical images and LiDAR point clouds in order to obtain the object-level model of a single building, as shown in Figure 5b.
In order to verify the effect of our object-level joint descriptor in cross-dimensional matching tasks, an object-level data set composed of building images and corresponding point clouds is needed. In addition, to learn how to extract the SDD accurately, a building model corresponding to cross-dimensional data is necessary. Therefore, a data set consisting of object-level optical images, object-level point clouds, and 3D models of the corresponding buildings is produced. The numbers of various types of data points in the data set are shown in Table 2.

2.4.1. Optical Images

The Institute of Geodesy and Photogrammetry, ETH Zurich, provides airborne optical images of the Zurich region that are published in the ISPRS data set [49]. These optical images cover an area nearby the center of Zurich (Switzerland) and the ETH Hoenggerberg. The region comprises the center of the quarter of Hoengg, including different residential areas with different types of buildings (flat roofs, hip roofs, etc.)
The Zurich Hoengg data set is based on aerial photographs collected over Zurich in 1995. The data set consists of four aerial images of the city of Zurich taken at an average image scale of ca. 1:5000. The 23 cm × 23 cm color photographs were scanned at 14 µm, yielding color images of approximately 840 Mb. Each image is approximately 16,500 × 16,400 pixels. The photography equipment was flown ca. 1050 above the ground with 65% forward and 45% sideward overlap. The camera was a Leica RC20. We annotated, segmented, and produced a data set containing only 1 building per image, a total of 11,094 images of 453 houses.

2.4.2. Point Clouds

Twice in 2014 and again from 2017 to 2018, the Federal Government conducted two high-resolution laser scans of the geographical area of the canton of Zurich (E/N Min:2669255/1223895; E/N Max:2716900/1283336[m]) using the airborne LiDAR system method (minimum point density of 5 Pkt/m2). The reference system, CH1903+_LV95, has a position accuracy of ±0.2 (m) and a height accuracy of ±0.1 (m). The Swiss SURFACE 3D product and spatial data set (GIS-ZH No.24) were created. We segmented the object-level buildings and saved individual buildings as point cloud files, yielding a total of 906 point cloud files of 453 buildings.

2.4.3. Three-Dimensional Model and SDF Ground Truth

The CityGML data of Zurich is available at “https://3D.bk.tudelft.nl/opendata/ (accessed on 24 January 2022)”, which includes the 3D models of buildings. After object-level splitting, SDF is extracted as the ground truth and input into the training network.

2.5. Evaluation Metrics

After extracting the SDD, PointNet [42] is used to classify the descriptors for cross-dimensional matching. In order to verify the precision of the SDD in presenting the 3D geometric information of the object, the EMD [50] is used to calculate the similarity between the SDD extracted from the 2D image and the SDD extracted from the 3D point cloud. EMD can intuitively describe the distance of the SDDs in the descriptor space. The smaller the EMD is, the more similar the SDDs are. Moreover, to verify the matching effect of our method, instance accuracy and category accuracy are used as metrics to evaluate the matching accuracy of the total sample and matching category accuracy of the building, respectively.

2.5.1. Instance Accuracy

Instance accuracy refers to the percentage of correctly matched data in the total data to be matched, as shown in (8):
I n s _ a c c = TP TP + FP × 100 %
where TP is the correct matching sample, and FP is the wrong matching sample.

2.5.2. Category Accuracy

Each building object is regarded as a class, and the category accuracy is the average accuracy of all the categories, as shown in (9):
C a t _ a c c = 1 N n = 1 N A n × 100 %
where An is the accuracy of each category, and N is the total number of categories.

2.6. The Training Platform and Parameter Settings

All experiments were carried out in the same environment. Training and testing of the network were conducted under Ubuntu 18.04. The hardware environment included a single Intel i7 9700 CPU, and the GPU acceleration used a single NVIDIA RTX 2080 Super with 8 GB of memory. In training, the batch size was 64, the Adam optimizer is used, the learning rate was set to 1 × 10−4, and the epoch was 1500. We conducted multiple sets of experiments under different hyperparameter setting conditions and judged the optimal hyperparameter settings according to the experimental results. The number of points sampled from the training models was 100,000, with a distance in the range of −0.3 to 0.3 (based on the normalized model).
In these experiments, the 2D images were divided into training, validation, and test data in a ratio of 6:2:2. The 3D point clouds scanned in 2014 were used for training and validation, and the 3D point clouds scanned in 2017–2018 were used for testing.

3. Experiment and Results

In this section, we verify the effect of SDD extraction and the accuracy of cross-dimensional matching, respectively, under different encoder combinations based on a large number of ablation experiments. We evaluate the accuracy of different encoder combinations’ SDDs by calculating the distance between descriptors in the descriptor space and evaluate its influence on the accuracy of cross-dimensional matching according to the sample accuracy and category accuracy. Specifically, the EMD is a description of the descriptor quality. The smaller the EMD is, the better the descriptor’s description of the target’s three-dimensional information will be. While the instance accuracy is the accuracy of all the data to be matched, the category accuracy is the average accuracy of all the categories (building ID). These are representations of the descriptor-matching ability.

3.1. Extraction Effects of Different Descriptors for Cross-Dimensional Matching

3.1.1. Extraction Effect under Different Encoder Combinations

The control variable method is used to intuitively obtain the influences of different encoder combinations on the matching results. Specifically, when performing the image-point-cloud-matching task, the image descriptors are extracted by the encoder combination “ResNet18 + JDEM-CBN-RES” and matched with the point cloud descriptors extracted by various encoder combinations. The same goes for the point-cloud-matching images. When performing the task of image point cloud matching, the descriptor of the point cloud is extracted by the encoder combination “PointNet + JDEM-CBN-RES”. In addition, since our EMD calculation method does not normalize the SDD in advance, the range of the results is not fixed, and its value is related to the size of the building model. Firstly, the superiority of the SDD over the SOPD is verified. The results are shown in Table 3 and Table 4.
From the above experimental results, it can be seen that the distance in the descriptor space between the extracted SDD and ground truth is closer than that of the SOPD. In other words, the SDD describes the 3D geometric invariance of the object more accurately. The following section verifies the influence of the proposed MAL on the SDD extraction effect.

3.1.2. Extraction Effect under Different Proposed Methods

In this experiment, the number in “JDEM-CBN-RES+MAL-0%” refers to the ratio of MAL to the original function multistage weight loss. The results are shown in Table 5.
From the experimental results, it can be seen that the addition of MAL improves the extraction effect of the SDD. Meanwhile, the setting of different weights for MAL has no obvious causal relationship.
Finally, we identified the value of the k of PointNet-k that is most suitable for our proposed data set. The results are shown in Table 6.
According to the above experimental results, it can be seen that the value of k influences the extraction effect of the SDD. Because the value of k determines the spatial coding range of PointNet-k, when k is small, the nearest neighbor points are too few in number to support PointNet-k in learning. This is why the quality of the SDD extracted by “PointNet-2 + JDEM-CBN-RES” and “PointNet-6 + JDEM-CBN-RES” does not improve. In conclusion, PointNet-k works best when k is 8. The change curve of the SDD distance with different ratios of MAL and different parameters k of PointNet-k is shown in Figure 6.

3.2. Cross-Dimensional Matching Accuracy under Different Encoder Combinations

In order to accurately evaluate the effectiveness of different encoder combinations and the proposed method for cross-dimensional matching tasks, we continued to use the sample accuracy and category accuracy to evaluate the cross-dimensional matching results. The experiment reported in this chapter also used the control variable method, and the default condition settings were the same as those in Section 3.1. In this experiment, the comparison between the best matching results based on the SDD and the SOPD matching results, which were divided into 2D images matching 3D point clouds and 3D point clouds matching 2D images, was conducted first. The results are shown in Table 7 and Table 8.
According to the experimental results, it can be seen that the instance accuracy is generally higher than the category accuracy, which indicates that there is an imbalance in the matching ability between categories. In other words, the matching accuracy of some buildings is significantly higher than that of other buildings. Furthermore, the experimental results show that our cross-dimensional matching method is superior to the baseline in terms of both instance accuracy and category accuracy.
Next, the ablation experiment was divided into three parts. The first part verified the superiority of the SDD compared with the SOPD. The results are shown in Table 9 and Table 10. The second part verified the effectiveness of MAL, and the results are shown in Table 11. Finally, the value of the encoder k that is most suitable for our proposed data set was identified, and the results are shown in Table 12.
Through the above experimental results, the same conclusion as that outlined Section 3.1 can be drawn: a higher matching accuracy can be obtained by extracting the SDD, as compared with the SOPD; MAL improves the matching accuracy; and the weight of MAL does not affect its functionality. Since the same category contains multiple building images taken from different angles, the quality of descriptors extracted from building images taken from different angles and the instance accuracy and category accuracy of cross-dimensional matching are improved after using MAL. Therefore, it can be proved that MAL improves the adaptability of the image encoder module to images taken from different angles.
It is evident in Figure 7 that with the increase in k, the matching accuracy of PointNet-k undergoes a process of decline–rise–decline. The SDD distance curve in Figure 7 can provide a similar result to that in Figure 6. This is because the value of k determines the spatial coding range of PointNet-k. When k is small, the nearest neighbor point is too small to support PointNet-k learning. When k is too large, the wrong nearest neighbor causes PointNet-k to learn the wrong spatial encoding. The point cloud density of the data determines the optimal value of k. In conclusion, it can be found that PointNet-k helps to improve the accuracy of cross-dimensional matching. For our data set, the best matching result can be obtained when k = 8.

4. Discussion

This section can be divided into two parts. The first part provides a discussion to further explore the specific advantages of our approach, the need for further research, and the potential applications of this research. The second part provides a discussion to explain the previous experimental results, which includes further elaboration of the experimental results, an analysis of the impact on the experimental data set, and future research directions.
Our method is a new method called JDEM which aims to extract a joint descriptor named SDD from cross-dimensional data. Through this method, the inherent modal differences between 2D images and 3D point cloud data are solved, and object-level matching relationships can accurately be established between different dimensional data. This research is of great practical value. For example, in the case of a UAV losing its control signal and positioning signal due to targeted interference, the use of the data matching method can help the UAV to complete self-positioning so that it can continue to complete the established task independently under offline conditions. However, due to the wide range of shooting angles and the great changes in light conditions in traditional image matching tasks, the image–image matching positioning method requires the storage of a large number of images of the UAV as a database. With the increase in the number of 2D images taken from different angles and lighting conditions in the same area in the database, the accuracy of positioning will also increase, but this will consume a lot of terminal storage space. The use of 3D point cloud data can solve this problem. Firstly, because the 3D point cloud data composed of coordinates do not have shooting angle and light condition problems, the image-point-cloud-matching method requires less storage space at the terminal than the image-matching positioning method. Secondly, the image-point-cloud-matching method is not affected by the quality or number of images in the database, which reduces the possibility of error matching caused by the data imbalance problem in the database. Therefore, our proposed method has good application value.
In this study, our method was compared with other methods for object-level cross-dimensional matching. Despite the better results achieved with our network, several issues remain to be addressed. Firstly, the joint descriptor proposed for our method is based on the three-dimensional features of the structure and does not contain the color information of the building. Therefore, it is easy to mismatch the faces of similar structures or the same building structure with different colors. There are many ways to obtain point clouds with color texture information [51]. If the color information in such point cloud data can be added to the descriptor extraction process, the above problems can effectively be alleviated. Moreover, a problem that must be addressed is the data imbalance. Due to the characteristics of deep learning, structures with more occurrences in the training set can achieve better learning results. In comparison, structures with lower occurrences in the data set are difficult to understand and accurately learn. The data of just over 400 buildings are not enough to fully support the generalization learning of the network. Therefore, more buildings are necessary for the data set.

5. Conclusions

This paper proposed a new object-level cross-dimensional building-matching method based on the SDD. The SDD is a descriptor extracted by the JDEM which uses the 3D geometric invariance of buildings to achieve the cross-dimensional matching of buildings. In addition, MAL was designed to improve the adaptability of the image encoder module for images taken at different angles. A 2D-3D-CDOBM data set was proposed to verify the effectiveness and robustness of our method. A large number of experimental results showed that our object-level descriptor extraction method achieves state-of-the-art outcomes. The effectiveness of our proposed module was also proven. Although this method marks apparent progress compared with the baseline, SDD can still only describe the 3D geometry of the object and needs a better resolution when handling different buildings with similar shapes. This is a challenging issue. Scenario-level cross-dimensional data matching may be an effective solution, which is the critical issue that we are committed to addressing.

Author Contributions

Conceptualization and funding acquisition, C.Z., Y.Y. and W.H.; methodology and writing of the original draft, W.W. and Y.Y.; processing and analysis of cross-dimensional data, visualization, and editing of the manuscript, N.S. and S.F.; ablation experiment and revision of the manuscript, Q.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62271159, No. 62071136, No. 62002083, No. 61971153); Heilongjiang Outstanding Youth Foundation (YQ2022F002); Heilongjiang Postdoctoral Foundation (LBH-Q20085 and LBH-Z20051); Fundamental Research Funds for the Central Universities Grant (3072022QBZ0805, 3072021CFT0801, and 3072022CF0808; and High-Resolution Earth Observation Major Project (grant No. 72-Y50G11-9001-22/23).

Data Availability Statement

The proposed data set will be available to download at ‘https://github.com/HEU-super-generalized-remote-sensing/Cross-dimensional-Object-level-Matching-Data-Set.git (accessed on 4 June 2023)’. For the code, please contact the corresponding author via email.

Acknowledgments

The authors express their thanks for the provision of the data sets by ISPRS and EuroSDR, released in conjunction with the ISPRS scientific initiative 2014 and 2015, led by ISPRS ICWG I/Vb.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, L.; Li, H.; Dai, Y. Efficient global 2D-3D matching for camera localization in a large-scale 3D map. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2372–2381. [Google Scholar]
  2. Sattler, T.; Leibe, B.; Kobbelt, L. Efficient & effective prioritized matching for large-scale image-based localization. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1744–1756. [Google Scholar] [PubMed]
  3. Song, Y.; Chen, X.; Wang, X.; Zhang, Y.; Li, J. Are large-scale 3-D models really necessary for accurate visual localization? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1637–1646. [Google Scholar]
  4. Kundu, J.N.; Rahul, M.V.; Ganeshan, A.; Babu, R.V. Object pose estimation from monocular image using multi-view keypoint correspondence. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; Springer: Cham, Switzerland, 2018; pp. 298–313. [Google Scholar]
  5. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Mur-Artal, R.; Montiel JM, M.; Tardos, J.D. ORB-SLAM: A versatile and accurate monocular SLAM system. IEEE Trans. Robot. 2015, 31, 1147–1163. [Google Scholar] [CrossRef] [Green Version]
  7. Jared, H.; Schonberger, J.L.; Dunn, E.; Frahm, J.M. Reconstructing the world in six days. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  8. Arandjelović, R.; Zisserman, A. DisLocation: Scalable descriptor distinctiveness for location recognition. In Proceedings of the Computer Vision—ACCV 2014: 12th Asian Conference on Computer Vision, Singapore, 1–5 November 2014; Revised Selected Papers, Part IV 12. Springer International Publishing: Cham, Switzerland, 2015; pp. 188–204. [Google Scholar]
  9. Cao, S.; Snavely, N. Graph-Based Discriminative Learning for Location Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 700–707. [Google Scholar]
  10. Johns, E.D.; Yang, G.Z. Pairwise probabilistic voting: Fast place recognition without RANSAC. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part II 13. Springer International Publishing: Cham, Switzerland, 2014; pp. 504–519. [Google Scholar]
  11. Jin Kim, H.; Dunn, E.; Frahm, J.M. Learned contextual feature reweighting for image geo-localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2136–2145. [Google Scholar]
  12. Gálvez-López, D.; Tardos, J.D. Bags of binary words for fast place recognition in image sequences. IEEE Trans. Robot. 2012, 28, 1188–1197. [Google Scholar] [CrossRef]
  13. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN architecture for weakly supervised place recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 5297–5307. [Google Scholar]
  14. Arandjelovic, R.; Zisserman, A. All about VLAD. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1578–1585. [Google Scholar]
  15. Song, Y.; Chen, X.; Wang, X.; Zhang, Y.; Li, J. 6-DOF image localization from massive geo-tagged reference images. IEEE Trans. Multimed. 2016, 18, 1542–1554. [Google Scholar] [CrossRef]
  16. Torii, A.; Taira, H.; Sivic, J.; Pollefeys, M.; Taira, H.; Okutomi, M.; Pajdla, T. Are large-scale 3D models really necessary for accurate visual localization? IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 814–829. [Google Scholar] [CrossRef] [Green Version]
  17. Sattler, T.; Weyand, T.; Leibe, B.; Kobbelt, L. Image Retrieval for Image-Based Localization Revisited. In Proceedings of the British Machine Vision Conference, Aberdeen, UK, 20–24 November 2012; Volume 1, p. 4. [Google Scholar]
  18. Díaz, R.; Fowlkes, C.C. Cluster-wise ratio tests for fast camera localization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 19–28. [Google Scholar]
  19. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  20. Salas-Moreno, R.F.; Newcombe, R.A.; Strasdat, H.; Kelly, P.H.J.; Davison, A.J. Slam++: Simultaneous localisation and mapping at the level of objects. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1352–1359. [Google Scholar]
  21. Agarwal, S.; Furukawa, Y.; Snavely, N.; Simon, I.; Curless, B.; Seitz, S.M.; Szeliski, R. Building rome in a day. Commun. ACM 2011, 54, 105–112. [Google Scholar] [CrossRef]
  22. Snavely, N.; Seitz, S.M.; Szeliski, R. Photo tourism: Exploring photo collections in 3D. In Proceedings of the ACM Siggraph 2006 Papers, Boston, MA, USA, 30 July–3 August 2006; pp. 835–846. [Google Scholar]
  23. Furukawa, Y.; Hernández, C. Multi-view stereo: A tutorial. Found. Trends Comput. Graph. Vis. 2015, 9, 1–148. [Google Scholar] [CrossRef] [Green Version]
  24. Ng, P.C.; Henikoff, S. SIFT: Predicting amino acid changes that affect protein function. Nucleic Acids Res. 2003, 31, 3812–3814. [Google Scholar] [CrossRef] [Green Version]
  25. Wang, B.; Chen, C.; Cui, Z.; Qin, J.; Lu, C.X.; Yu, Z.; Zhao, P.; Dong, Z.; Zhu, F.; Trigoni, N.; et al. P2-net: Joint description and detection of local features for pixel and point matching. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 16004–16013. [Google Scholar]
  26. Xing, X.; Cai, Y.; Lu, T.; Cai, S.; Yang, Y.; Wen, D. 3Dtnet: Learning local features using 2D and 3D cues. In Proceedings of the 2018 International Conference on 3D Vision (3DV), Verona, Italy, 5–8 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 435–443. [Google Scholar]
  27. Feng, M.; Hu, S.; Ang, M.H.; Gim Hee, L. 2D3D-matchnet: Learning to match keypoints across 2-D image and 3-D point cloud. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, BC, Canada, 20–24 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 4790–4796. [Google Scholar]
  28. Pham, Q.-H.; Uy, M.A.; Hua, B.-S.; Nguyen, D.T.; Roig, G.; Yeung, S.-K. Lcd: Learned cross-domain descriptors for 2D-3D matching. Proc. AAAI Conf. Artif. Intell. 2020, 34, 11856–11864. [Google Scholar] [CrossRef]
  29. Li, J.; Lee, G.H. DeepI2P: Image-to-point cloud registration via deep classification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15960–15969. [Google Scholar]
  30. Jeon, Y.; Seo, S.W. EFGHNet: A Versatile Image-to-Point Cloud Registration Network for Extreme Outdoor Environment. IEEE Robot. Autom. Lett. 2022, 7, 7511–7517. [Google Scholar] [CrossRef]
  31. Wang, Q.; Gu, Y.; Tuia, D. Discriminative Multiple Kernel Learning for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2016, 54, 3912–3927. [Google Scholar] [CrossRef]
  32. Dusmanu, M.; Rocco, I.; Pajdla, T.; Pollefeys, M.; Sivic, J.; Torii, A.; Sattler, T. D2-net: A trainable cnn for joint description and detection of local features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8092–8101. [Google Scholar]
  33. Luo, Z.; Zhou, L.; Bai, X.; Chen, H.; Zhang, J.; Yao, Y.; Li, S.; Fang, T.; Quan, L. Aslfeat: Learning local features of accurate shape and localization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6589–6598. [Google Scholar]
  34. Bai, X.; Luo, Z.; Zhou, L.; Fu, H.; Quan, L.; Tai, C.-L. D3feat: Joint learning of dense detection and description of 3D local features. In Proceedings of the IEEE/CVF Conference on Computer VISION and pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 6359–6367. [Google Scholar]
  35. Li, Y.; Su, H.; Qi, C.R.; Fish, N.; Cohen-Or, D.; Guibas, L.J. Joint embeddings of shapes and images via cnn image purification. ACM Trans. Graph. 2015, 34, 1–12. [Google Scholar] [CrossRef] [Green Version]
  36. Yan, Y.; Wang, W.; Su, N.; Wang, Z. Cross-Dimensional Object-Level Matching Method for Buildings in Airborne Optical Image and LiDAR Point Cloud. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  37. Wang, Q.; Gu, Y.; Yang, M.; Wang, C. Multi-attribute smooth graph convolutional network for multispectral points classification. Sci. China Technol. Sci. 2021, 64, 2509–2522. [Google Scholar] [CrossRef]
  38. Wang, Q.; Gu, Y. A discriminative tensor representation model for feature extraction and classification of multispectral LiDAR data. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1568–1586. [Google Scholar] [CrossRef]
  39. Park, J.J.; Florence, P.; Straub, J.; Newcombe, R.; Lovegrove, S. Deepsdf: Learning continuous signed distance functions for shape representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 165–174. [Google Scholar]
  40. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3D classification and segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 652–660. [Google Scholar]
  43. Dai, J.; Li, Y.; He, K.; Sun, J. R-fcn: Object detection via region-based fully convolutional networks. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; p. 29. [Google Scholar]
  44. Mescheder, L.; Oechsle, M.; Niemeyer, M.; Nowozin, S.; Geiger, A. Occupancy networks: Learning 3D reconstruction in function space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4460–4470. [Google Scholar]
  45. Zhao, C.; Zhang, C.; Yan, Y.; Su, N. Shape Reconstruction of Object-Level Building from Single Image Based on Implicit Representation Network. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  46. Russell, B.C.; Torralba, A.; Murphy, K.P.; Freeman, W.T. Labelme: A database and web-based tool for image annotation. Int. J. Comput. Vis. 2008, 77, 157–173. [Google Scholar] [CrossRef]
  47. CloudCompare. Available online: https://github.com/CloudCompare/CloudCompare (accessed on 30 May 2023).
  48. Cignoni, P.; Callieri, M.; Corsini, M.; Dellepiane, M.; Ganovelli, F.; Ranzuglia, G. Meshlab: An open-source mesh processing tool. In Proceedings of the Eurographics Italian Chapter Conference, Salerno, Italy, 2–4 July 2008; pp. 129–136. [Google Scholar]
  49. Nex, F.; Remondino, F.; Gerke, M.; Przybilla, H.-J.; Bäumker, M.; Zurhorst, A. ISPRS benchmark for multi-platform photogrammetry. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2015, 2, 135. [Google Scholar] [CrossRef] [Green Version]
  50. Rubner, Y.; Tomasi, C.; Guibas, L.J. The earth mover’s distance as a metric for image retrieval. Int. J. Comput. Vis. 2000, 40, 99–121. [Google Scholar] [CrossRef]
  51. Hu, Q.; Yang, B.; Khalid, S.; Xiao, W.; Trigoni, N.; Markham, A. Towards Semantic Segmentation of Urban-Scale 3D Point Clouds: A Dataset, Benchmarks and Challenges. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4975–4985. [Google Scholar]
Figure 1. Comparative diagram of pixel-level cross-dimensional matching and object-level cross-dimensional matching. Pixel-level cross-dimensional matching takes pixels in the image and points in the point cloud as the core for matching, while object-level cross-dimensional matching takes the building objects in the image and the building objects in the point cloud as the core for matching.
Figure 1. Comparative diagram of pixel-level cross-dimensional matching and object-level cross-dimensional matching. Pixel-level cross-dimensional matching takes pixels in the image and points in the point cloud as the core for matching, while object-level cross-dimensional matching takes the building objects in the image and the building objects in the point cloud as the core for matching.
Remotesensing 15 02974 g001
Figure 2. The diagram of the SDF and SOPF. The red line in (b,c) represents the boundary of the building in (a), green represents the interior of the boundary, while blue represents the exterior of the boundary, and the darker the color in different positions, the larger the descriptor value in that position.
Figure 2. The diagram of the SDF and SOPF. The red line in (b,c) represents the boundary of the building in (a), green represents the interior of the boundary, while blue represents the exterior of the boundary, and the darker the color in different positions, the larger the descriptor value in that position.
Remotesensing 15 02974 g002
Figure 3. Structural diagram of the cross-dimensional data matching method. P denotes the random sampling points. In the data input process, it is necessary to input images from different perspectives of the same building, using Multi-View Adaptive Loss (MAL) constraint training at the same time.
Figure 3. Structural diagram of the cross-dimensional data matching method. P denotes the random sampling points. In the data input process, it is necessary to input images from different perspectives of the same building, using Multi-View Adaptive Loss (MAL) constraint training at the same time.
Remotesensing 15 02974 g003
Figure 4. Structural diagram of different JDEMs.
Figure 4. Structural diagram of different JDEMs.
Remotesensing 15 02974 g004
Figure 5. A schematic diagram of the 2D-3D-CDOBM data set. (a) is 2D optical image, (b) is building model, and (c) is 3D point cloud.
Figure 5. A schematic diagram of the 2D-3D-CDOBM data set. (a) is 2D optical image, (b) is building model, and (c) is 3D point cloud.
Remotesensing 15 02974 g005
Figure 6. The change curve of the SDD distance with different ratios of MAL and different values of parameter k of PointNet-k.
Figure 6. The change curve of the SDD distance with different ratios of MAL and different values of parameter k of PointNet-k.
Remotesensing 15 02974 g006
Figure 7. The change curve of cross-dimensional matching accuracy with different ratios of MAL and different parameters k of PointNet-k.
Figure 7. The change curve of cross-dimensional matching accuracy with different ratios of MAL and different parameters k of PointNet-k.
Remotesensing 15 02974 g007
Table 1. Various encoder combinations.
Table 1. Various encoder combinations.
Image EncoderPoint Cloud Encoder
Part APart BPart APart B
ResNet18JDEM-CBN-RESPointNetJDEM-CBN-RES
ResNet34JDEM-CBN-NonePointNet-kJDEM-CBN-None
ResNet50JDEM-BN-RES JDEM-BN-RES
ResNet101JDEM-None-RES JDEM-None-RES
ResNet152
Joint Descriptor Extraction Module (JDEM).
Table 2. Details of our data set.
Table 2. Details of our data set.
Object-Level Optical ImageObject-Level Point CloudObject-Level 3D Model
11,094906453
Table 3. The EMD between descriptors extracted from 2D images and 3D point clouds under different encoder combinations.
Table 3. The EMD between descriptors extracted from 2D images and 3D point clouds under different encoder combinations.
Image EncoderSOPDSDD
Part APart B
ResNet18JDEM-CBN-RES74.608442.7909 (31.8175)
ResNet34JDEM-CBN-RES75.7076128.6485 (+52.9409)
ResNet50JDEM-CBN-RES77.066163.7674 (13.2987)
ResNet101JDEM-CBN-RES73.809045.9396 (27.8694)
ResNet152JDEM-CBN-RES76.768246.1051 (30.6631)
ResNet18JDEM-BN-RES77.965143.2792 (34.6859)
ResNet18JDEM-None-RES77.780243.2648 (34.5154)
ResNet18JDEM-CBN-None74.435843.1497 (31.2861)
In the table, the bold underline represents the best result among the comparison algorithms in the same column, and the content in parentheses shows the difference in evaluation indicators between the same row. If the Evaluation Metrics becomes smaller, it is displayed in green; if it becomes larger, it is displayed in red. Specifically, for the EMD, smaller values are better, while for the accuracy, larger values are better. The following tables are the same. Signed Distance Descriptor (SDD).
Table 4. The EMD between descriptors extracted from 3D point clouds and 2D images under different encoder combinations.
Table 4. The EMD between descriptors extracted from 3D point clouds and 2D images under different encoder combinations.
Point Cloud EncoderSOPDSDD
Part APart B
PointNetJDEM-CBN-RES74.608442.7909 (31.8175)
PointNetJDEM-BN-RES77.987246.2336 (31.7536)
PointNetJDEM-CBN-None75.431842.9513 (32.4805)
PointNetJDEM-None-RES76.375746.9127 (29.4630)
Table 5. The EMD between descriptors extracted from 2D images matching 3D point clouds using different ratios of MAL.
Table 5. The EMD between descriptors extracted from 2D images matching 3D point clouds using different ratios of MAL.
Image EncoderSDD
Part APart B
ResNet18JDEM-CBN-RES+MAL-0%42.7909
ResNet18JDEM-CBN-RES+MAL-5%42.1315 (0.6594)
ResNet18JDEM-CBN-RES+MAL-10%41.9423 (0.8486)
ResNet18JDEM-CBN-RES+MAL-20%41.5696 (1.2213)
ResNet18JDEM-CBN-RES+MAL-30%42.3517 (0.4392)
ResNet18JDEM-CBN-RES+MAL-40%41.7006 (1.0903)
ResNet18JDEM-CBN-RES+MAL-50%42.1304 (0.6605)
ResNet18JDEM-CBN-RES+MAL-60%42.6350 (0.1559)
ResNet18JDEM-CBN-RES+MAL-70%41.9715 (0.8194)
ResNet18JDEM-CBN-RES+MAL-80%42.0579 (0.7330)
ResNet18JDEM-CBN-RES+MAL-90%42.1455 (0.6454)
ResNet18JDEM-CBN-RES+MAL-100%41.5711 (1.2198)
Table 6. The EMD between descriptors extracted from 3D point clouds matching 2D images with different k of PointNet-k.
Table 6. The EMD between descriptors extracted from 3D point clouds matching 2D images with different k of PointNet-k.
Point Cloud EncoderSDD
Part APart B
PointNetJDEM-CBN-RES42.7909
PointNet-2JDEM-CBN-RES44.4586 (+1.6677)
PointNet-4JDEM-CBN-RES41.9119 (0.8790)
PointNet-6JDEM-CBN-RES42.9444 (+0.1535)
PointNet-8JDEM-CBN-RES41.1428 (1.6481)
PointNet-10JDEM-CBN-RES41.7385 (1.0524)
PointNet-12JDEM-CBN-RES42.2998 (0.4911)
Table 7. The best accuracy for 2D images matching 3D point clouds.
Table 7. The best accuracy for 2D images matching 3D point clouds.
SOPDSDD
Instance accCategory accInstance accCategory acc
72.2582%69.6282%76.8301% (+4.5719%)73.2332% (+3.605%)
Table 8. The best accuracy for 3D point clouds matching 2D images.
Table 8. The best accuracy for 3D point clouds matching 2D images.
SOPDSDD
Instance accCategory accInstance accCategory acc
75.2193%74.9485%79.8559% (+4.6366%)79.6946% (+4.7461%)
Table 9. The accuracy of 2D images matching 3D point clouds under different encoder combinations.
Table 9. The accuracy of 2D images matching 3D point clouds under different encoder combinations.
Image EncoderSOPDSDD
Part APart BInstance accCategory accInstance accCategory acc
ResNet18JDEM-CBN-RES72.2582%69.6282%75.5470% (+3.2888%)72.4163% (+2.7881%)
ResNet34JDEM-CBN-RES72.0759%69.4502%71.7247% (0.3512%)68.8125% (0.6377%)
ResNet50JDEM-CBN-RES72.0556%69.4846%74.5746% (+2.5190%)71.5551% (+2.0705%)
ResNet101JDEM-CBN-RES73.4063%71.1805%76.5532% (+3.1469%) 73.3839% (+2.2034%)
ResNet152JDEM-CBN-RES71.8395%68.8852%75.1351% (+3.2956%)72.2858% (+3.4006%)
ResNet18JDEM-BN-RES66.8828%63.0065%73.0821% (+6.1993%)68.3356% (+5.3291%)
ResNet18JDEM-None-RES68.4292%65.3175%73.3793% (+4.9501%)69.7284% (+4.4109%)
ResNet18JDEM-CBN-None71.1642%68.3775%76.2763% (+5.1121%)72.5993% (+4.2218%)
Table 10. The accuracy of 3D point clouds matching 2D images under different encoder combinations.
Table 10. The accuracy of 3D point clouds matching 2D images under different encoder combinations.
Image EncoderSOPDSDD
Part APart BInstance accCategory accInstance accCategory acc
PointNetJDEM-CBN-RES71.8985%71.6691%72.3997% (+0.5012%)72.2247% (+0.5556%)
PointNetJDEM-BN-RES75.2193%74.9485%76.4098% (+1.1905%)76.2344% (+1.2859%)
PointNetJDEM-None-RES71.2719%71.3812%73.8722% (+2.6003%)73.6228% (+2.2416%)
PointNetJDEM-CBN-None71.2719%71.3812%73.8722% (+2.6003%)73.6228% (+2.2416%)
Table 11. The accuracy of 2D images matching 3D point clouds using different ratios of MAL.
Table 11. The accuracy of 2D images matching 3D point clouds using different ratios of MAL.
Image EncoderSDD
Part APart BInstance accCategory acc
ResNet18JDEM-CBN-RES+MAL-0%75.5470%72.4163%
ResNet18JDEM-CBN-RES+MAL-5%76.1750% (+0.6208%)72.6685% (+0.2522%)
ResNet18JDEM-CBN-RES+MAL-10%76.3979% (+0.8509%)72.6124% (+0.1961%)
ResNet18JDEM-CBN-RES+MAL-20%75.9184% (+0.3714%)72.6139% (+0.1976%)
ResNet18JDEM-CBN-RES+MAL-30%76.0940% (+0.5470%)72.4914% (+0.0751%)
ResNet18JDEM-CBN-RES+MAL-40%76.2966% (+0.7496%)73.4532% (+1.0369%)
ResNet18JDEM-CBN-RES+MAL-50%76.3033% (+0.7563%)72.9010% (+0.4847%)
ResNet18JDEM-CBN-RES+MAL-60%76.3911% (+0.8441%)73.1598% (+0.7435%)
ResNet18JDEM-CBN-RES+MAL-70%75.7361% (+0.1891%)71.9418% (−0.4745%)
ResNet18JDEM-CBN-RES+MAL-80%76.6140% (+1.0670%)73.0335% (+0.6172%)
ResNet18JDEM-CBN-RES+MAL-90%76.7896% (+1.2426%)73.4325% (+1.0162%)
ResNet18JDEM-CBN-RES+MAL-100%76.8301% (+1.2831%) 73.2332% (+0.8169%)
Table 12. The accuracy of 3D point clouds matching 2D images using different k of PointNet-k.
Table 12. The accuracy of 3D point clouds matching 2D images using different k of PointNet-k.
Point Cloud EncoderSDD
Part APart BInstance accCategory acc
PointNetJDEM-CBN-RES76.4098%76.2344%
PointNet-2JDEM-CBN-RES73.1830% (−3.2268%)73.0390% (−3.1954%)
PointNet-4JDEM-CBN-RES76.9110% (+0.5012%)76.6826% (+0.4482%)
PointNet-6JDEM-CBN-RES77.8822% (+1.4724%)77.6392% (+1.4048%)
PointNet-8JDEM-CBN-RES79.8559% (+3.4461%)79.6946% (+3.4602%)
PointNet-10JDEM-CBN-RES78.3835% (+1.9737%)78.2929% (+2.0585%)
PointNet-12JDEM-CBN-RES76.3784% (−0.0314%)76.1185% (−0.1159%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, C.; Wang, W.; Yan, Y.; Su, N.; Feng, S.; Hou, W.; Xia, Q. A Novel Object-Level Building-Matching Method across 2D Images and 3D Point Clouds Based on the Signed Distance Descriptor (SDD). Remote Sens. 2023, 15, 2974. https://doi.org/10.3390/rs15122974

AMA Style

Zhao C, Wang W, Yan Y, Su N, Feng S, Hou W, Xia Q. A Novel Object-Level Building-Matching Method across 2D Images and 3D Point Clouds Based on the Signed Distance Descriptor (SDD). Remote Sensing. 2023; 15(12):2974. https://doi.org/10.3390/rs15122974

Chicago/Turabian Style

Zhao, Chunhui, Wenxuan Wang, Yiming Yan, Nan Su, Shou Feng, Wei Hou, and Qingyu Xia. 2023. "A Novel Object-Level Building-Matching Method across 2D Images and 3D Point Clouds Based on the Signed Distance Descriptor (SDD)" Remote Sensing 15, no. 12: 2974. https://doi.org/10.3390/rs15122974

APA Style

Zhao, C., Wang, W., Yan, Y., Su, N., Feng, S., Hou, W., & Xia, Q. (2023). A Novel Object-Level Building-Matching Method across 2D Images and 3D Point Clouds Based on the Signed Distance Descriptor (SDD). Remote Sensing, 15(12), 2974. https://doi.org/10.3390/rs15122974

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop