Next Article in Journal
Toward Privacy-Preserving Directly Contactable Symptom-Matching Scheme for IoT Devices
Next Article in Special Issue
Transform-Based Feature Map Compression Method for Video Coding for Machines (VCM)
Previous Article in Journal
Secure Multi-Level Privacy-Protection Scheme for Securing Private Data over 5G-Enabled Hybrid Cloud IoT Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optical and SAR Image Registration Based on Multi-Scale Orientated Map of Phase Congruency

1
Department of Electronic and Optical Engineering, Army Engineering University, Shijiazhuang 050003, China
2
Institute of Intelligent Electronic Information System, School of Integrated Circuits and Electronics, Beijing University of Technology, Beijing 100081, China
3
32398 Units of PLA, Beijing 100192, China
4
College of Electronic Engineering, National University of Defense Technology, Hefei 230037, China
*
Authors to whom correspondence should be addressed.
Electronics 2023, 12(7), 1635; https://doi.org/10.3390/electronics12071635
Submission received: 20 February 2023 / Revised: 15 March 2023 / Accepted: 22 March 2023 / Published: 30 March 2023
(This article belongs to the Special Issue Image/Video Processing and Encoding for Contemporary Applications)

Abstract

:
Optical and Synthetic Aperture Radar (SAR) images are highly complementary, and their registrations are a fundamental task for other remote sensing applications. Traditional feature-matching algorithms fail to solve the significant nonlinear radiation difference (NRD) caused by different sensors. To address this problem, a robust registration algorithm with the multi-scale orientated map of phase congruency (MSPCO) is proposed. First, a nonlinear diffusion scale space is established to obtain the scale invariance of feature points. Compared with the linear Gaussian scale space, the nonlinear diffusion scale space can better preserve the edge and texture information. Second, to ensure the quantity and repeatability of features, corner points and edge points are detected on the moment map of phase congruency, respectively, which is the foundation to the next feature matching. Third, the MSPCO descriptor is constructed via the orientation of phase congruency (PCO). PCO is highly robust to NRD, and the different scales of PCOs enhance the robustness of the descriptor. Finally, a feature-matching strategy based on an effective scale ratio is proposed, which reduces the number of comparisons among features and improves computational efficiency. The experimental results show that the proposed method is better than the existing feature-based methods in terms of the number of correct matches and registration accuracy. The registration accuracy is only inferior to that of the most advanced template matching method, and the accuracy difference is within 0.3 pixels, which fully demonstrates the robustness and accuracy of the proposed method in optical and SAR image registration.

1. Introduction

With high efficiency and a wide range, remote sensing has become one of the effective means for people to observe, describe, and analyze the Earth’s surface. The maturity of sensor technology provides a variety of ways for Earth’s observation, and the types of remote sensing images that people can obtain are increasingly diversified. Compared with a single sensor, multiple sensors can observe different features of the Earth’s surface. Combining them for analysis is helpful to obtain a more accurate understanding of the ground scene [1]. In various types of remote sensing images, optical and SAR images have evident complementarity. The optical imaging system is the earliest developed, and the technology is relatively mature. Optical images conform to people’s observation habits and have good interpretability. However, optical images are easily affected by bad weather, and accurate object information cannot be obtained in weather conditions such as clouds, rain, fog, and haze. SAR is an active microwave imaging system, which can observe the Earth in all-weather, all-day, and penetrate the clouds. It is sensitive to man-made targets, especially metal targets, but its imaging quality is poor, and it has serious speckle noise. For this reason, it is meaningful to conduct optical and SAR images simultaneously in image mosaics, image fusions [2], change detections [3], 3D reconstructions [4], navigations [5], and other applications.
Image registration is the process of overlaying two or more images of the same scene taken at different times, from different viewpoints, and/or by different sensors [6]. It is a necessary step for many remote sensing analysis tasks. Owing to the huge difference between them, their registration must solve the following problems:
(1)
Geometric differences: The imaging mode of the optical sensor is center projection, while SAR’s is side-looking. Different imaging modes will produce large geometric differences, and topographic relief will lead to serious geometric distortion between optical and SAR images, especially when the image contains mountains;
(2)
Radiation differences: An optical sensor passively collects the sunlight reflected from ground objects, while a SAR sensor actively transmits a microwave with a longer wavelength and collects the backscattered energy from ground objects, which is greatly affected by a surface’s roughness, slope, and complex dielectric constant;
(3)
Speckle noise: As a coherent imaging system, the SAR system will inevitably produce speckle noise, which will change the real gray information of the image and lead to errors in gray statistics.
To address the above issues, scholars have conducted a lot of research on the registration of optical and SAR images. Concretely, the registration approaches can be classified into two categories, namely, intensity based and feature based [7]. Intensity-based methods include mutual-information-based methods [8], correlation-based methods [9], and frequency-domain-based methods [10]. Mutual-information-based methods cannot register images from different sources [11]. It is sensitive to radiation differences. Since the image intensity and gradient information are sensitive to NRD, some scholars have studied the structural features of images to enhance the robustness of registration [12]. Some of them used local self-similarity (LSS) [13] to obtain structural features. Ye et al. [14] proposed a feature descriptor of dense local self-similarity (DLSS), defined a new similarity measure, and achieved a better registration effect. Xiong et al. [15] proposed rank-based (RLSS) features on the basis of DLSS. In addition, some researchers used phase congruency to obtain the structural features. Ye et al. [12] built a histogram of oriented phase congruency (HOPC) by replacing gradient information with phase congruency information, which improved the registration accuracy of multimodal images. Ye et al. [16] designed a robust feature descriptor based on channel features of oriented gradient (CFOG) and provided a general framework for template matching. Xiang et al. [17] mapped the image features of a 3D space and used the phase correlation method to register image pairs. Fan et al. [18] designed the phase congruency structural descriptor (PCSD) by combining LSS and phase congruency and achieved good registration results. Xiang et al. [19] improved the phase congruency model. In conclusion, the application of structural features improves the registration accuracy of template matching method. However, these methods are sensitive to rotation and scale change [20].
Feature-based methods are more suitable to deal with the local geometric distortion and radiation differences between multimodal images. First, feature-based methods extract the salient features in the image, including points, lines, and regions. Then, a reliable matching pair is established according to the distance of the feature descriptors. Finally, the parameters of the transformation model can be calculated. Scale-invariant feature transform (SIFT) [21] is the most classical method, which has scale and rotation invariance. However, it cannot register images with great differences, because it cannot solve radiation differences. SIFT has been improved by numerous scholars. FAN et al. [22] extracted feature points from the second layer to prevent the influence of from the speckle noise, which improved the registration effect of SIFT. Xu et al. [23] proposed the ILS-SIFT algorithm and introduced the iterative idea into the SIFT algorithm to realize image registration. Inspired by the SIFT algorithm, Ma et al. [24] designed the PSO-SIFT algorithm, which solved the radiation differences through a new gradient definition. Zhang et al. [25] used the Canny algorithm to extract image edges and remove edge points from candidate key points. According to the characteristics of the image pairs, Xiang [26] proposed the OS-SIFT algorithm, which used different operators to calculate the image gradient. Based on OS-SIFT, Zhang et al. [27] improved the ratio of exponential weighted average operator and adopted a new matching algorithm with cascaded sample consensus. These improved SIFT algorithms have a limited improvement on registration effects because they all build feature descriptors with gradient information, and gradient information is very susceptible to radiation differences [16].
In recent years, feature-based methods used structural features rather than gradient information, and structural features are more robust than radiation differences. Paul [28] proposed a new registration algorithm based on structure descriptors. Based on the maximum index map of phase congruency, Li et al. [29] designed the radiation-variation insensitive feature transform (RIFT) algorithm. Wang et al. [30] constructed descriptors by combining the orientation information of phase congruency with the map of maximum index, which improved the robustness of descriptors. Based on the improved maximal self-similarity (IMSD) feature detector and the oriented self-similarity (OSS) feature descriptor, Xiong et al. [31] registered image pairs successfully. Then, an optimized offset mean filtering (OMF) method was proposed to extract adjacent self-similarity (ASS) features, which improved registration efficiency and robustness [32]. The above methods have obtained satisfactory results for NRD, but they would be affected by the speckle noise in different degrees.
To address the above issues, we propose a novel registration algorithm based on nonlinear diffusion scale space (NDSS) and MSPCO. First, NDSS is established to obtain the scale invariance of features. Compared with the linear Gaussian scale space (GSS), NDSS can preserve the texture information, suppress the speckle noise, and provide more accurate positioning of features. Second, the MSPCO descriptors are constructed based on PCO with different scales. The descriptor consists of four parts, which are obtained from four different scales’ PCOs. PCO with a smaller scale describes the texture details of the image but is easily influenced by noise, while the PCO with a larger scale is just the opposite. The combination of the two effectively improves the registration robustness. Finally, a feature-matching strategy based on effective scale ratio (ESR) is proposed, which reduces the number of comparisons between features and improves computational efficiency.
Here, is the main work in this paper:
(1)
A novel MSPCO descriptor is constructed based on PCO with different scales. The descriptor contains four different scales of phase congruency information. PCO with a smaller scale ensures the accuracy of registration, and PCO with a larger scale can effectively suppress the speckle noise. The MSPCO descriptor improves the stability of registration;
(2)
The concept of ESR is proposed between two NDSSs. The application of the ESR matching strategy reduces the repeated and invalid comparisons of features and improves feature matching efficiency;
(3)
The MSPCO descriptor is robust in scale, rotation, and radiation differences. It can register any two images without limitations, in theory.
We organize this paper in the following parts: Section 2 introduces the proposed methods, including the establishment of the MSPCO descriptor and the detailed description of matching strategies based on ESR. The experimental verification of the MSPCO method and its comparison with other algorithms are given in Section 3. Section 4 is the conclusion.

2. Methodology

In Section 2, we proposed the MSPCO algorithm, which is robust to rotation, scale, and NRD between images. First, NDSSs are established for image pairs to detect feature points at each scale. Next, by performing FAST feature detection on the maximum and minimum moment map of phase congruency, respectively, edge and corner features with stable, reliable, and high repeatability are obtained. Then, based on the different scales of PCOs, a novel MSPCO descriptor is established, which is similar to the descriptor construction method in the SIFT algorithm. Finally, we use ESR to optimize the matching process and obtain correct matches by performing the nearest neighbor distance ratio (NNDR) matching strategy and the fast sample consensus (FSC) algorithm [32]. The flow chart of the MSPCO algorithm is shown in Figure 1.

2.1. Establishment of NDSS

The classical SIFT algorithm uses the Gaussian kernel to establish linear scale space, but the image becomes blurred after Gaussian filtering, thus resulting in inaccurate feature location and feature descriptors [33]. To make up for the deficiencies of GSS, we established NDSS. The nonlinear diffusion equation can be defined as [34]:
L ( x , y ) t = d i v ( c ( x , y , t ) L )
c ( x , y , t ) = g ( | L ( x , y , t ) | )
g = { 1 | L | 2 0 1 exp ( 3.315 ( | L | / k ) 8 ) | L | 2 0
where L is the input image, t represents the scale parameter, L is the gradient value of the image, c ( x , y , t ) represents the diffusion function, g is the conductivity function, and k represents a constant related to the diffusion speed. The smaller the k is, the more the edge information is retained.
Since the partial differential equation in Equation (1) has no analytical solution, the traditional numerical difference method cannot be used for the iterative solution, and the convergence speed is slow. To improve the convergence speed, the KAZE algorithm uses Additive Operator Splitting (AOS) [35] to solve the equation. Thus, the solution of the nonlinear diffusion equation is:
L k + 1 = [ I τ l = 1 m A l ( L k ) ] 1 L k
where τ is the time step, I denotes the identity matrix, l represents the direction, and A l is the derivative along the l th direction.
Different from the construction of GSS in SIFT, this paper only takes the first group, and there is no down sampling, i.e., all images in this group have the same size. The scale of each layer can be calculated by the initial scale σ 0 and the scale ratio k :
σ i = σ 0 k i , k = 2 1 / 3 , i [ 0 , 1 , , N 1 ]
where N is the total number of scale layers. We set σ 0 = 1.6 , N = 6 , and k = 2 1 / 3 .
Let I o ( x , y ) and I s ( x , y ) represent an optical image and a SAR image, respectively. The NDSS of the two is expressed as N D S S o p t = { I o 1 , I o 2 , I o 3 , I o 4 , I o 5 , I o 6 } and N D S S s a r = { I s 1 , I s 2 , I s 3 , I s 4 , I s 5 , I s 6 } . In this section, a brief comparison of NDSS and GSS is given. Comparison results are shown in Figure 2.
As we can see in Figure 2d,e, the difference between NDSS and GSS is apparent. With the increase in σ i , GSS becomes fuzzier, and detailed information is seriously lost. For NDSS, with the increase in σ i , the image has a relatively low degree of blur, and the image edges and details are well preserved. The experimental results fully verify the advantages of NDSS.

2.2. Feature Detection

Feature detection is an important step in image registration. The repeatability of features directly determines the number of correct matches. However, due to the serious NRD between optical and SAR images and the noise in SAR images, the effect of directly extracting features from SAR images is poor. It is difficult to extract the correspondence directly. Although the NRD between optical and SAR images is evident, their structural features have high similarity, and higher repeatability can be obtained by extracting corner points from the structural features. The phase information can preserve the structural features of the image and is invariant to image contrast, illuminance, scale, and rotation changes.
Oppenheim et al. [36] found that the phase information of the image can retain the main contour features. Then, Morrone and Owens [37] pioneered the phase congruency theory that pointed out that human access to image information mainly relied on the phase information and features always appeared in the place where the Fourier component was superimposed the most. Different from gradient-based feature detection, phase congruency is a frequency-domain feature detection method. Kovesi [38] extended the phase congruency theory and provided the Log Gabor Filter (LGF). According to RIFT [29], we can obtain the maximum and minimum moment map, marked as M Ψ and m Ψ . M Ψ and m Ψ reflect the edge and corner features of the image, respectively.
We extract edge features and corner features on M Ψ and m Ψ through the FAST algorithm [29]. The process of feature detection is shown in Figure 3.
The features in Figure 3 are evenly distributed, where the yellow points are the edge points detected on M Ψ , the red points are the corner points detected on m Ψ , and the green points represent the common part of both.

2.3. Feature Description

It is necessary to design a descriptor for the detection of features. Descriptor construction follows two principles: The first is robustness, i.e., the descriptor can overcome the influence of noise, local geometric difference, and NRD; and the second is uniqueness. The uniqueness of descriptors can reduce the mismatching rate of features. Due to the NRD between image pairs, the gradient information at the same location will produce significant differences. Therefore, gradient-based descriptors are often unreliable and have poor registration performance for multimodal images. In this paper, we construct feature descriptors based on multi-scale PCOs. The MSPCO descriptor is robust to the NRD.

2.3.1. PCO

Phase congruency has good robustness to NRD and is widely used for image registration. However, in previous algorithms, the amplitude of phase congruency is sensitive to speckle noise, resulting in large differences in local descriptors between image pairs, which degrades the accuracy of image registration. Compared with the amplitude, the orientation information represents the direction with the most drastic feature changes, which has stronger stability and is more suitable for descriptor construction.
We calculate the orientation of phase congruency by L G F s , o o d d ( x , y ) , and then the PCO of the image is obtained. PCO describes the directional changes of features at each pixel and can effectively resist NRD between images. To improve the robustness of registration, this paper constructs the multi-scale PCOs, as shown in Figure 4. Figure 5 is a schematic diagram of PCO construction at a single scale.
First, the images pairs are convolved with the odd symmetric filter of LGF, respectively, to obtain the PCO with s scales and o directions. At each scale, the projection of six directions on the x axis and y axis is calculated to obtain the horizontal component O x and the vertical component O y , which is calculated as follows:
O x = o ( o s , o ( θ o ) cos ( θ o ) )
O y = o ( o s , o ( θ o ) sin ( θ o ) )
where θ o represents the angle of the filter, and o s , o ( θ o ) is the convolution result of the odd filter at the direction o .
Then, the orientation of phase congruency O p c can be obtained by the arctangent function:
O p c = arctan ( O y , O x )
Due to the phenomenon of gradient inversion, the value O p c should be constrained, i.e.,
O p c = { O p c , O p c [ 0 ° , 180 ° ] O p c 180 ° , O p c [ 180 ° , 360 ° ]

2.3.2. MSPCO Feature Descriptor

The MSPCO descriptor is constructed based on the PCO. Figure 6 shows the construction process of the MSPCO descriptor.
From Figure 6, we can see that MSPCO is calculated based on a grid of sub-regions. First, we select a local image region with m × m pixels centered at each feature point. We divide the local region into n p × n p sub-regions. In this paper, we set m = 96 ( σ = σ 0 ) and n p = 4 . Next, for each sub-region, the PCO histogram is calculated separately. The range of PCO is [ 0 , 180 ° ] . According to the angle size, the PCO is evenly divided into six equal parts, each of which is 30°. Then, the PCO is composed of numbers { 0 , 1 , 2 , 3 , 4 , 5 } , thus the histogram dimension of each sub-region is n o = 6 . Finally, the descriptor is obtained by concatenating the histograms of sub-regions. That is MSPCO = [ S 1 , S 2 , S 3 , S 4 ] , where S i = [ S i H 1 , S i H 1 , , S i H 16 ] .

2.4. Feature Matching

After obtaining the features and their descriptors, the traditional registration method is to match the features from scales one by one, which needs a vast number of calculations. Only when the two image scales are at the same level or close to each other, an accurate and effective correspondence can be obtained. Therefore, this paper proposes a feature-matching strategy for ESR. The effective scale combination is the scale ratio that can be covered by the scale space. For N D S S o p t = { I o 1 , I o 2 , I o 3 , I o 4 , I o 5 , I o 6 } and N D S S s a r = { I s 1 , I s 2 , I s 3 , I s 4 , I s 5 , I s 6 } , E S R = { k 5 , k 4 , k 3 , k 2 , k 1 , 1 , k , k 2 , k 3 , k 4 , k 5 } . Previous algorithms compare features from different scales one by one, i.e., N × N = 36 comparisons are required. This paper proposes the concept of effective scale combination and reduces the number of calculations to 2 N 1 = 11 . Specifically, we extract feature points from two groups of NDSSs. Each group of NDSS contains six layers. Assuming that 500 feature points are extracted from each layer, 3000 feature points need to be extracted, respectively. According to the traditional feature matching algorithm, 9 × 10 6 ( 3000 × 3000 ) Euclidean distance calculations are required. This paper proposes an ESR matching strategy, which reduces the number of Euclidean distance calculations to 2.75 × 10 6 ( 500 × 500 ) . Its calculation quantity is reduced by 69%.
To gain illumination invariance, we, finally, normalize the feature vector. We can obtain initial matches through the NNDR matching strategy. Then, correct matches are selected by the FSC algorithm.

3. Experimental Results

In this section, we demonstrate the effectiveness of the MSPCO algorithm in optical and SAR image registration through verification experiments and contrast experiments. First, the experimental data and evaluation criteria used in this paper are given. The experimental data include synthetic image data and real image data, and we use the number of correct matches (NCM) and root mean square error (RMSE) as the evaluation metrics. Then, through validation experiments, the robustness of the descriptor and the ESR strategy are verified by using synthetic image data. Finally, we compare the MSPCO algorithm with five state-of-the-art algorithms, i.e., SIFT, SAR-SIFT, OS-SIFT, HOPC, and RIFT.

3.1. Datasets and Evaluation Criteria

The verification experiment was conducted on synthetic image pairs, which were generated from multispectral data. In general, different bands of the multispectral image has evident radiation differences. We regarded the R-band as the optical image and added multiplicative speckles in the B-band to generate the SAR image. Each band of the multispectral image had strict position consistencies, which were very useful for experimental evaluation. The synthetic image pairs are shown in Figure 7.
Six image pairs were obtained to perform the contrast experiment. Optical images mainly came from Google Earth, the Landsat satellite, and Google Maps. SAR was mainly provided by the GF 3 and TerraSAR-X satellites. The image covers different scenes, including urban, rural, and suburban areas. Table 1 describes the datasets that are shown in Figure 8. The registration performance of the MSPCO algorithm is evaluated quantitatively by two metrics: NCM and RMSE.

3.2. Robustness and Uniqueness Analysis of MSPCO

In this section, a synthetic image pair is used to verify the robustness and uniqueness of the MSPCO. As shown in Figure 7, there is a large radiation difference and even a gradient inversion phenomenon between image pairs. Moreover, there is evident multiplicative speckle noise in SAR images. All of these differences create significant challenges for the robustness of image registration. We compared SIFT, SAR-SIFT [39], OS-SIFT, HOPC, RIFT, and MSPCO and used Euclidean distance to quantify the differences between descriptors. Ten pairs of matching points were randomly selected, and the Euclidean distance d c between matching points was calculated for different descriptors to verify the robustness. The average Euclidean distance d n between one and the other nine non-corresponding points was calculated to verify the uniqueness. Repeat 10 times to increase the reliability of the experiment.
After normalization, the range of the Euclidean distance between descriptors is [ 0 , 2 ] . The larger the Euclidean distance, the lower the descriptor similarity. The smaller the distance, the higher the descriptor similarity. For a pair of correct matches, a small Euclidean distance value means the robustness of the matching. For two mismatched points, a large Euclidean distance indicates that the descriptor is more unique. It can be seen from Figure 9a that the MSPCO descriptor obtains the minimum Euclidean distance in the robustness analysis. SIFT and SAR-SIFT build feature descriptors based on gradient information, which are sensitive to radiation differences. HOPC, RIFT, and MSPCO are all established based on phase congruency and obtain a smaller d c , which indicates that phase congruency is robust to NRD. Among them, HOPC is established based on the amplitude of phase congruency. Li et al. proved that phase congruency is not suitable for feature description directly, thus the robustness of HOPC is slightly worse than RIFT and MSPCO. The RIFT algorithm built the maximum index map and achieved good radiation invariance. We built the MSPCO descriptor based on the multi-scale PCOs. Since the orientation information was more robust than the amplitude information, the best radiation invariance was obtained in this paper. As can be seen from Figure 9b, all descriptors have achieved good uniqueness.

3.3. Validation of ESR

The establishment of scale space is a common method to solve scale invariance, but it is followed by a huge amount of computation. Previous algorithms, which compare descriptors from different scales one by one, not only have low computational efficiency but also tend to produce mismatching, which will affect the registration results. In this paper, the ESR strategy is proposed, which can improve the registration speed and reduce the mismatching rate.
In the process of feature point description, the selection of a local support region is very important, and the same local support region can increase the similarity of descriptors. In the SIFT algorithm, the radius of the local support region is:
r a d i u s = 3 σ o c t × 2 × ( D + 1 ) + 1 2
Here, σ o c t is the scale of the scale space where the feature point is located ( D = 4 ). It is not difficult to see that the local support region becomes larger with the increase in σ o c t . In fact, images with larger scales usually have lower resolution, and the unit pixel contains more true ground distance. Therefore, in this paper, the local support region decreases with the increase in σ o c t , i.e.,
m = 96 × σ 0 σ i
Synthetic image pairs were selected to verify the matching strategy of ESR, whose image size was 300 × 300 . The synthetic SAR image was upsampled to 600 × 600 . The image was registered according to the ESR strategy described in Section 2.4. The real transformation was used to identify the correct matches by considering a threshold value of two pixels [40].
With the change in the scale σ , the local support regions of features also changed. Only when the local support regions of the features had sufficient overlap area could their descriptor achieve high similarity. As shown in Figure 10, the most matches were obtained when E S R = k 3 = 2 . This is in line with our expectations because the ratio of the ground sampling distance between the optical and SAR images was two. This experiment shows that correct matches can only be obtained when the local support regions of features are the same or close. The registration strategy of ESR proposed in this paper can efficiently find the true scale ratio of the input image pairs, which not only reduces meaningless repeated comparisons between descriptors but also reduces the cross-scale mismatching rate between features to a certain extent.
To quantitatively analyze the impact of ESR strategy on feature matching efficiency, we analyzed the number and processing time required to calculate Euclidean distance, with respect to the number of points. Figure 11 shows the number of calculations with respect to the number of points. As seen, with the number of feature points increasing, using ESR can effectively reduce the number of Euclidean distance calculations. Figure 12 shows the processing time with respect to the number of points. Similarly, with the number of feature points increasing, using ESR can effectively reduce the processing time.

3.4. Contrast Experiments

In this section, we utilize real image pairs to evaluate the performance of the proposed algorithm and comprehensively evaluate the experimental results qualitatively and quantitatively. To better analyze the advantages of the MSPCO algorithm, it is compared with SIFT, SAR-SIFT, OS-SIFT, HOPC, and RIFT.
Figure 13, Figure 14, Figure 15, Figure 16, Figure 17 and Figure 18 show the comparison results of the MSPCO algorithm with SIFT, SAR-SIFT, OS-SIFT, HOPC, and RIFT on six pairs of real optical and SAR images. Since the descriptor of the SIFT algorithm uses gradient information, its registration results are heavily dependent on the similarity of the gradient map of the image. However, the gradient is very sensitive to NRD, thus SIFT performs poorly on multimodal image registration. All image pairs failed to be matched, indicating that SIFT can only register images with the same mode. Similar to SIFT, the descriptor of the SAR-SIFT algorithm is also built based on gradient information. Different from SIFT, given the serious speckle noise problem of SAR images, SAR-SIFT provides a new definition of the gradient by changing the traditional difference gradient operator into a ratio gradient operator, which reduces the influence of speckle noise to a certain extent and improves the registration effect between SAR images. However, the redefined gradient does not weaken the influence of NRD, thus the registration effect of the multimodal image is poor. In all six registration experiments, no successful matching was achieved. The OS-SIFT algorithm is a registration algorithm specially used for optical and SAR images. According to the different characteristics of optical and SAR images, different gradient calculation forms are selected to obtain better registration results. In this paper, OS-SIFT completed the registration, except for the second and the sixth pairs of images. The descriptor of OS-SIFT was also constructed based on gradient information, which was robust to NRD to some extent. The second pair of images had a significant NRD, which directly led to the failure of the registration. For the sixth pair of images, there were more dark areas in the SAR image. The overall intensity of the image was weak, and the edge and texture information was less. Additionally, there were few non-zero values in the gradient map of the SAR image, and most of the non-zero values were noise. The above three algorithms were all gradient information-based methods, among which OS-SIFT achieved the best registration effect. HOPC is a template matching method, which uses phase congruency to construct descriptors. HOPC has illumination and contrast invariance and is robust to NRD, but it does not have scale and rotation invariance. Because of the large-scale difference, HOPC fails to match the first image pairs. RIFT is a feature-based image registration method. RIFT proposes the concept of a maximum index map, which further improves the descriptor similarity. Similar to HOPC, RIFT has no scale invariance. The proposed algorithm constructs the MSPCO descriptor, which not only preserves image details but also reduces the effect of image noise. The proposed method has achieved good matching performance in all six pairs of images.
Table 2 shows the registration results of SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and the proposed method on six pairs of real optical and SAR images. Among them, the OS-SIFT algorithm has the smallest NCM and the largest RMSE. There are two main reasons: one is the inherent disadvantages of the gradient information-based descriptor, and the other is that the Harris feature detection algorithm obtains fewer feature points, which inevitably leads to fewer matches. The NCM of RIFT is 2–5 times larger than OS-SIFT, and its RMSE is about 0.2 pixels lower than OS-SIFT. The proposed method can register all optical and SAR image pairs, achieving the most NCM and the smallest RMSE compared to OS-SIFT and RIFT. HOPC is one of the best template matching methods, with the highest registration accuracy. Compared with HOPC, the maximum difference of RMSE is less than 0.3 pixels, which fully demonstrates the high registration accuracy of the proposed method.

4. Conclusions

In this paper, a new optical and SAR image registration method was proposed based on multi-scale PCOs. To ensure the number and repeatability of features, corner and edge points were detected on the maximum and minimum moment map of phase congruency, respectively. The MSPCO descriptor was constructed, which was highly robust to NRD. In addition, we proposed a new feature-matching strategy based on ESR, which reduced the number of comparisons between features and improved computational efficiency. The experimental results showed that the proposed method was superior to the existing feature-based methods. The registration accuracy was only inferior to the current most advanced template matching method, and the accuracy difference was within 0.3 pixels, which fully demonstrated the robustness and accuracy of the proposed method in optical and SAR image registration. It should be noted that the rotation invariance of the MSPCO descriptor was implemented in the same way as that of RIFT, using a circular convolution sequence. In the future, we will continue to study the rotation invariance of phase congruency.

Author Contributions

Methodology, L.J.; software, L.J.; validation, J.D. and J.Z.; investigation, S.H.; resources, L.J.; data curation, J.Z.; writing original draft preparation, L.J.; writing—review and editing, L.J., J.D., S.H., J.Z. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under grant number 62171467; the National Natural Science Foundation of China under Grant 61901502; in part by the Research Plan Project of NUDT under Grant ZK22-27; the General Program of Natural Science Foundation of Hebei Province of China under grant number F2020506036; and the Research Innovation Development Funding of Army Engineering University Shijiazhuang Campus under grant number KYSZJQZL1902.

Data Availability Statement

All implementation details, sources, and data are available upon request from the corresponding author.

Acknowledgments

The authors would like to thank the anonymous reviewers for their helpful and constructive comments. We would like to thank the proposer of the OS-SIFT, HOPC, and RIFT algorithms for their part in the code.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Paul, S.; Pati, U.C. A comprehensive review on remote sensing image registration. Int. J. Remote Sens. 2021, 42, 5396–5432. [Google Scholar] [CrossRef]
  2. Wu, Y.; Fan, J.; Li, S.; Wang, F.; Liang, W. Fusion of synthetic aperture radar and visible images based on variational multiscale image decomposition. J. Appl. Remote Sens. 2017, 11, 025006. [Google Scholar] [CrossRef]
  3. Jia, L.; Li, M.; Zhang, P.; Wu, Y. SAR Image Change Detection Based on Correlation Kernel and Multistage Extreme Learning Machine. IEEE Trans. Geosci. Remote Sens. 2016, 54, 5993–6006. [Google Scholar] [CrossRef]
  4. Zhang, J.; Su, Q.; Tang, B.; Wang, C.; Li, Y. DPSNet: Multitask Learning Using Geometry Reasoning for Scene Depth and Semantics. IEEE Trans. Neural Networks Learn. Syst. 2021, 1–12. [Google Scholar] [CrossRef]
  5. Markiewicz, J.; Abratkiewicz, K.; Gromek, A.; Ostrowski, W.; Samczyński, P.; Gromek, D. Geometrical Matching of SAR and Optical Images Utilizing ASIFT Features for SAR-based Navigation Aided Systems. Sensors 2019, 19, 5500. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Zitová, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef] [Green Version]
  7. Feng, R.; Shen, H.; Bai, J.; Li, X. Advances and Opportunities in Remote Sensing Image Geometric Registration: A systematic review of state-of-the-art approaches and future research directions. IEEE Geosci. Remote Sens. Mag. 2021, 9, 120–142. [Google Scholar] [CrossRef]
  8. Liu, D.; Mansour, H.; Boufounos, P.T. Robust Mutual Information-Based Multi-Image Registration. In Proceedings of the IGARSS 2019-2019 IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019. [Google Scholar]
  9. Pratt, W.K. Correlation Techniques of Image Registration. IEEE Trans. Aerosp. Electron. Syst. 1974, AES-10, 353–358. [Google Scholar] [CrossRef]
  10. Hast, A. Robust and Invariant Phase Based Local Feature Matching. In Proceedings of the 2014 22nd International Conference on Pattern Recognition, Stockholm, Sweden, 24–28 August 2014. [Google Scholar]
  11. Suri, S.; Reinartz, P. Mutual-Information-Based Registration of TerraSAR-X and Ikonos Imagery in Urban Areas. IEEE Trans. Geosci. Remote Sens. 2010, 48, 939–949. [Google Scholar] [CrossRef]
  12. Ye, Y.; Shan, J.; Bruzzone, L.; Shen, L. Robust Registration of Multimodal Remote Sensing Images Based on Structural Similarity. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2941–2958. [Google Scholar] [CrossRef]
  13. Shechtman, E.; Irani, M. Matching Local Self-Similarities across Images and Videos. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar] [CrossRef] [Green Version]
  14. Ye, Y.; Shen, L.; Hao, M.; Wang, J.; Xu, Z. Robust Optical-to-SAR Image Matching Based on Shape Properties. IEEE Geosci. Remote Sens. Lett. 2017, 14, 564–568. [Google Scholar] [CrossRef]
  15. Xiong, X.; Xu, Q.; Jin, G.; Zhang, H.; Gao, X. Rank-Based Local Self-Similarity Descriptor for Optical-to-SAR Image Matching. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1742–1746. [Google Scholar] [CrossRef]
  16. Ye, Y.; Bruzzone, L.; Shan, J.; Bovolo, F.; Zhu, Q. Fast and Robust Matching for Multimodal Remote Sensing Image Registration. IEEE Trans. Geosci. Remote Sens. 2019, 57, 9059–9070. [Google Scholar] [CrossRef] [Green Version]
  17. Xiang, Y.; Tao, R.; Wan, L.; Wang, F.; You, H. OS-PC: Combining Feature Representation and 3-D Phase Correlation for Subpixel Optical and SAR Image Registration. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6451–6466. [Google Scholar] [CrossRef]
  18. Fan, J.; Wu, Y.; Li, M.; Liang, W.; Cao, Y. SAR and Optical Image Registration Using Nonlinear Diffusion and Phase Congruency Structural Descriptor. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5368–5379. [Google Scholar] [CrossRef]
  19. Xiang, Y.; Tao, R.; Wang, F.; You, H.; Han, B. Automatic Registration of Optical and SAR Images Via Improved Phase Congruency Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 5847–5861. [Google Scholar] [CrossRef]
  20. Sedaghat, A.; Mohammadi, N. Illumination-Robust remote sensing image matching based on oriented self-similarity. ISPRS J. Photogramm. 2019, 153, 21–35. [Google Scholar] [CrossRef]
  21. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  22. Fan, B.; Huo, C.; Pan, C.; Kong, Q. Registration of Optical and SAR Satellite Images by Exploring the Spatial Relationship of the Improved SIFT. IEEE Geosci. Remote Sens. Lett. 2012, 10, 657–661. [Google Scholar] [CrossRef] [Green Version]
  23. Xu, C.; Sui, H.; Li, H.; Liu, J. An automatic optical and SAR image registration method with iterative level set segmentation and SIFT. Int. J. Remote Sens. 2015, 36, 3997–4017. [Google Scholar] [CrossRef]
  24. Ma, W.; Wen, Z.; Wu, Y.; Jiao, L.; Gong, M.; Zheng, Y.; Liu, L. Remote Sensing Image Registration With Modified SIFT and Enhanced Feature Matching. IEEE Geosci. Remote Sens. Lett. 2016, 14, 3–7. [Google Scholar] [CrossRef]
  25. Zhang, W. Combination of SIFT and Canny Edge Detection for Registration Between SAR and Optical Images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 1–5. [Google Scholar] [CrossRef]
  26. Xiang, Y.; Wang, F.; You, H. OS-SIFT: A Robust SIFT-Like Algorithm for High-Resolution Optical-to-SAR Image Registration in Suburban Areas. IEEE Trans. Geosci. Remote Sens. 2018, 56, 3078–3090. [Google Scholar] [CrossRef]
  27. Zhang, X.; Wang, Y.; Liu, H. Robust Optical and SAR Image Registration Based on OS-SIFT and Cascaded Sample Consensus. IEEE Geosci. Remote Sens. Lett. 2021, 19, 1–5. [Google Scholar] [CrossRef]
  28. Paul, S.; Pati, U.C. Automatic optical-to-SAR image registration using a structural descriptor. IET Image Process. 2020, 14, 62–73. [Google Scholar] [CrossRef]
  29. Li, J.; Hu, Q.; Ai, M. RIFT: Multi-Modal Image Matching Based on Radiation-Variation Insensitive Feature Transform. IEEE Trans. Image Process. 2019, 29, 3296–3310. [Google Scholar] [CrossRef]
  30. Wang, L.; Sun, M.; Liu, J.; Cao, L.; Ma, G. A Robust Algorithm Based on Phase Congruency for Optical and SAR Image Registration in Suburban Areas. Remote Sens. 2020, 12, 3339. [Google Scholar] [CrossRef]
  31. Xiong, X.; Jin, G.; Xu, Q.; Zhang, H. Self-Similarity Features for Multimodal Remote Sensing Image Matching. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 12440–12454. [Google Scholar] [CrossRef]
  32. Xiong, X.; Jin, G.; Xu, Q.; Zhang, H.; Wang, L.; Wu, K. Robust Registration Algorithm for Optical and SAR Images Based on Adjacent Self-Similarity Feature. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1. [Google Scholar] [CrossRef]
  33. Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE Features; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  34. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef] [Green Version]
  35. Weickert, J.; Romeny, B.T.H.; Viergever, M.A. Efficient and reliable schemes for nonlinear diffusion filtering. IEEE Trans. Image Process. 1998, 7, 398–410. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Oppenheim, A.; Lim, J. The importance of phase in signals. Proc. IEEE 1981, 69, 529–541. [Google Scholar] [CrossRef]
  37. Morrone, M.; Owens, R. Feature detection from local energy. Pattern Recognit. Lett. 1987, 6, 303–313. [Google Scholar] [CrossRef]
  38. Kovesi, P. Image features from phase congruency. Videre J. Comput. Vis. Res. 1999, 1, 1–26. [Google Scholar]
  39. Dellinger, F.; Delon, J.; Gousseau, Y.; Michel, J.; Tupin, F. SAR-SIFT: A SIFT-Like Algorithm for SAR Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 453–466. [Google Scholar] [CrossRef] [Green Version]
  40. Xiong, X.; Jin, G.; Xu, Q.; Zhang, H. Robust SAR Image Registration Using Rank-Based Ratio Self-similarity. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2358–2368. [Google Scholar] [CrossRef]
Figure 1. Flowchart of the proposed method.
Figure 1. Flowchart of the proposed method.
Electronics 12 01635 g001
Figure 2. Comparison results of GSS and NDSS: (a) GSS; (b) NDSS; (c) Optical image; (d) The fifth layer of GSS; (e) The fifth layer of NDSS.
Figure 2. Comparison results of GSS and NDSS: (a) GSS; (b) NDSS; (c) Optical image; (d) The fifth layer of GSS; (e) The fifth layer of NDSS.
Electronics 12 01635 g002aElectronics 12 01635 g002b
Figure 3. Flowchart of feature detection.
Figure 3. Flowchart of feature detection.
Electronics 12 01635 g003
Figure 4. Multi-scale PCOs.
Figure 4. Multi-scale PCOs.
Electronics 12 01635 g004
Figure 5. Illustrative example of the PCO calculation.
Figure 5. Illustrative example of the PCO calculation.
Electronics 12 01635 g005
Figure 6. Main processing chain of the proposed MSPCO descriptor.
Figure 6. Main processing chain of the proposed MSPCO descriptor.
Electronics 12 01635 g006
Figure 7. Synthetic image pair from Quickbird: (a) R-band; (b) B-band; (c) Synthetic SAR image.
Figure 7. Synthetic image pair from Quickbird: (a) R-band; (b) B-band; (c) Synthetic SAR image.
Electronics 12 01635 g007
Figure 8. Six image pairs:(a) Google Earth and (b) GF 3; (c) Landsat and (d) TerraSAR-X; (e) Google Earth and (f) TerraSAR-X; (g) Google Earth and (h) TerraSAR-X; (i) Google Earth and (j) GF 3; (k) Google Maps and (l) GF 3.
Figure 8. Six image pairs:(a) Google Earth and (b) GF 3; (c) Landsat and (d) TerraSAR-X; (e) Google Earth and (f) TerraSAR-X; (g) Google Earth and (h) TerraSAR-X; (i) Google Earth and (j) GF 3; (k) Google Maps and (l) GF 3.
Electronics 12 01635 g008
Figure 9. Robustness and uniqueness analysis of the descriptors: (a) Robustness analysis; (b) Uniqueness analysis.
Figure 9. Robustness and uniqueness analysis of the descriptors: (a) Robustness analysis; (b) Uniqueness analysis.
Electronics 12 01635 g009
Figure 10. The relationship between scale estimation and NCM.
Figure 10. The relationship between scale estimation and NCM.
Electronics 12 01635 g010
Figure 11. Calculations with respect to the number of points.
Figure 11. Calculations with respect to the number of points.
Electronics 12 01635 g011
Figure 12. Processing time with respect to the number of points.
Figure 12. Processing time with respect to the number of points.
Electronics 12 01635 g012
Figure 13. Experimental results of pair 1. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 13. Experimental results of pair 1. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g013
Figure 14. Experimental results of pair 2. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 14. Experimental results of pair 2. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g014
Figure 15. Experimental results of pair 3. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 15. Experimental results of pair 3. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g015aElectronics 12 01635 g015b
Figure 16. Experimental results of pair 4. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 16. Experimental results of pair 4. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g016
Figure 17. Experimental results of pair 5. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 17. Experimental results of pair 5. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g017
Figure 18. Experimental results of pair 6. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Figure 18. Experimental results of pair 6. The registration results from (af) are SIFT, SAR-SIFT, OS-SIFT, HOPC, RIFT, and MSPCO, respectively.
Electronics 12 01635 g018aElectronics 12 01635 g018b
Table 1. Details of the experimental data.
Table 1. Details of the experimental data.
NoImage SourceSize (Pixels)PolarizationGSDDateLocation and Description
1Google Earth 406 × 406 N/A1.8 mXX/11/2016Suburban area in Nanjing, Jiangsu, China
GF 3 750 × 750 HH1 mXX/11/2016
2Landsat 600 × 600 Band 330 mXX/05/2007Suburban area in Rugen, Germany
TerraSAR-X 600 × 600 N/A30 mXX/03/2008
3Google Earth 528 × 520 N/A3 mXX/11/2007Urban area, significant noise in SAR data
TerraSAR-X 534 × 513 N/A3 mXX/12/2007
4Google Earth 628 × 618 N/A3 mXX/03/2009Urban area, local distortions, significant noise in SAR data
TerraSAR-X 628 × 618 N/A3 mXX/01/2008
5Google Earth 590 × 590 N/A3 mXX/12/2016Urban area in Fuzhou, Jiangxi, China
GF 3 590 × 590 VV3 mXX/11/2016
6Google Maps 520 × 520 N/A0.95 mN/ASuburban area in Newport News, Virginia, USA
GF 3 500 × 500 VV1 mXX/02/2018
Table 2. The results of image registration.
Table 2. The results of image registration.
NoCriterionSIFTSAR-SIFTOS-SIFTHOPCRIFTOURS
1NCM————52————142
RMSE————1.78————1.62
2NCM——————199232245
RMSE——————1.761.861.64
3NCM————73171344712
RMSE————1.961.421.791.68
4NCM————60161355711
RMSE————1.881.451.681.57
5NCM————127146287957
RMSE————1.911.501.721.55
6NCM——————182374
RMSE——————1.651.951.80
“——” means failure to register.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jia, L.; Dong, J.; Huang, S.; Liu, L.; Zhang, J. Optical and SAR Image Registration Based on Multi-Scale Orientated Map of Phase Congruency. Electronics 2023, 12, 1635. https://doi.org/10.3390/electronics12071635

AMA Style

Jia L, Dong J, Huang S, Liu L, Zhang J. Optical and SAR Image Registration Based on Multi-Scale Orientated Map of Phase Congruency. Electronics. 2023; 12(7):1635. https://doi.org/10.3390/electronics12071635

Chicago/Turabian Style

Jia, Leilei, Jian Dong, Siyuan Huang, Limin Liu, and Junning Zhang. 2023. "Optical and SAR Image Registration Based on Multi-Scale Orientated Map of Phase Congruency" Electronics 12, no. 7: 1635. https://doi.org/10.3390/electronics12071635

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop