1. Introduction
Polarimetric Synthetic Aperture Radar (PolSAR) technology stands as a pivotal technique in high-resolution microwave imaging. PolSAR systems use both power and relative phase information from imagery to reflect differences in target scattering characteristics. As a result, PolSAR technology holds significant potential and value in various applications. These include unsupervised land cover segmentation, target detection and recognition, and blind source separation. Additionally, PolSAR technology is valuable for estimating parameters such as soil moisture and biomass. SAR calibration ensures that the pixel values in the image accurately reflect the actual physical characteristics of the ground objects. This improves the accuracy of image classification and target detection. Additionally, calibrated SAR images can be effectively fused with other types of remote sensing data, such as optical images and LIDAR data, providing more comprehensive and accurate surface information [
1,
2,
3]. However, design flaws in radar systems and variations in antenna design and manufacturing can affect the transmission and reception of signals under different environmental conditions, leading to data distortion. Channel imbalance and crosstalk, primarily caused by inconsistent channel performance and energy leakage within the polarimetric channels, are common issues. These errors can disrupt the relative relationships between polarization channels, causing inaccurate reflections of the scattering characteristics of observed terrain features and thereby affecting subsequent application processing [
4,
5]. Calibration estimates and corrects for imbalances and crosstalk between polarimetric channels [
6]. Absolute (radiometric) calibration is not considered here.
The calibration targets can be classified into two categories [
7]: point targets and distributed targets. Point targets require the selection of artificially deployed calibration targets as reference objects for calibration parameter estimation. Classical algorithms for this include the Whitt [
8] and PARC [
9] algorithms, of which the former is susceptible to variations in attitude and the latter is disadvantaged by the mandatory use of external power sources, making it impractical for widespread deployment in scenes. Distributed targets, on the other hand, use naturally occurring targets with well-known polarimetric scattering characteristics instead of artificial calibrators. For instance, Ainsworth [
10] and Quegan [
11] developed algorithmic models that combine both point and distributed targets. Both types of algorithms require low system crosstalk to function accurately, as high crosstalk (amplitude > 0.05) impairs their performance. Additionally, Ainsworth’s algorithm exhibits weak robustness to noise. Further research has been conducted on crosstalk calibration. In the study by Shi [
12], methods for estimating crosstalk and imbalance when corner reflectors are unavailable were proposed. Pincus [
13] derived the calibration technique for circularly polarized antennas. Han conducted an in-depth analysis of circularly polarized SAR calibration techniques under various assumptions [
14].
However, there has been limited effort dedicated to fully estimating the complete polarimetric distortion model through model analysis. In [
15], Villa introduced a numerical optimization technique based on covariance matching to preliminarily examine the anticipated accuracy metrics and identify uncertainties associated with distorted parameters. Comet presented a statistically optimal method evaluated for detection purposes, demonstrating results comparable to or better than those achieved with the generalized maximum-likelihood ratio index. This approach has found successful application in SAR-related fields, such as polarimetric decomposition of In-SAR stacks [
16]. Building on these studies, Zhang [
17,
18] established a loss function for estimation crosstalk and imbalance parameters, iteratively predicting these parameters. However, this method heavily relies on the setting of initial values. Poor initial values may lead to convergence to local optima or yield erroneous analytical results. This paper elaborates on the reasons behind this phenomenon and proposes an enhanced method for estimating distorted parameters. Our contribution proposes an enhanced method for estimating distorted parameters. Building upon Zhang’s work, we analyze the underlying causes of these challenges and introduce an initial search mechanism to mitigate these shortcomings. Specifically, we introduce an outlier detection mechanism to identify outliers, which are subsequently processed using an initial search method known as the interior point method. This approach delivers an improved estimation and calibration algorithm. It not only addresses the issue of inaccurate analytical solutions caused by inappropriate initial values but also generates practical results suitable for practical applications.
The structure of this paper unfolds as follows.
Section 2 lays out the preliminaries essential to this study. Building upon this foundation, in
Section 3, we scrutinize the original algorithm, identifying potential vulnerabilities, devising a method for detecting erroneous values, and presenting corresponding solutions. Subsequently, in
Section 4, numerical simulation experiments are conducted, aimed at validating the important conclusions drawn on the practical use of initial search techniques. In
Section 5, real data are used to demonstrate the practical use of the initial search algorithm. Conclusions are drawn in
Section 6.
3. Troubles and Solutions
3.1. Trouble and Cause
Upon meticulous examination of the iterative process, it becomes apparent that it introduces supplementary parameters, thereby leading to a potential misinterpretation within the computational framework. In particular, within the context of deriving the least squares solution outlined in Equation (20), it necessitates to transition from the domain of the real numbers to the complex space.
Following is a more detailed explanation. This deviation stems from the observation that, during the initial stage of the iterative process, inaccurate estimations lead to distorted power values. To clarify, in the theoretically deduction, the imaginary component of the power, , associated with the correct set of distortion parameters, , ought to be zero. However, during the preliminary estimation process, the function often fails to accurately determine the parameter set, . Consequently, the model’s predictions of power values incorporating these inaccurate parameters do not align fully with the intended values for data restoration. This discrepancy is evident in the analytical solution for power predicted by Equation (20), which yields complex values. Furthermore, the transition of the real-valued unknown power vector from 6 to 12 introduces ambiguity, rendering the characterization of the solution an underdetermined problem.
Nevertheless, owing to the iterative inherent of estimation in this methodology, characterized by a progressive convergence toward a final value, encountering such a challenge is inevitable. Despite Zhang’s attempt to convert the complex matrix into the real domain through matrix transformation, the influence of the complex domain persists. His objective is to facilitate the translation of data from the complex domain to the real domain, with the goal of enhancing efficiency. Nevertheless, this operation does not negate the effect of the imaginary part; instead, it integrates the effects originating from the imaginary part into the real domain. Consequently, the inaccuracy persists, and even minor discrepancies can lead to significant deviations from the real value.
Using the
Figure 1 for analysis, within the model, the real-valued parameters of power predominantly function as scalers. However, as the prediction process unfolds, the incorporation of imaginary components expands the influence of power, transitioning it from solely a scaler to a facilitator of rotation within the model’s framework. This expansion introduces additional pathways within the model, leading to deviations from the true values.
Luckily, the power parameters in SAR exhibit a considerable magnitude compared to distortion parameters. As a result, perturbations in power parameters provide discernible insights into the impacts attributed to distortion parameters, making them distinguishable to some extent.
3.2. Outlier Exclude
Given the polarized outcomes observed with the Comet algorithm, it becomes imperative in practical prediction processes to assess algorithmic results against predefined criteria; this evaluation helps to determine the need for recalibration based on initial values.
In this study, Monte Carlo simulations were conducted to simulate the estimation results of the model under various conditions and derive reasonable discrimination rules. Estimation results that fail to accurately produce parameters were regarded as outliers. In the experiments conducted in
Section 4, the experimental results indicate that the Quegan algorithm produces more stable results compared to those of the Comet algorithm, albeit with room for improvement in precision. Therefore, data predicted by the Comet algorithm were compared against the results of the Quegan algorithm to determine the effectiveness.
In the actual judgment process, the only data available are these distorted parameters. Therefore, we chose to use the Quegan algorithm, which is relatively stable but requires improvement, as a reference to evaluate the quality of the prediction results. The most straightforward comparison lies in the differences between the results of the two algorithms. Hence, the magnitude of the difference between the Comet and Quegan algorithm results served as input features, enabling the determination of whether initial value estimation is warranted based on predefined thresholds.
The specific algorithmic process involves obtaining results from both the Comet and Quegan algorithms separately, then operating the difference treatment outlined in the subsequent equation:
where
denotes the results estimated by the Comet method, while
denotes Quegan’s counterpart.
Before conducting model training, it is essential to categorize the obtained results into outlier or truth, labeled as 0 and 1, respectively. This step transforms the problem into a binary classification task with five-dimensional features. To address this, we employ KNN algorithm for binary classification to model the outcomes.
Given the training set
, where
, in the training phase of this problem, the calculation of point distances performs optimally when utilizing the Chebyshev distance, that is,
Given the data’s five-dimensional nature, we employ the
kd-tree for nearest neighbor search. The
kd-tree is employed to expedite the search for the closest neighboring points to a target point [
22]. During the construction phase of the
kd-tree, a dimension is selected, and the dataset is sorted according to the values along that dimension. Subsequently, the dataset is partitioned into two subsets based on the median value, thus forming left and right subtrees. This partitioning process is recursively applied to each subtree until each leaf node contains a single data point.
During the search process, initiated from the root node, the target node is compared with the current node, and the search direction is determined based on the value of the target point along the current dimension, moving either toward the left or right subtree. This iterative process continues until a leaf node is reached, where the data of that node are recorded as the current nearest neighbor. Subsequently, the search algorithm backtracks to the parent node to check for closer points; if found, the nearest neighbor is updated. This backtracking process continues until the root node is reached, thereby completing the search operation.
Using the Chebyshev distance metric, the method seeks the k samples in the training set closest to
which are then represented
. According to the principle of majority voting, the error rate of the k-nearest neighbor training samples for the input instance
is defined as the proportion of k-nearest neighbor training samples whose labels differ from the input label:
When determining the value of
k, the objective is to minimize the error rate. Therefore, the training function is designed to optimize this objective:
in which
denotes an indicator function, specifically as:
Further elaboration on the specific process is omitted here; detailed information can be found in [
23].
3.3. Initial Search
Referring to Equation (19), we observe that the loss function is not subjected to a differentiation process. The loss function remains undifferentiated, thereby avoiding the problem. However, due to the high dimensionality and the nonlinearity of the function, obtaining an accurate solution without processing is challenging. Nevertheless, a middle-ground approach can be considered. We derive an initial rough solution from the undifferentiated loss function and utilize it as the starting point for Equation (20). This step allows the parameters to approach the result before the iteration starts, thereby reducing the likelihood of convergence to local optima. We signify these unknown 16 parameters into a
parameter vector
The objective function for initial value is set as follows:
To recalibrate parameters flagged as outliers, an initial pre-processing is executed using the interior point method [
24]. The specific iterative procedure unfolds as outlined below:
- (a)
Boundary Setting
The range of unknown parameters is established based on Quegan’s prediction results, following the constraint:
Let’s refer to the feasible region as
S =
In detail, extensive experiments have demonstrated that setting the boundaries of the crosstalk and imbalance parameters centered around the results obtained from the Quegan algorithm, with an estimation range set to 0.1, yields more favorable search outcomes. Additionally, the power boundaries are subtly expanded based on the values of the calibrated observation matrix. The function
will be centered around the Quegan result, denoted as:
where
and
denote the range of initial search, while
and
represent the Quegan’s calibration outcome.
- (b)
Parameter Updating
Given the set range, we define the Lagrangian function as:
where
is a small positive number, ensuring that
as
approaches the boundary. This constraint ensures that iteration points remain within the feasible domain. Given the smallness of
, the function value
closely approximates
. Thus, solving the problem is simplified to:
The function updates iteration steps by employing the Hessian matrix. The formula for Hessian function corresponds to this Lagrangian equation:
The specific iteration equation is:
in which
. In this iteration,
is updated based on the previous value
, the inverse of Hessian matrix
, and the gradient of the Lagrangian function
.
- (c)
Matrix Approximation
Because of the complex nature of function solution, calculating the inverse matrix of the Hessian matrix presents a formidable challenge, necessitating the utilization of iterative techniques to approximate this outcome. Note that, in this procedure, our objective is to ascertain the zero points of a function through the combined application of Newton’s method for locating stationary points and the BFGS algorithm [
25]. As both methodologies involve iterative steps, these processes are conducted concurrently. The outlined approach unfolds as follows:
where
. In the first step,
is the identity matrix.
In the process of initial value exploration, it is imperative to acknowledge the potential inaccuracies introduced by the utilization of interior-point algorithms. These inherent errors arise from the optimization algorithms potentially halting at a certain distance from the boundaries of inequality constraints rather than converging closely to them. Consequently, such circumstances may result in a degree of deviation in the final optimization outcome, as the algorithm fails to thoroughly explore the solution space adjacent to the constraint boundaries. Hence, subsequent precise computations are deemed necessary following the acquisition of initial values.
- (d)
Algorithm steps
- 1.
Take the penalty factor , and the error ,
- 2.
Select the initial point and set the number of iterations to k = 1,
- 3.
Construct the hessian matrix corresponding to the Lagrange function , update the iterative parameters according to the Formulas (31)–(32), and k = k + 1,
- 4.
Determine whether the stop criteria are met:
If , then stop iterating and output , otherwise repeat step 3.
At the outset of the search, the interior-point method typically achieves linear error reduction, progressing to quadratic convergence near the optimal solution, facilitated by its utilization of the Hessian matrix. Despite the problem’s non-convex nature, encountering local minima is inevitable, although multiple adjustments to the initial point can mitigate this.
The initial value search yields preliminary results, and in cases where convergence fails, expanding the parameter search range may be necessary, albeit at the cost of increased computational complexity and reduced efficiency. Each iteration of the interior-point method operates with a complexity of , where n denotes the number of variables. Given 16 variables in the initial value search process, the computational demands are substantial.
Furthermore, computing and storing the Hessian matrix are both time- and space-intensive tasks. To streamline this process, the quasi-Newton algorithm approximates the Hessian matrix, simplifying computations.
3.4. Algorithms
In subsequent sections of the paper, we will use the term “Comet with Initial Search” interchangeably with “Comet IS” to refer to experiments conducted with this additional search step. Before conducting the experiment, Algorithms –3—the Quegan algorithm, the Ainsworth algorithm, and the Comet IS algorithm—are briefly summarized in this paper.
Algorithm 1 Quegan |
(1). Crosstalk estimation based on second-order statistics of natural terrain features: |
, , , , . |
(2). Estimating the magnitude and phase of cross-polarization channel imbalance: |
, , , . |
(3). Computing co-polarization channel imbalance based on corner reflectors. |
Algorithm 2 Ainsworth |
|
by |
|
The crosstalk corrections are initially zero. |
(3) Iterative updating of distorted parameters: |
for |
a. Apply the total correction to the observed covariance: . , (12)
|
|
, . The detailed explanations of these equations can be found in [10]. |
, according to |
|
e. Update crosstalk by: |
|
is performed based on the new crosstalk matrix P as defined by Equation (4): |
|
|
. end |
Algorithm 3 Comet IS |
(1). Segment the SAR image. |
(2). Estimate parameters by Comet method and Quegan method, respectively. |
(3). Compute the difference between the outcomes obtained from the two methods. |
(4). Classify results based on KNN model |
|
(5). Search the position of outlier and compute the initial value according to Equation (29). |
(6). Introduce it into Comet algorithm for recalibration. |
(7). Perform interpolation processing on the κ-set. |
(8). Calculate the h value in conjunction with κ and the triangular reflector. |
4. Simulation Test
Given the unknown true distortion parameters in real data, our exploration of the optimality of the Comet method relies mostly on simulation experiments, wherein the true distortion parameters are ascertainable. This approach allows us to investigate the effectiveness of the methods as model metrics.
This paper delves into the Comet calibration methodology, which hinges upon the utilization of standard targets. To gauge the accuracy of the prediction algorithm, various combinations of and within the H/V polarization covariance matrix are employed. The simulation analysis presented in this section primarily assesses the viability of target selection parameters and the efficacy of the calibration method across the chosen targets.
To further investigate the Comet algorithm, we employ Monte Carlo simulations to explore various combinations of and . Additionally, this study integrates classic algorithms such as Quegan and Ainsworth for a comprehensive comparative evaluation of algorithmic efficacy.
4.1. Genarating Variables
Initially, random observation data are generated according to the specified parameters,
- (1)
ideal covariance matrix elements:
By specifying the values of
and
, the generator can produce the covariance matrix. Based on the polarimetric ratios defined in (12), (13), and (14), the ideal covariance matrix elements of the covariance matrix are calculated as follows:
The function U (0, 1) denotes a random number generator which yields a single random number uniformly distributed within the interval (0, 1).
Set
, and then the real covariance matrix can be obtained by combining (
). Here, we set
and define the range of covariance matrix parameters as
and
, that is:
The notation signifies a uniformly sampled vector, where the elements start from and terminate at , with an interval of between successive elements.
- (2)
distortion parameters:
To facilitate calibration simulation, random polarimetric distortion parameters must be added to the real covariance matrices. These parameters, integral to the calibration process, are defined as follows:
In our experiment, the predictions of and were not taken into consideration; therefore, they were simply set to 1.
Given the number of samples N as 1000 and set SNR = 20 dB. The detailed steps for constructing the noise matrix are as follows:
(a)
(b)
(c) , ,
(d) if , output ;
else return step (c).
4.2. On the Practice of Comet
To assess the performance of three algorithms in parameter prediction, the Euclidean norms of the differences between the predicted values and the actual computational results were computed for each algorithm. In this study, the accuracy of parameters is assessed by computing the
norm between the estimated parameters and the actual results. Specifically,
where
represents the configured parameter, whereas
denotes the estimated outcome.
The results of the computation were converted to dB conversion in
Figure 2. In the Ainsworth dataset, white spots indicate matrix singularity, making distortion parameters unattainable. Similarly, red regions are unsuitable for the Ainsworth algorithm combinations. While the outcomes in these regions may not deviate as drastically from the actual scenario as observed with the white points, they still exhibit considerable disparities from the ground truth. Conversely, concerning the Comet algorithm, it is notable that while there is some overlap between erroneous parameter estimation and function loss at certain points, a significant portion of the points does not exhibit a correlated relationship. An evident example is the occurrence of curves in both the Comet error plot and the function loss plot. The arcs in the error plot depict the algorithm’s subpar estimations, whereas the finer arcs in the function loss plot demonstrate comparatively better predictive outcomes. This disparity arises due to the shared segments of the two plots, stemming from constraints such as finite maximum iteration counts, non-convergent initial parameters, or oscillations during convergence. Conversely, the contradictions between the two plots serve as substantial evidence affirming the error source analysis outlined in
Section 3.
The first experiment is devoted to illustrating whether the Comet algorithm outperforms the Quegan and Ainsworth algorithms or not. The findings reveal a noteworthy stability in outcome variance when utilizing the Quegan and Ainsworth methods. Notably, while Ainsworth displays superior accuracy over the Quegan algorithm, it occasionally confronts challenges in estimation feasibility at isolated data points. In contrast, the Comet method displays a dichotomy in outcomes: successful predictions exhibit considerably smaller errors compared to the two former algorithms, yet a notable frequency of erroneous estimations persists.
Analogous conclusions can be drawn from the statistical data of the three algorithms shown in
Table 1. It is noteworthy that, due to certain failures in Ainsworth’s algorithm, irregular results were excluded from the statistical analysis. The statistical findings reveal that, apart from a few exceptional points, the overall performance of the Ainsworth algorithm remains relatively stable, with generally favorable prediction outcomes and a total range length of less than 12 dB. The average variance of this algorithm is
9.4803 dB, indicating its superiority among the three algorithms. However, the Quegan algorithm falls short of Ainsworth’s results in both optimal and average estimation variances. Nevertheless, this does not imply that Ainsworth’s algorithm consistently outperforms Quegan’s. For instance, within the interval
, the Quegan algorithm not only exhibits greater stability but also yields more accurate predictions. On the other hand, the Comet algorithm demonstrates superior performance in successful predictions (−18.1582 dB) but the poorest performance in erroneous predictions (10.8171 dB). In comparison to the range of lengths of less than 15 and 12 for the former two algorithms, the length associated with the Comet algorithm approaches approximately 20 dB.
This observation corroborates our initial investigation, highlighting the persistent risk of inaccurate estimations even when employing minimal loss functions. It underscores the importance of further research to mitigate such errors and enhance the reliability of our predictions.
4.3. KNN Classification
In this experiment, the classification data groups were established by categorizing them according to the Euclidean norm of the difference between predicted and actual results.
A total of 6600 sets of data were collected, comprising 927 outlier sets and 5673 true data. The decision threshold distinguishing outliers from true values is set at 0.15, based on the calculation result from Equation (34). This threshold, determined experimentally, provides optimal classification performance. It also accounts for an average error difference of 0.05 across the 10 distortion parameters (real and imaginary parts) considered during its selection process. Prior to the experiment, the dataset was divided into training and testing sets in a 7:3 ratio. In the process of utilizing the
kd-tree, a bucket size of 50 was set to balance search speed and accuracy. As shown in
Figure 3, the function stabilizes around the 29th iteration. At this point, the observed minimum target value and the estimated minimum function value are 0.006 and 0.0060048, respectively.
Fortunately, for classification training, we utilized the ‘fitcknn’ function available in MATLAB. To optimize the performance of the classification model, we employed the Bayesian optimization algorithm due to its effectiveness in exploring complex parameter spaces and maximizing model accuracy. The specific configurations of training parameters, including those for the ‘fitcknn’ function and the Bayesian optimization algorithm, are provided in detail in
Appendix A.
After training, the testing data were evaluated using the trained model, yielding the results depicted in
Figure 4.
Figure 4a,b illustrate the actual classification and predicted outcomes, respectively. Furthermore, cross-validation loss was employed as a complementary metric to assess the model’s performance on the testing set. The testing set demonstrated strong performance on the trained model, with target function loss of the 0.0060048, indicating effective outlier identification capabilities. For more detailed training results, readers are encouraged to refer to the author’s GitHub page.
4.4. Comet with Initial Search
After identifying the outliers, experiments were re-conducted at the corresponding positions. Prior to re-conducting the experiments, an additional initial value search step was introduced to refine the experimental process.
The initial value was set as recommended in
Section 3 and the Lagrange coefficient λ was set to 0.02. During the experiment, unreasonable values were detected using the KNN model, prompting an initial value search. Some areas required multiple adjustments, with a maximum of four adjustments needed in certain cases. Following the incorporation of initial value search, a noticeable reduction in the discrepancy between predicted results and actual set parameters was observed. The results after re-experimentation are shown in
Figure 5. However, it became apparent that the estimation performance for arch positions slightly lag behind other parts. Nevertheless, to some extent, compared to the original algorithm, this approach demonstrates an improvement in mitigating the impact of the complex domain. Furthermore, even in areas of slightly inferior performance, the algorithm consistently outperforms both the Quegan and Ainsworth algorithms.
Statistical data analysis reveals consistent findings displayed in
Table 2, indicating that the inclusion of an initial value search process constrains the prediction error range to below −1.0132 dB. On average, the overall predictive outcomes appear quite favorable. However, upon scrutinizing the analysis of the optimal prediction points, it becomes evident that despite the inclusion of the initial value search, the recalculation of outlier points fails to surpass the initially computed results. This suggests that while the initial value search process helps constrain prediction errors on average, it may not fully address underlying issues within the algorithm. Instead, it serves more as a remedial measure to mitigate the impact of these inherent issues.
5. Experiment on Real Data
In this section, the empirical data were acquired by the Microwave Microsystems Research and Development Department at the Aerospace Information Research Institute, Chinese Academy of Sciences. The data, obtained at the Luo Ding Airport in Zhao Qing, Guangdong in May 2021, consist of an S-band quad-polarimetric image captured by the airborne experimental SAR system.
The aircraft, with active phased array antenna supports, flew at a speed of 74 m/s and at an altitude of 2600 m. The nadir angle of the airborne SAR radar is 45 downward. Employing a dual-illumination mode for target observation, the radar’s center of the scene was at 3989 m. The SAR operated at a frequency of 3.2 GHz, with a range sampling rate of 400 MHz, a signal bandwidth of 300 MHz, and a signal sampling interval of 1000 Hz. The radar achieved a resolution of 1 m, allowing for detailed imaging of the target area. Three sets of parallel data were obtained from this sampling, which were used for calibration of testing results in the actual data experiment.
Near the imaging airport, various artificial corner reflectors were deployed for data polarization quality assessment and calibration. Within the radar testing area, 16 trihedral corner reflectors and two bi-corner reflectors rotated at 45 degrees were positioned. The schematic of the placements is illustrated in
Figure 6b: the magenta plus sign marks the positions where a trihedral corner reflector is placed, and the yellow marks the position where a rotational bi-corner reflector is placed.
Three algorithms—the Ainsworth, Quegan, and Comet IS algorithm—were employed in the experiments. The results indicate that there are discernible differences among the outcomes generated by the three algorithms, with notable similarities observed between the results of the Comet IS algorithm and the Quegan algorithm. This similarity can be attributed to the outlier exclusion procedure employed earlier, where outlier detection was conducted based on the results of the Quegan algorithm. Using this result as a reference, we established the initial search range for revalidation, which influenced the outcomes of the Comet IS algorithm. Furthermore, upon observing the measurement results of Ainsworth in
Figure 7, it is evident that the amplitude values of the crosstalk coefficient “u” and “z” largely overlap. Additionally, around the position of approximately 2600, both latter two algorithms exhibit similar peaks, whereas the Ainsworth shows no significant fluctuations.
Figure 8 sequentially presents the SAR raw image after Pauli decomposition, followed by the image calibrated using the Quegan algorithm, Ainsworth algorithm, and Comet IS algorithm. Upon calibration, all algorithms exhibit varying degrees of success in restoring the information of the original image. Compared to the pre-calibrated Pauli color image, the post-calibrated image shows noticeable changes: fields and grasslands near the airport appear greener, indicating single or volume scattering, while clutter in dark target areas such as airport runways adjacent to grasslands and ponds near fields transitions from red to purple. This is consistent with the scattering mechanism influenced by the characteristics of the objects themselves.
To carefully analyze the calibration effects of algorithms on images, the central positions of the images were selected for magnification research in
Figure 9. Position 1 corresponds to areas with bare soil and vegetation, predominantly exhibiting surface and volume scattering, which appear as green and blue, respectively, in the Pauli decomposition. Position 2 corresponds to residential areas, primarily exhibiting double bounce scattering, resulting in a red appearance. Position 3 represents calm water bodies, exhibiting specular reflection, theoretically appearing black, or undergoing surface scattering upon reaching a certain roughness on the water surface, appearing blue [
26].
Before calibration, position 1 exhibited yellow and sporadic red hues. Following calibration with the three algorithms, position 1 predominantly appears green with sporadic blue hues. Moreover, after calibration with the Comet IS algorithm, blue is more pronounced at position 1. The built-up area at position 2 exhibits double bounce scattering characteristics both before and after calibration. Initially, position 3 appears mainly black before calibration, yet still interspersed with some red. After calibration, the results obtained using the Ainsworth algorithm reveal sporadic green hues in addition to black at position 3. The Quegan and Comet IS algorithms show sporadic blue hues at position 3.
However, solely relying on visual observation makes it difficult to distinguish the calibration effects of the three algorithms. Therefore, further validation was conducted using corner reflectors to assess the accuracy of each algorithm’s performance and its ability to restore the SAR data.
During the validation process, we utilized 16 trihedral corner reflectors, denoted as TH1 to TH16, along with two bi-static corner reflectors rotated at 45
, labeled as RH1 and RH2. Subsequently, we employed the returned values from points with established reflector positions to estimate residual errors. The calculation process proceeded as follows:
Among these, xtalk, camp, cph, xcmp, and xcph represent polarization isolation, co-polarization channel imbalance magnitude, co-polarization channel imbalance phase, cross-polarization channel imbalance magnitude, and cross-polarization channel imbalance phase, respectively.
and
denote the number of TH and RH. The first three evaluation metrics were calculated using trihedral corner reflectors, while the remaining evaluation metrics were calculated using rotated bi-static corner reflectors. The statistics of these three data are presented in
Table 3,
Table 4 and
Table 5.
For a more intuitive comparison, in the detailed evaluation table, we show the results of the first group of every corner reflector in
Table 6,
Table 7 and
Table 8, in order. From the overall evaluation table, the Comet IS algorithm exhibited favorable results in the xtalk, cph, and xph metrics.
Generally speaking, the Quegan algorithm and the Comet IS algorithm exhibit significant advantages over the Ainsworth algorithm in terms of xtlak. However, in specific areas, the Ainsworth algorithm’s calibration results show superiority. The results of the first set of data are analyzed as follows: in the TH4 region, the xtalk values for the Quegan algorithm and the Comet IS algorithm are 1.1195 and 17.1912, while Ainsworth’s xtalk is 29.6012. Similar data results also appear in regions like TH5 and TH16. For the camp and cph, the error results of the three algorithms are very similar, with Ainsworth’s results slightly weaker. Compared to the other two algorithms, the Quegan algorithm achieves better results in terms of cross-polarization channel imbalance. The results from the second and third sets of data closely resembled those from the first set.
Analyzing each algorithm individually reveals that for the Quegan algorithm, xtalk at the positions of TH2, TH4, and TH16 is not ideal, consistent with the results of the simulation experiments. In the simulation experiments, when both
and
are relatively high, the calibration results are weaker compared to other positions. At positions TH12 and TH13, xtlak is higher, but the cross-polarization channel imbalance and phase recovery effects are lower. It can be inferred from
Table 9 that this may be related to the higher eta values of the observed data at corresponding positions. In contrast, the Ainsworth algorithm has low xtalk at TH2, and the cross-polarization channel imbalance is also larger. Although this does not align with the simulation experiment results,
Table 9 reveals that the corresponding eta value is higher. At the TH5 position, the crosstalk isolation is high, which also corresponds to the simulation experiment results: under the combination
, and the corresponding eta, its value is close to 1, indicating good performance in both xtalk and camp, cph indicators. The results of the Comet IS algorithm show low xtlak at the TH2 position, which is consistent with the simulation experiment results. Under the combination
, the results of the simulation experiment are also weaker, but the performance in crosstalk isolation is better than the other two algorithms. Similar situations also occur at the TH16 position. Additionally, the results of Comet IS are generally stable, with good performance in xtalk. Particularly, the xmp results obtained by Comet IS near RH1 are significantly weaker than the other two algorithms. It can be observed from
Table 9 that the combination of
= [0.42, −5.8] at this point is also a set of combinations whose performance is slightly inferior in the simulation experiments. Similarly, the Quegan algorithm near RH2 also exhibits deficiencies.
Furthermore, in
Figure 10, the x- and co-pol signature of the corner reflector RH9 are plotted. The calibration results of the Quegan algorithm and the Comet IS algorithm are superior, while the results of the Ainsworth algorithm are relatively poor. This is consistent with the results presented in
Table 6,
Table 7 and
Table 8.
The analysis of individual corner reflectors indicates that no algorithm has an absolute advantage because each algorithm’s model may not fully correspond to the actual situation during modeling. The assumptions about noise made by these three algorithms are also quite simplistic, which is a compromise made to solve the model but does not reflect the actual situation. Moreover, noise during radar data collection is not constant, explaining why the experimental results of corner reflectors do not fully follow the simulation experiments. This limitation also applies to the Comet IS algorithm. Furthermore, in practical applications, due to the limited number of sampling points and the complexity of the environment, it is difficult to meet the premise of an infinite number of sampling points during the calibration process. On the contrary, in practical applications, a balance must be struck in terms of sample quantity.
6. Conclusions
This paper investigates techniques for polarimetric SAR calibration employing distributed targets. The establishment of Covariance Matching Estimation Techniques and the formulation of the loss function offer a more precise method for predicting distortion coefficients and scrutinizing result authenticity. However, challenges arise due to the analytical solution derived from a single differentiation and the introduction of unrealistically estimated parameters. This leads to an underdetermined equation and increased unknowns.
This paper’s contribution lies in evaluating the reasonableness of the solution and re-solving unreasonable outcomes. Prior to solving, an initial value search process is incorporated to confine the solution within a smaller range, mitigating to some extent the diffusion of real-domain solutions into the complex domain. With the successful establishment of an evaluation mechanism, the solving process can be iteratively conducted. Given that the training data almost completely cover real-world scenarios, incorrect data from initial training can be accurately identified during actual application. This allows for re-execution of predictions using an initial value search. Comprehensive learning data ensure the effectiveness and robustness of the recognition mechanism. The prediction process using an initial value search can converge more swiftly and accurately in the correct direction, resulting in precise distortion parameter estimation due to the rational setup of initial values.
The experimental results demonstrate the outstanding performance of the proposed Comet IS algorithm. During the simulation stage, Comet IS exhibited a significant reduction of 4.9419 dB in average error compared to its predecessor, Comet. Relative to classical algorithms, Ainsworth and Quegan, Comet IS showed reductions of 2.1211 dB and 7.7069 dB in average error, respectively.
In the real data experiment phase, all three algorithms—Comet IS, Quegan, and Ainsworth—exhibited a notable calibration effect on the original images, rendering the scattering characteristics of SAR images more pronounced. Results from corner reflectors further validate the excellent performance of the Comet IS algorithm across various indicators. These include polarization isolation (Ainsworth17.9926 dB, Quegan1.5226 dB), co-polarization channel imbalance magnitude (Ainsworth4.3862 dB, Quegan0.0180 dB), co-polarization channel imbalance phase (Ainsworth1.4125, Quegan0.0270), and cross-polarization channel imbalance phase (Ainsworth1.3231, Quegan0.1492).
Nevertheless, the problem at hand presents several areas requiring immediate improvement, such as significant computational overhead and prolonged computation time. In comparison to the Q and A algorithms, this algorithm exhibits lower computational complexity and higher efficiency. However, it delivers more accurate data and is better suited for scenarios requiring high-standard and high-quality data acquisition. Additionally, there may be instances necessitating multiple corrections of the initial value search range to attain reasonable outcomes. Furthermore, akin to other studies, this paper assumes similar noise characteristics, merely positing average noise power while overlooking varying effects of noise across different channels.