1. Introduction
The frequency-modulated continuous wave (FMCW) radar offers several advantages over traditional pulse radar systems, such as lightweight design, structural simplicity, low cost, and low probability of interception [
1]. As a result, FMCW radar has found widespread applications in fields such as short-range security inspection, autonomous driving, deformation monitoring, and vital sign detection [
2,
3,
4,
5]. To achieve finer target range resolution, radar systems require ultra-wide instantaneous bandwidth. However, such ultra-wide bandwidth imposes stringent demands on signal acquisition systems, making analog demodulation (dechirp) a crucial solution for processing ultra-wideband signals. However, dechirp technology requires exceptionally high linearity in the chirp modulation of transmitted and received signals, especially when the signal bandwidth is large. Failure to meet these linearity requirements can result in significant degradation of range resolution and even prevent accurate imaging [
6].
Unfortunately, existing transmit–receive chains inevitably introduce some degree of nonlinearity into the signal. For instance, linear frequency modulation (LFM) signals generated using digital direct synthesis (DDS) and phase-locked loops inherently exhibit certain nonlinear characteristics. Furthermore, nonlinear components such as frequency multipliers, mixers, and filters in the signal transmission path exacerbate the degradation of chirp linearity. Consequently, in ultra-wideband application scenarios requiring high-quality range resolution, nonlinear correction techniques are indispensable. These techniques play a pivotal role in determining the imaging quality of range profiles and are critical for ensuring the performance of FMCW radar systems [
7,
8].
Current nonlinear correction techniques can be broadly categorized into hardware-based and software-based approaches. Hardware methods aim to minimize chirp nonlinearity through real-time preprocessing by adjusting the signal generation circuitry. Common hardware techniques include predistortion voltage compensation, delay phase detection, delay phase-locked loops, and various predistortion methods addressing VCO modulation voltage distortion and DDS digital distortion [
9,
10,
11]. Although hardware methods can output high-linearity chirp signals in real time, their parameter designs are often susceptible to environmental factors such as temperature, resulting in insufficient stability.
In contrast, software correction methods have garnered significant attention due to their strong versatility, portability, and high correction accuracy. Unlike the real-time processing of hardware approaches, software methods employ digital signal processing techniques to estimate and correct nonlinear factors in the sampled intermediate-frequency (IF) signals.
Existing software correction methods primarily focus on estimating and compensating for the nonlinearities in the transmission chain. These methods typically model nonlinear phase errors as one of three types: polynomial phase, periodic phase, and stochastic phase. In practical radar systems, all three types of phase errors coexist, albeit with varying proportions depending on the system. Most current research focuses on the estimation and correction of polynomial phase errors. For example, Anghel et al. proposed the use of the high-order ambiguity function (HAF) to estimate polynomial coefficients and correct space-variant nonlinear errors through time-domain interpolation and resampling, making it suitable for engineering applications [
12]. Wang et al. employed the wavelet synchrosqueezed transform to derive the time-frequency relationship of the reference chirp signal for fitting nonlinear parameters. The residual video phase (RVP) method was then applied to compensate for nonlinearities in chirp signals. However, the WSST-based approach has strict limitations, such as requiring that the signal frequency must vary slowly [
13]. Similarly, Rongyao Zheng et al. combined the complete ensemble empirical mode decomposition with HAF to estimate polynomial phase errors and corrected them via time-domain resampling. These methods successfully eliminate polynomial nonlinear errors [
14].
Nevertheless, none of the aforementioned methods address the correction of periodic phase errors, and studies focusing on high-frequency periodic nonlinearities remain scarce. Ayhan et al. suggested that periodic nonlinearities may arise from various sources, such as DDS spurs or clock spurs, but no correction solution was proposed [
15]. Zhao et al. extracted phase information directly from the sampled signal and used a three-component sinusoidal model for fitting, which could partially mitigate the effects of periodic nonlinearities. However, this approach often fails to meet the requirements for direct sampling under ultra-wideband conditions [
16].
In summary, there is currently no effective method for periodic error correction. Modeling periodic errors as multi-sinusoidal models for parameter estimation poses a series of challenging problems. First, the parameter estimation of general multi-sinusoidal signals is typically based on frequency estimation, which is often realized through the Fourier transform of discrete-time samples. However, accurate frequency estimation requires a large number of samples, imposing strict demands on practical applications [
17]. Although methods based on eigenvalue decomposition, such as Multiple Signal Classification (MUSIC), Prony’s method, and Estimation of Signal Parameters via Rotational Invariance Techniques (ESPRIT) [
18,
19,
20], can achieve high-frequency resolution under limited data length and sampling rates, their performance degrades significantly in low signal-to-noise ratio (SNR) scenarios [
21,
22]. Second, periodogram-based methods and eigenvalue decomposition techniques require prior knowledge of the number of frequencies to be estimated. However, this information is typically unavailable in practical scenarios, necessitating the use of additional criteria such as the Akaike Information Criterion (AIC) [
23] or the Minimum Description Length (MDL) [
24] to determine the number of sinusoidal components in the signal. Finally, after obtaining the frequencies, a concise and efficient method is still needed to estimate the corresponding amplitude and phase parameters.
Building on recent advancements in frontier technologies such as deep learning [
25,
26], this paper employs a neural network-based approach to address the aforementioned challenges. The proposed method trains a neural network using simulated data. Experimental results demonstrate that the proposed neural network significantly outperforms existing frequency estimation methods, particularly under medium-to-high noise conditions, precisely estimating frequencies from data contaminated with noise of unknown variance. Furthermore, the paper adopts a neural network model to estimate the number of frequency components in the signal. Based on this foundation, the neural network model is first applied for the preliminary estimation of the frequencies associated with periodic sinusoidal errors. A phase-matching method is then proposed to extract amplitude and phase information, thereby determining the initial parameters of the periodic sinusoidal errors.
This paper proposes a joint estimation algorithm that combines Adaptive Moment Estimation (ADAM) [
27,
28] and the Limited-Memory Broyden–Fletcher–Goldfarb–Shanno (LBFGS) algorithm to achieve more accurate estimation of periodic errors [
29]. In the process of periodic error estimation, the ADAM algorithm leverages its adaptive learning rate adjustment capability to provide a rapidly convergent initial solution, while the LBFGS algorithm refines the parameters through precise second-order approximations, thereby improving the accuracy of the final estimation results. This joint estimation algorithm not only performs exceptionally well in handling periodic errors but also holds significant potential for application in other complex nonlinear optimization problems.
The remainder of this paper is organized as follows.
Section 2 provides a detailed introduction to the IF error model and the analysis of periodic phase distortion.
Section 3 presents the estimation and correction process for periodic phase errors.
Section 4 validates the effectiveness and superiority of the proposed method using real-world data. Finally,
Section 5 concludes the paper with a summary of the findings.
3. Estimation and Correction of Periodic Phase Errors
Based on the analysis in the previous section, periodic errors can generate paired echoes near the target, severely degrading the detection performance for nearby weak targets and far-range targets. To address this issue, periodic nonlinear correction is essential. Since nonlinearity is an intrinsic modulation characteristic of the system and independent of the target, this section first employs ideal single targets, such as delay lines or corner reflectors, to accurately estimate the system’s nonlinearity. The estimated parameters are then used to construct a matched Fourier transform, enabling space-variant nonlinear correction for multiple targets.
The phase of the beat echo signal is first obtained by phase unwrapping as follows:
In the equation, it should be noted that
;
and
are unknowns, but
can be treated as a constant, which does not affect the subsequent estimation of sinusoidal parameters. Prior to fitting, a DC removal process can be applied to eliminate
. The complete process of periodic error correction is illustrated in
Figure 1. First, the phase information of the beat signal is extracted using a corner reflector or delay line. Next, initial estimates of the periodic error are obtained using a deep learning model combined with a phase-matching method. Subsequently, precise optimization is performed by integrating the ADAM and LBFGS algorithms. Finally, the phase error is corrected using Fourier matching techniques based on the estimated parameters, yielding the fully corrected beat signal.
3.1. Initial Estimation of Periodic Errors Based on a Neural Network Model
3.1.1. Frequency Estimation Model
In this paper, a model suitable for estimating periodic phase errors was trained using simulated data. As shown in
Figure 2, the frequency estimation module is designed to transform the input signal into a frequency representation, enabling the extraction of frequency features from the signal. This module consists of an input layer, multiple convolutional layers, and an output layer, with ideal frequency representations generated using a narrow Gaussian kernel.
Input Layer: The input signal undergoes an initial transformation through a linear layer. This transformation projects the signal into a higher-dimensional internal representation space, enabling the model to extract a richer set of features for subsequent operations. The primary purpose of this transformation is to enhance feature representation, facilitating more accurate feature extraction during subsequent convolution operations. The use of a fully connected layer in this step allows the network to capture complex relationships within the input signal, ensuring that the initial features are adequately extracted for further processing in the network.
Convolutional Layers: The output from the input layer is processed through multiple one-dimensional convolutional layers. The design of the convolutional layers is inspired by the success of convolutional neural networks in image processing, with the primary advantage of effectively capturing local features within the signal. Each convolutional layer employs multiple filters, followed by a batch normalization layer and a ReLU activation function after the convolution operation. The convolutional kernels slide over the signal, generating feature maps that capture frequency information at various scales. This process enables the convolutional layers to extract local frequency characteristics from the signal. Batch normalization helps accelerate the training process, stabilize the model, and prevent issues such as gradient vanishing or explosion. Through this structure, the convolutional layers achieve efficient computation while extracting rich frequency features. By stacking multiple layers, the network progressively refines high-level features of the signal, enhancing its ability to represent complex frequency patterns.
Output Layer: A deconvolution layer (transposed convolution layer) is employed to upsample the internal representation to the desired frequency representation dimensions. The deconvolution layer, a specialized operation in convolutional neural networks, performs the reverse of the convolution operation. While convolution is typically used for downsampling—extracting features and reducing data dimensions using a sliding window—deconvolution is used for upsampling, restoring data dimensions through interpolation and kernel expansion. By upsampling through the deconvolution layer, the output frequency representation retains higher levels of detail, which facilitates subsequent tasks such as frequency estimation and counting. The high-resolution frequency representation produced by the deconvolution layer captures finer frequency variations within the signal, thereby enhancing the accuracy of frequency estimation.
Application of Narrowband Gaussian Kernels: Narrowband Gaussian kernels are utilized to generate the ideal frequency information during model training. The specific steps are as follows:
Define the Gaussian Kernel: A narrowband Gaussian kernel is mathematically defined with a mean frequency and a narrow standard deviation, ensuring it focuses on a specific frequency range.
Generate Ideal Frequency Representation: Using the Gaussian kernel, an ideal frequency representation is constructed by centering it on the target frequencies of the input signal. This representation serves as the ground truth for training the frequency estimation model.
Apply to Training Data: For each input signal, the corresponding ideal frequency representation is computed using the Gaussian kernel, creating the supervised labels required for model training.
Train the Model: The model learns to map the input signal to the ideal frequency representation by minimizing the error between its predictions and the Gaussian-kernel-generated frequency labels. This ensures the model is optimized for precise frequency estimation, even in complex or noisy environments.
This approach leverages the smooth and localized nature of Gaussian kernels, which not only provide a high-resolution reference for training but also enhance the network’s resilience to noise. The kernel function is designed to create an ideal, noise-resilient feature map that aids in more accurate frequency estimation during training.
Loss Function: The loss function utilizes the Mean Squared Error (MSE) to quantify the difference between the predicted frequency representation and the ground truth frequency representation.
The Mean Squared Error (MSE) loss function is an effective metric for quantifying the average deviation between predicted and true frequency values, making it particularly suitable for regression tasks. By minimizing the discrepancy between estimated frequencies and the ground truth, the MSE loss ensures that the network can consistently predict frequencies with outstanding performance. The training data consist of simulated signals with known frequency characteristics, which serve as a foundation for guiding network training.
For each input signal, the corresponding ideal frequency representation is computed using a Gaussian kernel. This process involves applying the kernel with a mean frequency and a narrow standard deviation, thus generating an idealized frequency map that accurately reflects the true frequency characteristics of the signal. These generated frequency labels are then utilized during the training process. The training data are specifically designed to cover a wide range of frequency values and signal-to-noise ratios (SNRs), thereby enabling the network to generalize reliably across various scenarios. To simulate real-world conditions, the dataset includes signals with varying numbers of frequency components and diverse noise levels.
Incorporating the structural design outlined previously, the frequency representation module efficiently extracts relevant frequency features from the input signals, generating high-resolution frequency maps. This architecture is capable of capturing the essential frequency components while also adapting to varying SNRs, ensuring optimal performance across different environments. The use of narrow Gaussian kernels enhances the precision of frequency estimation and significantly improves noise resilience. As a result, this approach delivers superior performance in frequency estimation tasks, particularly in scenarios involving high noise levels and complex signal conditions, where traditional methods may struggle.
3.1.2. Frequency Counting Model
The frequency counting module is designed similarly to the frequency estimation module, with the goal of determining the number of frequencies present in the signal based on its frequency representation. Like the frequency estimation module, it consists of an input layer, multiple convolutional layers, and an output layer. The input layer and convolutional layers share the same structure as those in the frequency estimation module. However, the output layer in the frequency counting module uses a linear layer to convert the internal representation into the final frequency count.
Additionally, when the frequency count is a discrete value, with a finite and predefined set of possible frequency counts, the model adjusts the loss function to the cross-entropy loss function:
3.1.3. Phase Match
Through the frequency estimation and frequency counting models, the initial frequency values
and the frequency count
k are obtained. The estimated frequency parameters are then used to construct the phase-matching function.
This is obtained by multiplying the phase matching function with the phase error signal.
As shown above, when
, the DC component is
. By traversing
over
, the phase parameter is found at the maximum DC component, while half of the DC component gives the amplitude parameter
. A one-dimensional search is sufficient to obtain
. Let
and update the error phase.
Repeat the phase-matching process to obtain the parameter values of the K sinusoidal components. The above process yields the initialization values , serving as the starting point for the subsequent optimization algorithm.
3.1.4. Optimization Algorithm
Since our objective function is nonconvex, we used the ADAM optimization algorithm during the coarse optimization phase of model training to update parameters. Compared to traditional gradient descent methods like SGD, ADAM is more efficient and robust, often achieving better performance in fewer iterations. However, it does not guarantee finding the global optimum. In practice, combining multiple optimizers often yields better results. Therefore, this study employs L-BFGS for fine optimization.
3.1.5. Coarse Optimization—ADAM
ADAM is widely used in deep learning optimization and is an efficient optimization algorithm. Based on the concept of stochastic gradient descent, it ensures relatively low computational cost by utilizing the first-order derivatives of the objective function. ADAM dynamically adjusts the learning rate for each parameter using first- and second-moment estimates, making it highly effective in handling sparse gradients and noisy problems.
The loss function for ADAM is the MSE:
where
represents the optimization parameter
and
is the input data to be fitted. The detailed optimization steps are shown below:
- 1.
Set the first moment estimate variable , second moment estimate variable , learning rate , first moment exponential decay rate , and second moment exponential decay rate ;
- 2.
Calculating the gradient:
where
t represents the current iteration number;
- 3.
Update first-order and second-order moment estimates and compute bias corrections:
- 4.
The parameters are updated using the corrected moment estimates:
After iterating for a certain number of times, the coarse-optimized periodic phase error parameters are obtained. These parameters will then be used for local optimization to obtain more accurate error parameters.
3.2. Precision Optimization—L-BFGS
In the second stage of optimization, the L-BFGS optimization algorithm is employed to further fine-tune the model parameters and improve the fitting accuracy. L-BFGS is a quasi-Newton method suitable for optimization problems requiring high-precision convergence. This method approximates the inverse of the Hessian matrix using limited memory, making it more efficient for large-scale optimization problems. In this paper, the L-BFGS algorithm is used to optimize the amplitude, frequency, and phase parameters, ensuring that the generated signal matches the real signal as closely as possible.
The algorithm steps are as follows:
- 1.
The initialized fitting parameters are set to the final results of the ADAM optimization. A history matrix is initialized to store the gradient and position vectors, with the history matrix size set to 20.
- 2.
Calculate the gradient
g using a method similar to the ADAM algorithm, which will not be elaborated here. Then, update the position vector difference and the gradient difference. These updates are used to record the historical information, which is subsequently employed to approximate the inverse of the Hessian matrix:
Let
, The first loop starts from the most recent historical record and iterates backwards, calculating the scalar
and updating the vector
q:
Calculate the scaling factor
for the initial matrix and initialize the vector
r:
The second loop starts from the furthest historical record and iterates forwards, calculating the scalar
and updating the vector
r:
Determine the search direction
, and refine the parameters.
- 3.
If the norm of the gradient
is smaller than the predefined tolerance
or if the maximum number of iterations is reached, the iteration stops. This completes the precise optimization process. In
Section 4, this paper will validate the algorithm through both simulations and measured data.
3.3. Correction Using Matched Fourier Transform
Existing nonlinear correction methods typically use the time-domain interpolation resampling method to eliminate polynomial nonlinearity and achieve multi-target focusing. However, under periodic nonlinearity, solving transcendental equations to determine interpolation points becomes challenging. To address this, this paper adopts the more versatile matched Fourier transform to achieve multi-target nonlinear correction. With an analytical expression of nonlinearity, multi-target focusing can be effectively performed.
The multi-target differential frequency signal with sinusoidal period phase error can be expressed as follows:
where
Q is the target number and
is the echo delay of the
q target. Its Fourier transform is as follows:
Let
, then the MFT of
is as follows:
where
is the derivative of
, and its expression is as follows:
The targets at different distances are focused into the ideal function, and the frequency of the target reflects the true distance of the target, thereby correcting the system’s nonlinearity.
The periodic amplitude of the new time axis
is as follows:
The transformation from the time axis t to the new time axis is a resampling process. The MFT of the intermediate frequency signal is equivalent to the Fourier transform on the new time axis . The one-dimensional range profile compression on is the result after correcting the periodic phase errors.