4.1. Dataset Description and Experimental Setup
4.1.1. Dataset Description
We use simulated data to validate the effectiveness of the proposed method; the data from reference [
29] was slightly augmented to create a parameter table including 11 waveforms, as shown in
Table 2. The parameter characteristics of different waveforms have significant statistical overlap.
Based on the prepared waveform parameters, radar work mode data simulation is conducted under two MFR signal models, respectively. Signal model 1 establishes a one-to-multiple mapping relationship between work mode and waveform, as shown in
Figure 5a, which uses the MFR two-layer mapping structure described in [
20]. Signal model 2 uses the more comprehensive mapping structure and scheduling law described in
Section 2, where the work mode is mapped to the waveform in a multiple-to-multiple manner through tasks, as shown in
Figure 5b.
We set the antenna scanning model and the spatial scene parameters of radar, target, and reconnaissance receiver, as shown in
Table 3 [
16]. The intercepted MFR pulse sequence can be simulated based on
Table 1 and
Table 3. The pulse sequence includes CF, PW, DTOA, and PA, in which the PA sequence has the antenna scanning characteristics of work mode.
Based on the established waveform parameters, the PDW sequences of work modes are simulated for the two MFR signal models. The MFR intercepted signal is inevitably affected by non-ideal conditions such as low signal noise ratio (SNR) transmission, which will lead to incomplete observation of the pulse sequences, and manifest as measurement error, pulse loss, etc [
30]. Measurement error is simulated by adding Gaussian distribution deviation to the CF, PW, DTOA, and PA parameters of the pulses, as shown in
Table 4. Additionally, we randomly drop
pulses from the pulse train to simulate the pulse lost scenario; the simulated pulse loss rate levels are shown in
Table 5. Further, different degrees of measurement error and pulse loss are combined to simulate complex scenarios;
Table 6 shows the simulation levels of pulse loss and measurement error. Finally, by constructing different sample pulse number distributions, the case where the sample length is not fixed in the interception work mode was simulate, the sample length distribution is shown in
Table 7.
Two sets of train datasets were generated for the two MFR signal models based on the three scenarios in
Table 4,
Table 5 and
Table 6, where each work mode category contains 1800 samples, a sample is a PDW sequence with 4000 to 10,000 pulses.
For two signal models, six test datasets
were generated based on the three scenarios in
Table 4,
Table 5, and
Table 6, respectively, for comparing the impact of different methods under non-ideal conditions across different signal models. Further, in the second signal model condition, the test dataset
is generated under the last complex scenario shown in
Table 6, with the sample length distribution conditions shown in
Table 7. This dataset is used to analyze the impact of sample length, segment length, and shift length on the method’s performance under complex scenario. In the test datasets, each category contains 300 samples.
The original train datasets and test datasets established above are organized into tensors in the shape of as the input for the methods that can handle sequence with variable length and further divided into tensors in the shape of as the input of the other methods.
4.1.2. Experimental Setup
The experimental platform is built on the Win 10 system equipped with an Intel(R) Xeon(R) Gold 6133 CPU @2.50 GHz and an NVIDIA RTX A6000. Simulation data were generated by MATLAB R2022a, and the deep learning network framework was developed using PyCharm 2022.2.3, based on Python 3.9 and PyTorch 1.12.
In the experiment, the CNN block has an input channel count of 4, an output channel count of 32, a convolution kernel size of 7, a convolution stride of 3, a maximum pooling layer count of 2, a pooling kernel size of 3, a stride of 2, and uses the ReLU activation function. The number of input channels of the first Bi-LSTM is 32, the number of layers is 1, and the size of the hidden layer is 32. The number of input channels for the second Bi-LSTM is 64, the number of layers is 1, and the size of the hidden layer is 64. In the attention layer, the number of input channels is 128, and the number of attention heads is 8. The number of input channels for the linear layer is 128, and the number of output channels is 6. Adam is chosen for optimization, with a learning rate of 0.0001, a training batch size of 64, and 80 training epochs.
The accuracy (acc) is used to verify the classification recognition ability. Accuracy is the proportion of correctly identified samples to the total samples, which is defined as follows:
where
is the total number of test samples.
Since the baseline method processing framework only handles segmented pulse segments, a PDW sequence sample of length
obtains
predicted results. Therefore, the majority of the multiple predicted results is taken as the recognition result output of the following work mode sample:
where
is the
-th segment’s predicted result.
4.2. Validation Result
This section presents three experiments to verify the feasibility of the proposed method and its recognition capabilities under different conditions. Experiment 1 is conducted to compare the recognition performance of the proposed method with baseline methods for two MFR signal models. Experiments 2 is conducted only for signal model 2 to verify the impact of sample length distribution, segment length, and shift length on recognition capabilities.
This experiment was carried out on dataset
of the two MFR signal models to verify the recognition performance of the proposed method in non-ideal environments. The baseline methods CNN [
15], bi-GRU [
18], bi-LSTM [
20], and CNN-LSTM were used to validate the dataset generated by the MFR model 1. Additionally, when validating the dataset generated by MFR model 2, the DP-ATCN [
21] method, which can handle long sequence inputs, was further introduced.
Figure 6 shows the recognition results of the signal model 1 dataset under three test scenarios. At a lower pulse loss level, there is a decreasing trend in the accuracy of these methods as the measurement error worsens, but they all can still maintain more than 97% of the category discrimination. Pulse loss can disrupt the temporal distribution of the original PRI parameter sequence with parameter range overlap and complex modulation, while the PW, RF, and PA parameter still have recognition separability within their respective ranges, resulting in a recognition accuracy above 94%. In the complex scenarios with measurement error and pulse loss, the accuracy shows a more obvious downward trend. However, due to the multi-dimensional features expanding separability space and the deep network model extracting effective features from noisy data, the proposed method and various baseline methods can both maintain high recognition accuracy more than 90%. The overall observation results show that under signal model 1, the proposed method is generally superior to the baseline methods. Although it did not show very an outstanding recognition accuracy, this result verifies that our method has competitive recognition performance under the condition of having a simple mapping structure. The superiority of our method is mainly manifested in the recognition under complex waveform mapping models.
Figure 7 shows the recognition results of the signal model 2 dataset under three test scenarios, while
Figure 8 shows the confusion matrices of the five methods for the last scenario in the complex dataset
. Due to its special processing structure, the proposed method can effectively extract dual-scale features, and its recognition performance in the three scenarios is significantly better than the comparison methods. DP-ATCN can also extract waveform scheduling features, but due to its lack of bidirectional causal feature extraction capabilities and the introduction of redundant features with a deeper model, its performance is slightly inferior to the proposed method. However, other waveform-based recognition methods are clearly unsuitable for recognition under this MFR model because they can only extract features within individual pulse segments, which ultimately resulting to poor recognition performances. With the increasing prevalence of non-ideal conditions, the separability of PRIs with parameter range overlap and complex modulation types is poor, while the preserved discriminability of PW, RF, and PA parameters within respective ranges ensure recognition separability. Therefore, the proposed method can maintain over 93% recognition accuracy in three scenarios, and other methods also show relatively stable recognition results. As shown in
Figure 5b, there is waveform reuse between TAS, TWS, and VSR, as well as between STT and MTT. However, the CNN, bi-GRU, bi-LSTM, and CNN-LSTM methods segment work mode samples as network inputs and cause the consistent pulse segment features across different modes, which leads to widespread misidentification. As a result, the overall recognition accuracy of these methods can only be maintained between 70% and 78%. The experimental results indicate that the proposed method has significant advantages when facing radar with complex waveform mapping models.
This experiment is conducted solely on the dataset
, which is based on the MFR signal model 2. The proposed method can handle sample inputs with variable length, and this experiment is used to verify the impact of sample length, segment length, and shift length distribution on its recognition ability. The recognition accuracy at five different sample length distribution levels and five sets of segment length with shift length selections is shown in
Table 8.
The proposed method exhibits differential performance when processing samples with varying sample length distribution. As shown in
Table 8, when the sample length is between 100 and 2000, the recognition performance is quite low. This is because the short samples only represent a small part of the MFR’s spatial scanning and waveform scheduling pulse sequence over a brief period, which is insufficient to reflect the complete work mode of the radar. This ultimately leads to incorrect discrimination between work modes of VSR, TWS, TAS, etc., which use the same waveform. When the sample length is within the other four distributions, the samples contain complete work mode information, and the proposed method can achieve over 90% recognition accuracy in complex scenario. It can also be noted from the
Table 8, as the sample length increases, the accuracy under each segment selection condition both shows a slight decline trend. This is because long samples are divided into more pulse segments, which affects the accurate extraction of features.
The selection of longer segment lengths does not follow with a positive impact on recognition in the complex scenario. As shown in
Table 8, under a fixed sample length, as the pulse segment length and step size increase, recognition performance exhibits a downward trend, which is more pronounced in cases of long samples and long segments. This is because, in the bi-LSTM structure of the inter-segment-scale feature extraction, longer segments require a larger time span to output the hidden state vector of the last time step, which increases the model’s complexity and introduces redundant information. Moreover, through compact segmentation, the local features of PRI sequences with complex modulation can be fully preserved.
This experiment verified that appropriate sample length and pulse segmentation methods could help the network extract more representative features, resulting in better recognition performance.