Next Article in Journal
Local Government Cybersecurity Landscape: A Systematic Review and Conceptual Framework
Previous Article in Journal
Damage Evolution and Fractal Characteristics of Granite under the Influence of Temperature and Loading
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Dual-Component Radar-Signal Modulation Recognition Method Based on CNN-ST

College of Information Science and Engineering, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(13), 5499; https://doi.org/10.3390/app14135499
Submission received: 18 May 2024 / Revised: 14 June 2024 / Accepted: 15 June 2024 / Published: 25 June 2024

Abstract

:
Dual-component radar-signal modulation recognition is a challenging yet significant technique for electronic reconnaissance systems. To improve the lower recognition performance and the higher computational costs of the conventional methods, this paper presents a randomly overlapping dual-component radar-signal modulation recognition method based on a convolutional neural network–swin transformer (CNN-ST) under different signal-to-noise ratios (SNRs). To enhance the feature representation ability and decrease the loss of the detailed features of dual-component radar signals under different SNRs, the swin transformer is adopted and integrated into the designed CNN model. An inverted residual structure and lightweight depthwise convolutions are used to maintain the powerful representational ability. The results show that the dual-component radar-signal recognition accuracy of the proposed CNN-ST is up to 82.58% at −8 dB, which shows the better recognition performance of the CNN-ST over others. The dual-component radar-signal recognition accuracies under different SNRs are all more than 88%, which verified the fact that the CNN-ST achieves better recognition accuracy under different SNRs. This work offers essential guidance in enhancing dual-component radar signal recognition under different SNRs and in promoting actual applications.

1. Introduction

Efficient and high-accuracy radar-signal modulation recognition has emerged as a pivotal requirement in modern electronic warfare [1,2]. With the ever-increasing complicated electromagnetic environment, the interceptors are likely to be affected by the existing multiple potential emitters and result in an overlapping of the radar signals. As the first step for electronic reconnaissance, the precise detection of the signal modulation can provide promising evidence for the subsequent parameter estimation and jamming implementation [3]. Therefore, automatically recognizing the modulation of overlapping radar signals was of great significance, and has become an essential topic in signal processing.
Radar-signal modulation recognition techniques have been investigated in recent decades, and mainly contain radar signal feature extraction and modulation recognition [4,5]. Previous research mainly used the handcrafted feature to train a multi-class classifier of single-component radar signals [6]. However, the conventional extracted features in single-component radar signal recognition, such as high-order moments and cumulants [7] and power spectral density, as well as instantaneous frequency and phase, were inadequate in the circumstance of overlapping radar signals. Moreover, the potential hybrid category was rather random, and the computational features led to poor generalization performance. Because of the enormous advantages of deep learning, multi-class classification algorithms based on deep learning have been proposed for recognizing dual-component radar signals [8,9], which shed light on the capability of highly complicated feature representation [10]. Prior work has largely proved that the extracted representations through deep learning models were more robust and efficient than the handcrafted features of computer vision, which were also potentially applicable to the signal process.
The convolutional neural network (CNN) has been adopted for recognizing related tasks and achieved encouraging performance on radar-signal modulation recognition [11,12,13]. That is because it possessed a robust learning ability feature and higher classification performance compared with the traditional methods, especially for converting radar signals into time-frequency images and automatically extracting various feature details of images. In particular, ResNet [3,14], U-shaped network (U-Net) [15], Asymmetric Convolution Squeeze-and-Excitation [16], and their variants have witnessed remarkable success in radar-signal modulation recognition. ResNet improved the network performance by increasing depth and using residual connections, while it took a lot of computational resources and may overfit. U-Net reduced the size of the feature map by using encoders to superimpose convolution and pooling operations, while it reduced the efficiency and lost spatial information. All of the architecture benefited from a skip connection which incorporated semantic features with fine-grained features. However, the radar-signal modulation recognition method based on CNN makes it difficult to effectively mine the characteristics of dual-component radar signals.
Recently, the transformer has become a deep learning architecture specifically designed to work with sequential data. The main functions include sequence modeling and understanding, efficient parallel computing, and long-distance dependency modeling. Therefore, it has also been utilized in computer vision [17]. In [18], a dynamic transformer was proposed for configuring some tokens, while it would require higher computation requirements. Following this structure, a swin transformer was proposed [19], which computed self-attention within non-overlapping local windows and leveraged the shifted window partition to build connections among the windows of each preceding layer. Although the swin transformer model has made great progress in image classification [20], it has not been utilized in radar-signal modulation recognition.
As can be known from the above references, some methods of radar-signal modulation recognition at the same signal-to-noise ratio (SNR) have been proposed. However, the dual-component radar signal recognition method under different SNRs has not been reported to date. A majority of the proposed radar-signal modulation recognition methods cannot accurately classify dual-component radar signals under different SNRs. The recognition accuracy of radar-signal modulation recognition was relatively lower, particularly under a lower SNR. There is an urgent need to propose an efficient method for recognizing dual-component radar signals under different SNRs and to improve the recognition accuracy. Moreover, for promoting the actual applications in modern electronic warfare, the proposed method would also be used to recognize multi-component radar signals under different SNRs by training the network parameters and designing a multi-label classifier.
Therefore, to solve the above deficiencies, this paper presents a novel network model called CNN–swin transformer (CNN-ST) for recognizing randomly overlapping dual-component radar signals under different SNRs, which integrates the swin transformer into the designed CNN model and improves the recognition accuracy. The important contributions include the following: (1) a novel randomly overlapping dual-component radar signal recognition method under different SNRs based on CNN-ST is first presented; (2) it is inspired by the swin transformer with powerful global modeling capability, which integrates into the CNN; (3) for extracting detailed radar signal features, an inverted residual structure and lightweight depthwise convolutions are adopted; and (4) the proposed CNN-ST model achieves better recognition accuracy under different SNRs.
The related works of radar-signal modulation recognition are reviewed in Section 2; the dual-component radar signals and data preprocessing are explained in detail in Section 3; the novel network CNN-ST for classifying dual-component radar signals under different SNRs is fabricated in Section 4; The influences of dual-component radar signals under different SNRs on the recognition performance are fully investigated in Section 5; and some important conclusions are summarized in Section 6.

2. Related Work

CNN-based methods for radar-signal modulation recognition are first summarized, and then an overview of recent applications of transformers in computer vision is provided, especially image recognition.

2.1. Radar Signal Modulation-Recognition Methods Based on CNN

Owing to the rapid development of CNN in recent decades, the recognition methods of radar signals based on CNN have been proposed [21,22,23]. In [24], a CNN-based radar signal modulation-recognition technique was proposed, which offered significant improvement over the recent radar-signal modulation recognition technology. In [25], a generic training strategy improved the prediction performance of radar-signal modulation recognition. In [26], a radar emitter signal classification algorithm based on CNN was presented, and the simulation results indicated superior recognition accuracy. In [27], a feature fusion algorithm for radar signal automatic modulation recognition using CNN was proposed, and the simulation results revealed superior accuracy. In [28], a cost-efficient CNN for radar-signal modulation recognition was proposed, and the results demonstrated excellent generalization ability. In [29], a CNN for low-complexity and robust modulation recognition was proposed, which achieved better recognition performance. In [30], a CNN-LSTM to address automatic modulation recognition was proposed, and experimental results demonstrated superior recognition performance. In [31], a multi-class learning framework based on CNN under the same SNR was proposed, and the results demonstrated superior performance over others. In [32], an efficient deep CNN with feature fusion at low SNR was presented, and recognition performance was up to 84.38% at −12 dB. Although the CNN-based network has demonstrated better robustness in the single-component radar signal, it cannot be applied completely to dual-component radar signals. Because of the existing difference in time-frequency images and features, the CNN-based network may demonstrate poor recognition performance for dual-component radar signals.

2.2. Vision Transformer

The transformer block is composed of a multi-head self-attention (MSA), multiple-layer perceptron (MLP), and layer normalization (LN) module [33]. Recent research has been trying to investigate the benefits of transformers in computer vision. In [17], the transformer first applied to image classification for replacing the CNN model was proposed. Chen et al. [34] presented an ImageNet benchmark, and maximally excavated the capability of the transformer. Kolesnikov et al. [35] conducted image recognition with a transformer structure. Wu et al. [36] dynamically extracted visual tokens, and used visual transformers to perform the visual tokens. Jiang et al. [37] conducted the first pilot study in building a generative adversarial network model using only pure transformer-based architecture. Yuan et al. [38] designed a novel vision transformer, which reduced the parameter count. Touvron et al. [39] produced the transformer structure and recognition accuracy achieved 83.1% on the ImageNet dataset. Wang et al. [40] introduced the pyramid vision transformer. Liu et al. [19] designed a swin transformer based on the shifted-window strategy. Following that, Zheng et al. [20] proposed the swin transformer and multi-layer perceptron (ST-MLP) to classify the strawberry appearance, and the results demonstrated superior recognition performance.
The CNN-based methods have improved radar-signal modulation recognition performance. Inspired by these excellent works, we adopt the swin transformer blocks to obtain the global context information for the CNN model. In this paper, a new framework including CNN and the swin transformer (CNN-ST) is proposed. CNN-ST first applies the swin transformer to the radar-signal modulation recognition field, which increases the recognition performance, especially for dual-component radar-signal modulation recognition under different SNRs.

3. Signal Model and Time-Frequency Transformation

The dual-component radar signals under different SNRs are adopted, and the time-frequency transformation technology is implemented to obtain the detailed features and classify the radar signal types.

3.1. Dual-Component Radar Signals under Different SNRs

For recognizing intra-pulse modulation radar signals, even quadratic frequency modulation (EQFM), linear frequency modulation (LFM), normal signal (NS), binary phase-shift keying (BPSK), binary frequency-shift keying (2FSK), sinusoidal frequency modulation (SFM), FRANK, and four frequency-shift keying (4FSK) are adopted in this paper. The Gaussian white noise (GWN) is utilized to disturb those signals, for simulating the realistic application environment in the actual battlefield. The received dual-component radar signal is written as [31]
y t = i = 1 k A i r e c t t / T i e j ( 2 π f c i t + i t + 0 i ) + n t     ( k = 1 ,   2 )
where n t is i-th noise component; k represents signal number; and A i , f c i , 0 i , and T i , denote amplitude, carrier frequency, initial phase, and pulse width of the signals, respectively.
This paper is mainly intended to classify dual-component signals under different SNRs, wherein dual-component signals are obtained by randomly overlapping two single-component signals, and thus k is set as 2. Given the received overlapping-signal sample dataset D = ( x i , t i ) 1 i N including N samples with eight types of signals, the i-th sample is represented as x i , and t i j = [ t i 1 , t i 2 , , t i 8 ] represents the true label vector x i , such that sample x with a label vector t = [ 0 , 1 , 0 , 0 , 0 , 0 , 1 , 0 ] indicates the presence of radar signals at the second and seventh positions, which are overlapped in x .

3.2. Time-Frequency Transformation

The common time-frequency transformation methods include the short-time Fourier Transform (STFT), the Wigner–Ville distribution (WVD), the Choi–Williams distribution (CWD), the pseudo Wigner–Ville distribution (PWVD), the smoothed pseudo Wigner–Ville distribution (SPWVD), and so on [41,42]. The SPWVD can express the radar signals in detail and effectively prevent cross-interference over other methods, and is adopted in this paper [43]. The SPWVD transformation is written as
S t , f = x ( t μ + τ / 2 ) · x * t ν τ / 2 h τ g μ e j 2 π f τ d μ d τ
where is the complex conjugate; h ( τ ) and g ( μ ) represent window functions; and x ( t ) is the analytic signal y ( t ) .
Through SPWVD time-frequency transformation, 2184 types of dual-component radar signals under different SNRs are acquired. To save space, Figure 1 only illustrates four types of time-frequency images (TFIs) of dual-component radar signals under different SNRs. Therein, Figure 1a represents the overlap of EQFM and SFM at −2 dB; Figure 1b represents the overlap of EQFM at −2 dB and SFM at 12 dB; Figure 1c represents the overlap of EQFM at 12 dB and SFM at −2 dB; and Figure 1d represents the overlap of EQFM and SFM at 12 dB.
Figure 1 demonstrates that SNR exerts an essential effect on TFIs. The higher SNR results in obtaining clearer TFIs. It means that the more detailed features are extracted for the clearer TFIs, and better classification performance can be achieved in the following analyses.

4. Network Model CNN-ST

The network model CNN-ST is proposed in this paper, for a detailed extraction of the features of signals and for accurately recognizing the signal types. Firstly, the model framework of CNN-ST is described, and the motivation for choosing the CNN-ST model is explained. Next, the network model CNN-ST is designed. Finally, the swin transformer is utilized for enhancing the feature representation ability and reducing the loss of detailed information of dual-component radar signals under different SNRs.

4.1. Model Framework

To accurately classify randomly overlapping dual-component radar signals under different SNRs, this paper presents a novel network model CNN-ST, which possesses robust feature-extraction capability. To improve the focusing and information expression ability of dual-component radar signals under different SNRs, the network architecture based on CNN and the swin transformer is designed for extracting in detail the radar signal features. Figure 2 shows the network architecture of the CNN-ST model.
Figure 2 demonstrates that TFIs obtained by time-frequency transformation are first entered into the designed CNN-ST model, the radar signal features are then extracted in detail, and the signal types are finally classified. The CNN-ST model consists of the first convolutional layer, an average pooling layer, three bottleneck convolution blocks, four swin transformer blocks, a second convolutional layer, a nonlinearity layer Hard swish (H-swish), a third convolutional layer, an average pooling layer, an 8-dimension fully connected layer, and a multi-label classifier. Within this, softmax is the activation function of the multi-label classifier. The bottleneck convolution block aims to increase expressive ability. The expansion ratio changes the output channel number. The swin transformer block possesses powerful global modeling capability. For the deepening of the network, the cost of applying the nonlinear activation function H-swish could be reduced, and could largely decrease the number of parameters. Table 1 lists the CNN-ST parameters.

4.2. CNN Structure

To extract in detail the radar signal features, the CNN model is designed. Additionally, to maintain the representational ability, the designed CNN removes the nonlinearity in the narrow layers. The structure of CNN in CNN-ST is illustrated in detail in Figure 3.
Figure 3 demonstrates that the designed CNN includes the following: the initial fully convolutional layer; an average pooling layer; two bottleneck convolutional blocks, where the expansion ratio is 1, and the stride length is 1; and a bottleneck convolution block, where the expansion ratio is 6, and the stride length is 2. Within this, the bottleneck convolution block with a stride length of 1 contains the following: a convolutional layer, a batch normalization, an activation function H-swish, an average pooling layer in which a kernel size is 2 × 2 and a stride length is 2, a convolutional layer in which the kernel size is 1 × 1 and the stride length is 1, an activation function H-swish, a depthwise convolutional layer in which t the kernel size is 3 × 3 and the stride length is 2, an activation function ReLU6, a convolutional layer in which the kernel size is 1 × 1 and the stride length is 2, a batch normalization, and a dropout. Finally, a residual connection is utilized. The expansion ratio of the bottleneck convolution block is set as 1. The bottleneck convolution block with a stride length of 2 is the same as the one with a stride length of 1, except that there is no residual connection.
Depthwise separable convolutions are an essential component of many deep CNN architectures, and are adopted in this paper. A nonlinearity swish was introduced in [44,45], which was utilized as a replacement for ReLU and improved the classification performance of the dual-component radar signals under different SNRs. The nonlinearity is written as
s w i s h x = x · σ x
Nonlinearity brings in the non-zero cost and improves recognition accuracy. This is because it is much more expensive to calculate the sigmoid functions on mobile devices. The sigmoid function R e L U 6 x + 3 6 is the same as in [46]. The small difference is that ReLU6 is used. Recently, a similar H-swish was proposed in [47], and the hard version of H-swish is defined as
H s w i s h x = x R e L U 6 x + 3 6

4.3. Swin Transformer Block

To improve the feature representation performance and decrease the loss of detailed information of dual-component radar signals under different SNRs, this paper integrates the swin transformer into the designed CNN model, which possesses powerful global modeling capability [48]. Figure 4 illustrates the swin transformer structure in detail.
Figure 4 demonstrates that the swin transformer structure mainly includes LN, W-MSA, MLP, MSA, and SW-MSA, and a residual connection is used after each module, wherein the LN layer is added before each W-MSA and SW-MSA. Therefore, the output s l of the l-th layer in the swin transformer is written as
s ^ l = W M S A L N s l 1 + s l 1
s l = M L P L N s ^ l + s ^ l
where s l and s ^ l represent the outputs of W-MSA and MLP. In W-MSA, the input is a patches sequence, s l 1 R L × D .
The outputs of SW-MSA and MLP modules are expressed as
s ^ l + 1 = S W M S A L N s l + s l
s l + 1 = M L P L N s ^ l + 1 + s ^ l + 1
where s ^ l + 1 and s l + 1 refer to the outputs of SW-MSA and MLP, respectively. The self-attention computation in W-MSA and SW-MSA is defined as
A t t e n t i o n z l = S o f t M a x Q K T / d + B V
Q = s l W Q ,   K = s l W K ,   V = s l W V
where W Q , W K , and W V R D × d represent the three projection matrices’ parameters; Q, K, and V R L × d represent the query, key, and value matrices, respectively; d denotes the dimension of query or key; and B R L × L represents the relative position bias.

5. Experiment, Results, and Analysis

The recognized performance of CNN-ST for randomly overlapping dual-component radar signals under different SNR is evaluated. To demonstrate better classification performance over others, CNN-DQLN and CNN-Softmax in [49] are also utilized for conducting comparative analyses.

5.1. Datasets and Training Parameters

The typical eight types of signals that include 2FSK, BPSK, 4FSK, FRANK, NS, EQFM, LFM, and SFM are adopted in this paper. Table 2 lists the detailed radar signal parameters. Note that the training, validation, and testing datasets contain randomly overlapping dual-component radar signals under different SNRs, which are preprocessed by using SPWVD transformation. To conduct the comparative analyses, the SNR ranges from −12 dB to 10 dB, and every 2 dB is selected. The number of each class of dual-component radar signals under different SNRs is set as 300, and a total of 1,419,600 samples are obtained in the training dataset. For the validation dataset, the number of each class of dual-component radar signals under different SNRs is set as 100, and a total of 473,200 samples were obtained. For the testing dataset, the number of each class of dual-component radar signals under different SNRs is set as 30, and a total of 5160 samples is obtained. The types and parameters of radar signals are the same as those in Reference [49].
The deep learning framework used in the analyses is Pytorch 1.11 and Python 3.9. Based on the computational complexity and recognition performance, the training parameters are set as follows: the batch size is 128; epoch is 100; epsilon is 0.001; momentum is 0.01; initial learning rate is set as 0.01; learning decay rate is 0.1 every 10 epochs; dropout is 0.8; weight decay is set as 1 × 10−5; and batch normalization with average decay is set as 0.999.

5.2. Results, Discussions, and Analysis

To explicitly explore the recognition accuracy of CNN-ST and conduct the comparative analyses, Figure 5 illustrates the overall recognition accuracies of CNN-ST, CNN-DQLN, and CNN-Softmax as a function of SNR. Therein, the testing dataset used in this subsection represents the dual-component radar signals under the same SNR, and the SNR used ranges from −10 dB to 10 dB.
Figure 5 demonstrates that an increase in SNR leads first to increasing overall recognition accuracies of CNN-ST, CNN-DQLN, and CNN-Softmax, and then leveling off. Therein, the increasing trend for CNN-ST at the lower SNR is faster than CNN-DQLN and CNN-Softmax, and achieves better recognition performance over others. The recognition performance of CNN-ST, CNN-DQLN, and CNN-Softmax are 50.88%, 37.37%, and 26.49%, at −10 dB, respectively. The recognition performance of CNN-ST is 13.51% higher than that of CNN-DQLN, and 24.39% higher than that of CNN-Softmax. The recognition accuracy of CNN-ST is 82.58% at −8 dB, while it is 74.74% and 57.02% for CNN-DQLN and CNN-Softmax, respectively. Moreover, the recognition accuracies can reach 100%, 98.77%, and 95.44% at 4 dB for CNN-ST, CNN-DQLN, and CNN-Softmax, respectively. Therefore, CNN-ST achieves superior recognition accuracy compared to CNN-DQLN and CNN-Softmax. The reason is that the adopted swin transformer in the CNN-ST can extract the more detailed features of the radar signals.
To clearly express the recognition performance of each type of radar signal, Figure 6 illustrates the variation in the recognition accuracy of eight types of signals with SNR.
Figure 6 demonstrates that the recognition accuracies of eight types of signals increase with increasing SNR, while the increasing trend levels off at the higher SNR. It means that the lower SNR exerts an important effect on recognition accuracy. Moreover, 2FSK and LFM demonstrate a superior recognition performance over the others. That is because the clearer the TFIs for 2FSK and LFM are, the easier the detailed features are extracted, and the better the recognition performance which can be obtained. The recognition accuracies of 2FSK, EQFM, 4FSK, BPSK, LFM, SFM, NS, and FRANK are 90.83%, 82.14%, 77.98%, 67.62%, 80.71%, 75.95%, 73.45%, and 71.91%, at SNR of −10 dB, respectively. Moreover, the recognition accuracies of all types of radar signals are basically up to 1 at larger than 0 dB. Therefore, the designed CNN-ST model demonstrates better recognition performance at the same SNR.
To investigate the influence of different SNRs on the recognition accuracy of dual-component signals, Figure 7 demonstrates the recognition accuracy as a function of different SNRs. Therein, the SNR is −12 dB to 4 dB; the X and Y axes represent the SNR under different radar signals; and the Z axis represents the classification accuracy. The dual-component radar signals include the following: overlapping 2FSK and 4FSK (2FSK-4FSK), overlapping 2FSK and BPSK (2FSK-BPSK), overlapping 2FSK and EQFM (2FSK-EQFM), overlapping 2FSK and FRANK (2FSK-FRANK), overlapping 2FSK and LFM (2FSK-LFM), overlapping 2FSK and NS (2FSK-NS), overlapping 2FSK and SFM (2FSK-SFM), overlapping LFM and 4FSK (LFM-4FSK), overlapping LFM and BPSK (LFM-BPSK), overlapping LFM and EQFM (LFM-EQFM), overlapping LFM and FRANK (LFM-FRANK), and overlapping LFM and SFM (LFM-SFM).
Figure 7 shows that the recognition accuracies of twelve types of randomly overlapping dual-component radar signals increase as different SNRs increase. When the different SNRs of the dual-component radar signal are higher than −2 dB, the increasing trend gradually levels off, and the recognition accuracies are up to 1. It means that the different SNRs of the dual-component radar signal exert less effect on the recognition accuracy, especially at the higher SNR. That is because the higher the different SNR is, the less the noise interferes with the TFIs is, and the easier it is for the effective features of the TFIs to be extracted. The recognition accuracies of 2FSK-4FSK, 2FSK-BPSK, 2FSK-EQFM, 2FSK-FRANK, 2FSK-LFM, 2FSK-NS, 2FSK-SFM, LFM-4FSK, LFM-BPSK, LFM-EQFM, LFM-FRANK, and LFM-SFM are up to 76.25%, 63.75%, 82.625%, 75.625%, 70.625%, 72%, 67.5%, 71.5%, 60.625%, 78.25%, 73.75%, and 70.625%, at different SNRs of −12 dB, respectively. When the SNR of 2FSK is −12 dB and of 4FSK it is 4 dB, the recognition accuracy of 2FSK-4FSK is up to 88.56%. For 2FSK at 4 dB and 4FSK at −6 dB, the recognition accuracy of 2FSK-4FSK is 1. When SNR of 2FSK is −4 dB and 4FSK is −2 dB, the recognition accuracy of 2FSK-4FSK is up to 1. For 2FSK at −6 dB and 4FSK at −8 dB, the recognition accuracy of 2FSK-4FSK is 93.125%. When SNR of 2FSK is −12 dB and BPSK is 4 dB, the recognition accuracy of 2FSK-BPSK is up to 81.875%. For 2FSK at 4 dB and BPSK at −6 dB, the recognition accuracy of 2FSK-BPSK is 98.75%. When SNR of 2FSK is −4 dB and BPSK is −2 dB, the recognition accuracy of 2FSK-BPSK is up to 96.125%. For 2FSK at −6 dB and BPSK at −8 dB, the recognition accuracy of 2FSK-BPSK is 85.25%. When SNR of 2FSK is −12 dB and EQFM is 4 dB, the recognition accuracy of 2FSK-EQFM is up to 92.625%. For 2FSK at 4 dB and EQFM at −6 dB, the recognition accuracy of 2FSK-EQFM is 1. When SNR of 2FSK is −4 dB and EQFM is −2 dB, the recognition accuracy of 2FSK-EQFM is up to 1. For 2FSK at −6 dB and EQFM at −8 dB, the recognition accuracy of 2FSK-EQFM is 96.5%. Note that the recognition accuracies of randomly overlapping dual-component signals are all more than 90% at −2 dB, which demonstrates better recognition performance. The reason is the powerful feature extraction and classification ability. Therefore, this work offers important experimental guidance in further enhancing recognition performance under different SNRs and promoting the actual applications.
The floating point of operations (FLOPs), parameters, and time are adopted to investigate computational complexity. Table 3 lists the computational complexity of CNN-ST and CNN-DQLN.
As can be seen from Table 3, the FLOPs, parameters, and the time of CNN-ST are all lower than CNN-DQLN. Therefore, the proposed CNN-ST demonstrates higher computational accuracy and lower computational complexity over others, and possesses enough computational novelty.

6. Conclusions

This paper presented a novel randomly overlapped dual-component radar signal recognition method under different SNRs based on a convolutional neural network–swin transformer (CNN-ST), for improving recognition performance. The overall model framework was first designed, and the swin transformer was subsequently adopted and integrated into the CNN model. An inverted residual structure and lightweight depthwise convolutions were used to maintain the powerful representational ability. The influences of the different SNRs on the recognition performance of dual-component radar signals were experimentally investigated. The results demonstrated that recognition accuracies of CNN-ST were up to 82.58% at −8 dB, which showed better recognition performance over others. The recognition accuracies of randomly overlapping dual-component radar signals under different SNRs were all more than 90% at −2 dB, which verified the superior recognition performance of the proposed CNN-ST model. The recognition accuracies of 2FSK-4FSK, 2FSK-BPSK, 2FSK-EQFM, 2FSK-FRANK, 2FSK-LFM, 2FSK-NS, 2FSK-SFM, LFM-4FSK, LFM-BPSK, LFM-EQFM, LFM-FRANK, and LFM-SFM are up to 76.25%, 63.75%, 82.625%, 75.625%, 70.625%, 72%, 67.5%, 71.5%, 60.625%, 78.25%, 73.75%, and 70.625% at SNR of −12 dB, respectively. This work provided essential guidance in enhancing recognition performance under different SNRs and promoting the actual applications.
For considering the more complex scenarios with randomly overlapping radar signals, future research will focus on classifying multi-component signals under different SNRs and promoting the actual applications in modern electronic warfare.

Author Contributions

C.W.: conceptualization, validation, writing—original draft. Q.Z.: resources, funding acquisition, writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was financially supported by the National Natural Science Foundation of China (Grant No. 62073123), the Key Research & Development and Promotion Project of Henan Province (Grant No. 242102211002), and the High-Level Talent Research Start-up Fund Project of Henan University of Technology (2023BS040).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Nomenclature

SNRSignal-to-noise ratio
CNNConvolutional neural network
CNN-STConvolutional neural network–swin transformer
CNN-DQLNConvolutional neural network and Deep Q-Learning Network
CNN-SoftmaxConvolutional neural network with fully connected Softmax layers
ResNetResidual neural network
EQFMEven quadratic frequency modulation
LFMLinear frequency modulation
NSNormal signal
BPSKBinary phase-shift keying
2FSKBinary frequency-shift keying
SFMSinusoidal frequency modulation
4FSKFour frequency-shift keying
GWNGaussian white noise
STFTShort-Time Fourier Transform
WVDWigner–Ville Distribution
CWDChoi–Williams Distribution
PWVDPseudo Wigner–Ville Distribution
SPWVDSmoothed Pseudo Wigner–Ville Distribution

References

  1. Meng, F.; Chen, P.; Wu, L.; Wang, X. Automatic modulation classification: A deep learning enabled approach. IEEE Trans. Veh. Technol. 2018, 67, 10760–10772. [Google Scholar] [CrossRef]
  2. Qu, Q.Z.; Wei, S.J.; Liu, S.; Liang, J.D.; Shi, J. Jrnet: Jamming recognition networks for radar compound suppression jamming signals. IEEE Trans. Veh. Technol. 2020, 69, 15035–15045. [Google Scholar] [CrossRef]
  3. Qi, P.H.; Zhou, X.Y.; Zheng, S.L.; Li, Z. Automatic modulation classification based on deep residual networks with multimodal information. IEEE Trans. Cogn. Commun. Netw. 2021, 7, 21–33. [Google Scholar] [CrossRef]
  4. Kishore, T.R.; Rao, K.D. Automatic intrapulse modulation classification of advanced lpi radar waveforms. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 901–914. [Google Scholar] [CrossRef]
  5. Si, W.; Wan, C.; Zhang, C. Towards an accurate radar waveform recognition algorithm based on dense cnn. Multimed. Tools Appl. 2021, 80, 1779–1792. [Google Scholar] [CrossRef]
  6. Wu, G.R.; Kim, M.J.; Wang, Q.; Munsell, B.C.; Shen, D. Scalable high-performance image registration framework by unsupervised deep feature representations learning. IEEE Trans. Biomed. Eng. 2017, 64, 250. [Google Scholar] [CrossRef]
  7. Huang, S.; Yao, Y.; Wei, Z.; Feng, Z.; Zhang, P. Automatic modulation classification of overlapped sources using multiple cumulants. IEEE Trans. Veh. Technol. 2017, 66, 6089–6101. [Google Scholar] [CrossRef]
  8. Huang, S.; Jiang, Y.; Qin, X.; Gao, Y.; Feng, Z.; Zhang, P. Automatic modulation classification of overlapped sources using multi-gene genetic programming with structural rick minimization principle. IEEE Access 2018, 6, 48827–48839. [Google Scholar] [CrossRef]
  9. Gao, J.P.; Shen, L.X.; Gao, L.P. Modulation recognition for radar emitter signals based on convolutional neural network and fusion features. Trans. Emerg. Telecommun. Technol. 2019, 30, e3612. [Google Scholar] [CrossRef]
  10. Yu, Z.Y.; Tang, J.L.; Wang, Z. Gcps: A cnn performance evaluation criterion for radar signal intrapulse modulation recognition. IEEE Commun. Lett. 2021, 25, 2290–2294. [Google Scholar] [CrossRef]
  11. Huynh-The, T.; Doan, V.S.; Hua, C.H.; Pham, Q.V.; Nguyen, T.V.; Kim, D.S. Accurate lpi radar waveform recognition with cwd-tfa for deep convolutional network. IEEE Wirel. Commun. Lett. 2021, 10, 1638–1642. [Google Scholar] [CrossRef]
  12. Liu, L.T.; Li, X.Y. Unknown radar waveform recognition system via triplet convolution network and support vector machine. Digit. Signal Process. 2022, 123, 103439. [Google Scholar] [CrossRef]
  13. Zhang, X.L.; Zhang, J.Z.; Luo, T.Z.; Huang, T.Y.; Tang, Z.P.; Chen, Y.; Li, J.S.; Luo, D.P. Radar signal intrapulse modulation recognition based on a denoising-guided disentangled network. Remote Sens. 2022, 14, 1252. [Google Scholar] [CrossRef]
  14. Hong-hai, Y.; Xiao-peng, Y.; Shao-kun, L.; Ping, L.; Xin-hong, H. Radar emitter multi-label recognition based on residual network. Def. Technol. 2022, 18, 410–417. [Google Scholar] [CrossRef]
  15. Jiang, W.K.; Li, Y.; Liao, M.M.; Wang, S.F. An improved lpi radar waveform recognition framework with ldc-unet and ssr-loss. IEEE Signal Process. Lett. 2022, 29, 149–153. [Google Scholar] [CrossRef]
  16. Wei, S.; Qu, Q.; Wang, M.; Wu, Y.; Shi, J. Automatic modulation recognition for radar signals via multi-branch acse networks. IEEE Access 2020, 8, 94923–94935. [Google Scholar] [CrossRef]
  17. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, La Jolla, CA, USA, 5–7 May 2021; pp. 1–22. [Google Scholar]
  18. Wang, Y.; Huang, R.; Song, S.; Huang, Z.; Huang, G. Not all images are worth 16 × 16 words: Dynamic vision transformers with adaptive sequence length. In Proceedings of the 35th Conference on Neural Information Processing Systems, Sydney, Australia, 6–14 December 2021; pp. 11960–11973. [Google Scholar]
  19. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the 18th IEEE/CVF International Conference on Computer Vision, Virtual, Online, Canada, 11–17 October 2021; pp. 1–14. [Google Scholar]
  20. Zheng, H.; Wang, G.H.; Li, X.C. Swin-mlp: A strawberry appearance quality identification method by swin transformer and multi-layer perceptron. J. Food Meas. Charact. 2022, 16, 2789–2800. [Google Scholar] [CrossRef]
  21. Guo, Q.; Yu, X.; Ruan, G. Lpi radar waveform recognition based on deep convolutional neural network transfer learning. Symmetry 2019, 11, 540. [Google Scholar] [CrossRef]
  22. Shengliang, P.; Hanyu, J.; Huaxia, W.; Hathal, A.; Yu, Z.; Mazrouei, S.M.; Yu-Dong, Y. Modulation classification based on signal constellation diagrams and deep learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 718–727. [Google Scholar]
  23. Wan, J.; Yu, X.; Guo, Q. Lpi radar waveform recognition based on cnn and tpot. Symmetry 2019, 11, 725. [Google Scholar] [CrossRef]
  24. Kong, S.-H.; Kim, M.; Linh Manh, H.; Kim, E. Automatic lpi radar wave form recognition using cnn. IEEE Access 2018, 6, 4207–4219. [Google Scholar] [CrossRef]
  25. Oktay, O.; Ferrante, E.; Kamnitsas, K.; Heinrich, M.; Bai, W.J.; Caballero, J.; Cook, S.A.; de Marvao, A.; Dawes, T.; O’Regan, D.P.; et al. Anatomically constrained neural networks (acnns): Application to cardiac image enhancement and segmentation. IEEE Trans. Med. Imaging 2018, 37, 384–395. [Google Scholar] [CrossRef] [PubMed]
  26. Wang, F.; Yang, C.; Huang, S.; Wang, H. Automatic modulation classification based on joint feature map and convolutional neural network. IET Radar Sonar Navig. 2019, 13, 998–1003. [Google Scholar] [CrossRef]
  27. Zhang, Z.; Wang, C.; Gan, C.; Sun, S.; Wang, M. Automatic modulation classification using convolutional neural network with features fusion of spwvd and bjd. IEEE Trans. Signal Inf. Process. Over Netw. 2019, 5, 469–478. [Google Scholar] [CrossRef]
  28. Huynh-The, T.; Hua, C.H.; Pham, Q.V.; Kim, D.S. Mcnet: An efficient cnn architecture for robust automatic modulation classification. IEEE Commun. Lett. 2020, 24, 811–815. [Google Scholar] [CrossRef]
  29. Tunze, G.B.; Huynh-The, T.; Lee, J.-M.; Kim, D.-S. Sparsely connected cnn for efficient automatic modulation recognition. IEEE Trans. Veh. Technol. 2020, 69, 15557–15568. [Google Scholar] [CrossRef]
  30. Zhang, Z.; Luo, H.; Wang, C.; Gan, C.; Xiang, Y. Automatic modulation classification using cnn-lstm based dual-stream structure. IEEE Trans. Veh. Technol. 2020, 69, 13521–13531. [Google Scholar] [CrossRef]
  31. Si, W.J.; Wan, C.X.; Deng, Z. Intra-pulse modulation recognition of dual-component radar signals based on deep convolutional neural network. IEEE Commun. Lett. 2021, 25, 3305–3309. [Google Scholar] [CrossRef]
  32. Si, W.J.; Wan, C.X.; Deng, Z.A. An efficient deep convolutional neural network with features fusion for radar signal recognition. Multimed. Tools Appl. 2022, 82, 2871–2885. [Google Scholar] [CrossRef]
  33. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; pp. 1–15. [Google Scholar]
  34. Chen, H.T.; Wang, Y.H.; Guo, T.Y.; Xu, C.; Deng, Y.P.; Liu, Z.H.; Ma, S.W.; Xu, C.J.; Xu, C.; Gao, W.; et al. Pre-trained image processing transformer. In Proceedings of the 18th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, Online, Canada, 11–17 October 2021; pp. 12294–12305. [Google Scholar]
  35. Kolesnikov, A.; Beyer, L.; Zhai, X.; Puigcerver, J.; Yung, J.; Gelly, S.; Houlsby, N. Big transfer (bit): General visual representation learning. In Proceedings of the 16th European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 491–507. [Google Scholar]
  36. Wu, B.; Xu, C.; Dai, X.; Wan, A.; Zhang, P.; Tomizuka, M.; Keutzer, K.; Vajda, P. Visual transformers: Token-based image representation and processing for computer vision. arXiv 2020, arXiv:20200559351. [Google Scholar]
  37. Jiang, Y.; Chang, S.; Wang, Z. Transgan: Two transformers can make one strong gan. In Proceedings of the 35th Conference on Neural Information Processing Systems, Online, 6–14 December 2021; pp. 14745–14758. [Google Scholar]
  38. Yuan, L.; Chen, Y.; Wang, T.; Yu, W.; Shi, Y.; Tay, F.E.; Feng, J.; Yan, S. Tokens-to-token vit: Training vision transformers from scratch on imageNet. In Proceedings of the 18th IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, Online, Canada, 11–17 October 2021; pp. 558–567. [Google Scholar]
  39. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jegou, H. Training data-efficient image transformers & distillation through attention. In Proceedings of the International Conference on Machine Learning (ICML), Electr Network, 18–24 July 2021; pp. 7358–7367. [Google Scholar]
  40. Wang, W.H.; Xie, E.Z.; Li, X.; Fan, D.P.; Song, K.T.; Liang, D.; Lu, T.; Luo, P.; Shao, L.; IEEE. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the 18th IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, Online, Canada, 11–17 October 2021; pp. 548–558. [Google Scholar]
  41. El Khadiri, K.; Elouaham, S.; Nassiri, B.; El Melhoaui, O.; Said, S.; El Kamoun, N.; Zougagh, H. A comparison of the denoising performance using capon time-frequency and empirical wavelet transform applied on biomedical signal. Int. J. Eng. Appl. 2023, 11, 358–365. [Google Scholar] [CrossRef]
  42. Dliou, A.; Latif, R.; Laaboubi, M.; Maoulainine, F.; Elouaham, S. Noised abnormal ECG signal analysis by combining EMD and Choi-Williams techniques. In Proceedings of the 2012 IEEE International Conference on Complex Systems, Agadir, Morocco, 5–6 November 2012; pp. 1–5. [Google Scholar]
  43. Ma, N.; Wang, J. Dynamic threshold for spwvd parameter estimation based on otsu algorithm. J. Syst. Eng. Electron. 2013, 24, 919–924. [Google Scholar] [CrossRef]
  44. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  45. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for activation functions. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018; pp. 1–13. [Google Scholar]
  46. Courbariaux, M.; Bengio, Y.; David, J.P. Binaryconnect: Training deep neural networks with binary weights during propagations. In Proceedings of the International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 1–9. [Google Scholar]
  47. Avenash, R.; Viswanath, P. Semantic segmentation of satellite images using a modified cnn with hard-swish activation function. In Proceedings of the 14th International Conference on Computer Vision Theory and Applications, Barcelona, Spain, 25–27 February 2019; pp. 413–420. [Google Scholar]
  48. Lin, A.; Chen, B.; Xu, J.; Zhang, Z.; Lu, G.; Zhang, D. Ds-transunet: Dual swin transformer u-net for medical image segmentation. IEEE Trans. Instrum. Meas. 2022, 71, 4005615. [Google Scholar] [CrossRef]
  49. Qu, Z.; Hou, C.; Hou, C.; Wang, W. Radar signal intra-pulse modulation recognition based on convolutional neural network and deep q-learning network. IEEE Access 2020, 8, 49125–49136. [Google Scholar] [CrossRef]
Figure 1. SPWVD transformation of four types of dual-component signals under different SNRs. (a). EQFM (−2 dB)-SFM (−2 dB); (b) EQFM (−2 dB)-SFM (12 dB); (c) EQFM (12 dB)-SFM (−2 dB); (d) EQFM (12 dB)-SFM (12 dB).
Figure 1. SPWVD transformation of four types of dual-component signals under different SNRs. (a). EQFM (−2 dB)-SFM (−2 dB); (b) EQFM (−2 dB)-SFM (12 dB); (c) EQFM (12 dB)-SFM (−2 dB); (d) EQFM (12 dB)-SFM (12 dB).
Applsci 14 05499 g001
Figure 2. Model framework of CNN-ST.
Figure 2. Model framework of CNN-ST.
Applsci 14 05499 g002
Figure 3. Structure of CNN model.
Figure 3. Structure of CNN model.
Applsci 14 05499 g003
Figure 4. Structure of swin transformer block.
Figure 4. Structure of swin transformer block.
Applsci 14 05499 g004
Figure 5. Overall recognition accuracies of CNN-ST, CNN-DQLN, and CNN-Softmax as a function of SNR.
Figure 5. Overall recognition accuracies of CNN-ST, CNN-DQLN, and CNN-Softmax as a function of SNR.
Applsci 14 05499 g005
Figure 6. Variation in the recognition performance of eight types of signals with SNR. (a) 2FSK, EQFM, 4FSK, BPSK; (b) LFM, SFM, NS, FRANK.
Figure 6. Variation in the recognition performance of eight types of signals with SNR. (a) 2FSK, EQFM, 4FSK, BPSK; (b) LFM, SFM, NS, FRANK.
Applsci 14 05499 g006
Figure 7. Recognition accuracy of dual-component radar signals as a function of different SNRs.
Figure 7. Recognition accuracy of dual-component radar signals as a function of different SNRs.
Applsci 14 05499 g007aApplsci 14 05499 g007b
Table 1. CNN-ST parameters.
Table 1. CNN-ST parameters.
OperatorsLayersSizeExpansion RatioStrideOutput
Convolution13 × 3-2224 × 224 × 32
Average pooling12 × 2-2112 × 112 × 32
Bottleneck1-11112 × 112 × 16
Bottleneck1-6256 × 56 × 24
Bottleneck1-1228 × 28 × 32
Swin transformer2---14 × 14 × 64
Swin transformer2---14 × 14 × 96
Convolution11 × 1-17 × 7 × 160
H-swish1---7 × 7 × 160
Convolution11 × 1-17 × 7 × 1280
Average pooling17 × 7-11 × 1 × 1280
Fully connected1---1 × 1 × 8
Table 2. Parameters of various radar signals.
Table 2. Parameters of various radar signals.
TypesParametersRanges
2FSK Carrier   frequency   f 1 ,   f 2 0.01 to 0.46
Bandwidth   f N/32 to N/8
BPSKBarker codes[5, 7, 9, 13]
Carrier   frequency   f 0 0.1 to 0.4
T s N/32 to N/16
4FSK Carrier   frequency   f 1   to   f 4 0.1 to 0.4
T s N/32 to N/8
EQFM Carrier   frequency   f 1 ,   f 2 0.05 to 0.4
Bandwidth   f 0.05 to 0.3
FRANK Carrier   frequency   f 0 0.1 to 0.4
T s N/100 to N/50
Phase number M[4, 5, 6, 7]
LFM Initial   frequency   f c 0.01 to 0.45
Bandwidth   f 0.05 to 0.4
NS Carrier   frequency   f 0 0.1 to 0.4
SFM Carrier   frequency   f 0 0.05 to 0.15
Bandwidth   f 0.05 to 0.35
Table 3. Computational complexity of two methods.
Table 3. Computational complexity of two methods.
ParameterCNN-STCNN-DQLN
FLOPs (G)3.75.2
Parameters (M)20.634.9
Time (ms)3359
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wan, C.; Zhang, Q. A Novel Dual-Component Radar-Signal Modulation Recognition Method Based on CNN-ST. Appl. Sci. 2024, 14, 5499. https://doi.org/10.3390/app14135499

AMA Style

Wan C, Zhang Q. A Novel Dual-Component Radar-Signal Modulation Recognition Method Based on CNN-ST. Applied Sciences. 2024; 14(13):5499. https://doi.org/10.3390/app14135499

Chicago/Turabian Style

Wan, Chenxia, and Qinghui Zhang. 2024. "A Novel Dual-Component Radar-Signal Modulation Recognition Method Based on CNN-ST" Applied Sciences 14, no. 13: 5499. https://doi.org/10.3390/app14135499

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop