Next Article in Journal
Uncontrolled Two-Step Iterative Calibration Algorithm for Lidar–IMU System
Previous Article in Journal
Hunting Network Anomalies in a Railway Axle Counter System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Working Mode Recognition of Non-Specific Radar Based on ResNet-SVM Learning Framework

College of Electronic Countermeasure, National University of Defense Technology, Hefei 230071, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3123; https://doi.org/10.3390/s23063123
Submission received: 11 January 2023 / Revised: 9 March 2023 / Accepted: 10 March 2023 / Published: 14 March 2023
(This article belongs to the Section Radar Sensors)

Abstract

:
Mode recognition is a basic task to interpret the behavior of multi-functional radar. The existing methods need to train complex and huge neural networks to improve the recognition ability, and it is difficult to deal with the mismatch between the training set and the test set. In this paper, a learning framework based on residual neural network (ResNet) and support vector machine (SVM) is designed, to solve the problem of mode recognition for non-specific radar, called multi-source joint recognition framework (MSJR). The key idea of the framework is to embed the prior knowledge of radar mode into the machine learning model, and combine the manual intervention and automatic extraction of features. The model can purposefully learn the feature representation of the signal on the working mode, which weakens the impact brought by the mismatch between training and test data. In order to solve the problem of difficult recognition under signal defect conditions, a two-stage cascade training method is designed, to give full play to the data representation ability of ResNet and the high-dimensional feature classification ability of SVM. Experiments show that the average recognition rate of the proposed model, with embedded radar knowledge, is improved by 33.7% compared with the purely data-driven model. Compared with other similar state-of-the-art reported models, such as AlexNet, VGGNet, LeNet, ResNet, and ConvNet, the recognition rate is increased by 12%. Under the condition of 0–35% leaky pulses in the independent test set, MSJR still has a recognition rate of more than 90%, which also proves its effectiveness and robustness in the recognition of unknown signals with similar semantic characteristics.

1. Introduction

Radar mode recognition is the process of identifying radar functions and working modes, based on parameters and styles extracted from the unknown received signals. With the continuous development of phased array, multi-beam, and signal processing technology, modern multi-function radar (MFR) acquires new abilities. MFR can flexibly change direction and signal style, accomplish stable tracking and monitoring, and can effectively carry out diversified combat tasks such as search, tracking, monitoring, and guidance [1]. Radar mode recognition technology plays an indispensable role in fields like ELINT, cognitive radio, and combat decision-making [2,3]. However, with the radar working style becoming more diversified, the parameter variability is significantly enhanced, and the electromagnetic space becomes more complex. People cannot effectively identify MFR behavior patterns through simple methods such as template matching or statistical histograms [4]. Mode recognition in the modern complex environment has become an important and challenging problem. This paper attempts to propose an effective solution, which combines an artificial intelligence algorithm with radar signal processing knowledge.
Radar behavior mode usually requires corresponding signal style and parameter range, because of its unique functional requirements. The conventional radar receivers receive intra-pulse waveforms, obtained by sampling, as well as the pulse description words (PDW), including radio frequency (RF), pulse width (PW), pulse amplitude (PA), direction of arrival (DOA), and time of arrival (TOA) [5], as shown in Figure 1. Because of relatively pure electromagnetic environment, simple signal parameters, and dull working mode, radar signals could realize behavior pattern recognition by comparison of PDW in previous times [6]. However, with the emergence of complex modulation patterns, such as jitter and stagger, simple parameter comparison has been unable to complete this recognition task. Methods based on statistical analysis and behavioral reasoning have been applied to pulse sorting and recognition [7]. With the development of phased array, radar waveform and styles are more changeable, and the aliasing between parameters is increasingly serious. Therefore, people have turned their attention to the radar intra-pulse waveform, trying to distinguish radar operating modes through waveform differences between signals [8]. However, the wide parameter range and random characteristics of radar signals, make the conventional methods decreasingly effective, when translated from theory to application. The above methods are all based on the analysis and identification of the differences between intra-pulse or inter-pulse data, ignoring the processing laws of the radar itself, that the purpose of radar signal diversification is to better serve the indicators and functions of radar, rather than simple differentiation.
For this reason, this paper proposes a multi-information source recognition and learning framework of radar patterns, called MSJR, which can not only analyze the difference of the extracted signal data, but also calculate the values of functional indicators under typical circumstances, such as detection range, resolution, and ambiguity. It can realize the pattern recognition of non-specific radar, by extracting features from three dimensions: range of parameters, functional indicators, and regularities of data. To this end, we also need to pay attention to a key problem, that is, how to express the complex data regularities, which can be solved through extensive application of deep neural networks.
The main contributions of this paper are as follows:
  • A machine learning method, with radar knowledge embedded, is proposed. The framework creatively applies the radar principles to the dataset, and extracts the radar functional indicators from it, of which the input vector of classifier consists, along with pulse sequences and modulation types, and its feature vector is degraded by the histogram method. Because of the correlation between signal features, MSJR is more beneficial to extract features include the distribution features of data, making the model more extensible and able to identify unknown signals that meet this feature.
  • We propose an MSJR learning framework, based on deep residual network and a support vector machine, for non-specific radar behavior pattern recognition. As an alternative to a single neural network or classifier, the framework optimizes two subnetworks, through model cascades, aiming at pattern recognition accuracy. This framework can better utilize the powerful data expression ability of deep learning, and make use of the classification ability of SVM. Moreover, it is more conducive to identifying patterns in complex environments with serious parameter overlap. Training the MSJR learning framework, can extract the information contained in the signal more efficiently than a single algorithm.
  • The radar operating mode dataset is proposed, with reference to the actual radar, and the proposed method is evaluated on the four types of radar operating datasets commonly used by MFR, with its range and working mode covering most of the information displayed in the public. It also covers modes like target detection, ranging, tracking, speed detection, along with seven inter-pulse modulation types, e.g., constant, jitter, dwell and switch, stagger, slip, wobbulated, and hybrid modulation. In addition, a certain amount of noise, and 0–50% leaky pulses, are added in some cases, which brings more challenges to classification. The experimental results show that the proposed model can achieve more than 90% recognition accuracy under the condition of 10% measurement error and 35% leaky pulse, which is impossible to achieve with previous methods.
The rest of this paper is arranged as follows. Firstly, we review the relevant work in the field of radar operating mode recognition in Section 2. Section 3 describes the radar pattern recognition task and gives the representation method of radar behavior. In Section 4, the MSJR learning framework, based on ResNet24-1D and SVM, is established in four submodules. Section 5 reports datasets, experimental designs, and experimental results, to evaluate and compare the performance of the proposed framework with other recognition technologies. Section 6 concludes the article.

2. Related Work

With the diversification of radar systems and the complexity of electromagnetic space, the methods to classify radar signals using the statistical characteristics of signals, such as the histogram method and pulse repetition interval (PRI) transform method [9], have been unable to adapt to the task of radiation source identification and classification. The flexible parameter style and overlapping working range make the statistical characteristics of signals no longer obvious, therefore, mining intrinsic features of signals for recognition and classification, has become the main research direction. These methods can be summarized into two categories: knowledge-driven and data-driven recognition methods.
Knowledge-driven methods exploit the signals according to the known expert knowledge, and extract the internal differences for classification and recognition. The authors of [10,11] proposed high-order statistics of signals, to replace the original signal, for identification after transformation. In [12,13,14], the authors used short-time Fourier transform (STFT), wavelet transform, Wiger–Ville distribution, and other methods, to conduct time–frequency analysis, then they extracted the signal pulse characteristics for identification, which could distinguish the subtle differences of radar electrical signals. The radar ambiguity function was effectively used in [15], which represented the autocorrelation of the signal, and the analysis of the ambiguity function can also accurately locate different radiation sources. As parameters vary quickly, Visnevski tried to describe radar pulses with a multi-level structure in [16], and put forward the concept of “radar word”. He resolved a fixed sequence of pulses into several levels, e.g., radar words, phrases, sentences, paragraphs, to represent radar behavior, and established the relationship between radar words on the basis of a hidden Markov chain. However, it is difficult to solve the contradiction between recognition accuracy and operation complexity caused by the effective encoding length of words [17]. Starting from the semantic relation of radar context, the authors of [18] calculated the confidence interval of unstable evidence based on Dempster–Shafer (D–S) evidence theory, and the identification results exceeded 90%. The common problem of knowledge-driven algorithms, is that the recognition ability depends on the stability of the selected features. Once the prior information is biased, or the selected features do not work on some patterns, the recognition ability will deteriorate greatly. Therefore, more scholars have turned their interests to data-driven algorithms.
Data-driven methods optimize the known model, by learning a large amount of data, to achieve classification and recognition. The widely-used traditional classifier and deep learning network are the typical examples. In [19,20], a clustering algorithm was used to help SVM extract features, but it could not identify complex signal styles. Since deep neural networks are widely used in radar recognition, the authors of [21,22,23,24] used a denoising auto-encoder (DAE), a convolutional neural network (CNN), a residual neural network (ResNet), and a recurrent neural network (RNN), respectively, to achieve their recognition rates of more than 90%, under the given conditions, which verified the advantages of the deep learning method in the field of signal recognition, but at the same time, it also exposed the weakness of a single network facing a complex environment. For the improvement of multi-objective tasks, the authors of [25,26,27] used the method of adding windows in the time domain, to process data, designed the time processing module and the threshold function of the selection window, and thus realized multi-objective classification. The authors of [28,29] proposed a comprehensive recognition method, based on a traditional classifier and deep learning network. It confirmed the feasibility of identifying unknown signals from known signals, which employed a classifier to train the deep learning network and deduce the center vector of the known data.
As for the convolutional residual network and support vector machine model used in the proposed framework, there are many excellent improvement methods in the field of engineering. In view of the difference in results caused by different structures of CNN network, a modified D–S algorithm was designed in [30], which optimizes the network architecture, and achieved higher accuracy in the identification of surface cracks of concrete structures. The work in [31], proposed a comprehensive recognition method that combines long short-term memory networks (LSTM) and CNN, in which CNN was used to extract hidden features in low SNR environments, and LSTM was used to establish the semantic relationship. However, the network could not preserve the global characteristics of the sample, causing the loss of information. In [32], the network adaptively adjusted the activation function threshold according to the input data, and filtered the noise to the maximum extent possible while retaining the original features. However, while removing the noise, the model also removed some valid information. As an excellent nonlinear classifier, SVM’s performance is always restricted by hyperparameters. In [33,34], the regularization parameters and kernel functions of SVM were optimized based on the chaotic differential evolution algorithm and the Black Widow algorithm, respectively. Compared with the conventional artificial fish swarm algorithm and particle swarm optimization algorithm, the regularization parameters and kernel functions of SVM were more effective in thermal process recognition and EEG recognition, but the burden of operation speed was heavier. The authors of [35] proposed a parallel hybrid algorithm for SVM parameter optimization, that combined mainstream sequential minimum optimization with stochastic gradient descent, which achieved a great improvement in operation speed, but its accuracy was not as high as the above algorithms.
In the past work, most researchers have used the means of improving the network model and optimizing the hyperparameters, to increase the signal recognition ability. The methods of extracting target features from deep learning networks and classifiers are essentially data-driven, by dividing the differences of data in different working modes. One of the reasons why it is difficult to further improve the recognition efficiency, is that the data is polluted by noise, and the machine’s learning ability cannot extract more and more effective features. In this case, the recognition ability of the model can be improved by manually defining features, through the description of radar rules. For the task of radar mode recognition, we should pay more attention to the functional differences generated by the signals, rather than the differences between the signals. Therefore, we learn from the authors of [28,36], to encode the radar information, which maps the radar performance indicators into the machine learning model. An integrated recognition framework is used to integrate the advantages of data-driven and knowledge-driven methods. The experiment proved that this idea was feasible.

3. Problem Formulation

This paper aims to solve the problem of automatic recognition of radar behavior mode, under complex conditions. This section describes the functional characteristics of the radar behavior mode, and provides the definitions of six modulations, then uses a hierarchical method to establish a mathematical model for radar mode recognition.

3.1. Radar Behavior Mode

To meet the functional requirements of radar, MFR needs to optimize the signal parameters of different modes from the aspects of beam scheduling and waveform design. Radar behavior can be limited to four categories: velocity search, ranging, tracking, and search while track [37]. In this study, the airborne phased array radar is taken as an example, to define different modes from the perspective of non-cooperative signals.

3.1.1. Velocity Search (VS)

VS mode is a radar operating mode that can provide the farthest detection range. It mainly detects targets by detecting their speed. This mode uses high PRF at a fixed period, which can maximize the radar coherent processing period, to improve the detection range, and avoid Doppler ambiguity. At this time, the head-on target is in the low clutter area, so sidelobe noise generated by echo ambiguity can be avoided. The target can be detected in the Doppler domain by using constant PRF.

3.1.2. Range while Search (RWS)

RWS mode is the most commonly used search mode of radar, which is characterized by fast detection of multiple targets and provision of range, azimuth, altitude, and other information. According to the repetition frequency, it can be divided into high pulse repetition frequency (HPRF) and medium pulse repetition frequency (MPRF) waveforms. The two waveforms cooperate with each other, to provide rapid detection of a large airspace situation. In HPRF mode, the pulse parameters are relatively fixed, and the range ambiguity is mainly solved by three-stage linear frequency modulation technology. Although the range ambiguity and range illusion can be solved to a certain extent, the accuracy is low. For the MPRF waveform, radar uses dwell conversion or stagger to solve range ambiguity, which improves the accuracy, but decreases the gain of coherent processing and detection range.

3.1.3. Single Target Tracking (STT)

STT mode, is the working mode with the highest accuracy and the most concentrated energy. It can accurately and continuously obtain the target’s azimuth, distance, speed, and other information. The antenna beam can constantly change with the target, to lock it, which can guide the precision weapon to strike. Since this mode aims to track known targets, within the range of the known range gate and Doppler frequency, it does not consider the ambiguity problem, and has the most flexible waveform design method. In order to ensure stable tracking and accuracy for weapons to strike, various modulation, e.g., constant, sliding, jitter, stagger, and periodic, can be adopted in different situations.

3.1.4. Trace and Search (TAS)

TAS is a composite mode, working on tracking, which is mainly used for multi-target tracking in the target airspace, while ensuring a certain searching function. In this mode, the radar utilizes beam scheduling, to ensure stable illumination to the tracking target, ensure that the tracking is not lost, and have high accuracy. TAS has two repetition frequencies: HPRF and MPRF, and has a similar working behavior to RWS when searching, while its tracking mode is similar to STT. The resource division is determined by target number and scheduling method.

3.2. Pulse Modulation Mode

A radar signal can be modeled as a collection of discrete signals, with PRI as period. Let T P , n Z + , respectively, be the pulse period and pulse sequence number, then the radar signal, y ( t ) , can be expressed as:
y ( t ) = n = 1 Z ( μ ( t n T ) μ ( t n T τ ) )
where μ ( t ) represents the single pulse signal waveform, and τ represents the pulse width. Using the same method, we introduce the following six modulations and give the mathematical expression of the pulse parameter Y ( n ) , of the nth pulse; M is the number of pulses in a modulation period. In an ideal environment, the pattern of modulation is clear, but the distribution of data between pulses is very vulnerable to the interference of lost pulses.
The receiver’s reception of electronic signals in a noisy environment is not complete, and the signal intercepted by the receiver may be highly distorted. The signals obtained by electronic reconnaissance are usually mixed signals, of various radiation sources and environmental noise in the receiver frequency band, and radar pulses will be offset and lost to some extent. Due to the limitations of the receiver’s bandwidth and sensitivity, under the limitation of the fixed false alarm rate in the signal detection link, with the decrease in the signal-to-noise ratio, the radar will generate more lost pulses [38]. The degree of lost pulse is affected by receiver performance and the false alarm rate setting. This paper mainly studies the changes in radar pulse descriptors, so it is more appropriate to use a more uniform standard of lost pulse, instead of signal to noise ratio, to measure noise. Figure 2 describes the pulses of six modulations under an ideal environment and with 30% leaky pulses.

3.2.1. Constant

Radar parameters are not completely fixed, and it is generally believed that the variation is less than 1% of the average value, which can be considered as constant modulation, Y ( n ) = ( 1 + r a n d ( 0.01 , 0.01 ) ) × i = 0 n Y ( i ) / n . This type of signal has strong stability and is usually used for searching and tracking radars, especially those using moving target indication (MTI) technology and pulse doppler (PD) technology.

3.2.2. Stagger

This mode regularly converts its value around a stable period. Y ( n ) = S ( n % M ) , S is a set of stagger group sequences. Stagger modulation can eliminate the blind speed of moving targets, and PD radar can also solve range and speed ambiguity, through PRI stagger.

3.2.3. Jitter

The parameter value randomly jitters around a central value. The jitter value is a random sequence, symmetrically distributed in the interval T , T . The relative jitter amount is usually 1 % 30 % , Y ( n ) = 1 + r a n d 0.3 , 0.3 × Y ¯ ( n ) . This jitter feature is usually regarded as an electronic countermeasure against some interference.

3.2.4. Dwell and Switch

The radar pulse sequence is composed of several constant waveforms. Each value stays for a certain time and then converts to the next value. Y ( n ) = D ( n % M ) , D is a set of variable sequences. High and medium PRF radars usually adopt D&S modulation, to solve the problem of range and speed ambiguity.

3.2.5. Sliding

The value monotonously increases or decreases, and then quickly returns to another extreme value when reaching one, Y ( n ) = Y m i n + ω × ( n % M ) , ω denotes the sliding change rate. This method can be used to solve the blind distance, and its parameter changes continuously. Sliding modulation can also be used to optimize scanning in angle of pitch, usually the maximum value is less than six times the minimum value.

3.2.6. Wobbulated

The parameter value usually changes periodically, in the form of sine, cosine, or a triangle wave, and its change rate, α , is close to the sine change rate, which can reach 5% of the average value. The change frequency, ω , can reach up to 50 Hz, Y ( n ) = Y m i n + α sin n ω + 2 π × n % M / M . This modulation style is usually used for missile guidance, accurate ranging, and refraining from the shielding effect.
Besides these six widely used modulations described above, there are other modulations such as scheduling, burst, etc. This paper mainly studies the above six modulations and their hybrid forms.

3.3. Hierarchical Radar Pattern Recognition Task

According to the hierarchical structure, on which MFR depends, this paper defines the radar signal as four levels, with reference to the “radar word” model, which are, the working mode, processing period, pulse sequence and characteristic parameters. Then, the task of this paper can be expressed as reverse identification of the top-level radar behavior mode, using the set of characteristic parameters received by the bottom layer. Figure 3 shows the hierarchical model of the structure.
Definition 1. 
Working mode B R , R = A , B , C , D , is the set of four behavior modes. B is a set of processing cycle sequences P N , B = P 1 , P 2 , P 3 P N , signal processing in the same period is consistent.
Definition 2. 
Each processing period contains pulse sequences, p n , with the same processing method, P N = p 1 , p 2 , p 3 p n , n represents the number of pulses in the processing period, 1 n L , L represents the number of pulses in the maximum processing period.
Definition 3. 
The pulse is represented by k feature parameter sets, which take PDW and its extended features. Then the ith pulse characteristic set can be expressed as ( R F i , P W i , P A i , D O A i ,   T O A i p a r a m e t e r k i ).
The previous research on behavior pattern recognition [36,39,40] mainly considered a single application scenario. The recognition model was established under the condition that the radar model was specific, as well as the fluctuation range of the signal was small, so it had little scalability. In the complex electromagnetic environment, the following three challenges must be solved, to accurately identify the radar operating mode: (1) The signal may come from different radar models, and the unknown signal may be outside the preset range. (2) Because of the change in radar beam and the leaky pulse, caused by low interception technology, the regulation of pulses will be seriously damaged. (3) The difference of radar modulations leads to great deviations in extracting signal features from the defective data. In order to meet the above challenges, and correctly identify the radar working mode, this paper divides the pattern recognition task into three sub-tasks, namely, data regulation extraction, calculation of performance indicators, and feature selection and classification. The regulation extraction will enable the network to focus on the regulations of data distribution, instead of mapping behavior patterns from the heterogeneous overlapping data. The indicator calculation is a human intervention identification process. Based on radar principles, the features expressed in the dataset can be refined to the maximum extent, and interference information can be filtered. The selection and classification of features needs to compare with the real tags, and automatically select the features with the greatest weight, for pattern recognition.

4. Proposed MSJR Learning Framework

In order to meet the challenge of pattern recognition of different types of radars, in a complex electromagnetic environment, a new MSJR learning framework embedded radar knowledge, is proposed in this paper. The general diagram of the proposed scheme is shown in Figure 4.
The proposed learning framework mainly includes four steps: (1) ResNet24-1D is used to identify modulation; (2) we extract performance indicators from signal processing, then establish the classification feature sets, to conduct a multi-source fusion model; (3) constant threshold normalization for training and testing; (4) SVM is used to identify behavior patterns.
The reason for the step-by-step implementation, is that the timing information of pulses will be destroyed after feature extraction through modeling, and the step-by-step implementation can make the retained information more complete. It should be noted that ResNet24-1D and SVM are trained in a cascading manner, and ResNet will be jointly optimized, according to the recognition of modulation and the behavior mode.

4.1. Residual Neural Network for Modulation Recognition

In the learning framework, we use the deep residual network to solve the problem of initial modulation recognition. A residual neural network, is a neural network that gives consideration to both learning ability and efficiency. It has achieved great success in computer vision and medical image recognition, and is promoted in semantic recognition, signal recognition, and other fields. The application of ResNet for modulation recognition, instead of RNN networks or other new networks, such as Transformer, is mainly based on three considerations:
  • The number of pulses in the sample is 8000∼16,000, so the model needs to have sufficient expression ability to extract signal patterns which are overlapped and seriously missing in long samples;
  • The model structure should avoid being too large and complex, and the cost performance ratio should be considered;
  • It can, not only have a strong representation ability, but also effectively avoid gradient problems caused by deep networks.
ResNet meets the above requirements [41]. The structure of the proposed network model is shown in Figure 5.
The input of the network is the PDW parameter, after leveling, and the sample size is [1 × 8000 × 5]. The network extracts features through a 1D-convolution layer, three residual modules, and a full connection layer, and the filter size of the convolution layer is set to 3 × 1. Each convolution layer is matched with a batch normalization layer, and a ReLU activation function, to solve the problem of gradient explosion and gradient disappearance. After the last residual module, a global average pooling is used, to extract the important features of each channel, and the full connection layer and SoftMax activation function are used to output a [7 × 1] one-hot label. The traditional convolutional neural network, classifies through directly connected convolution layers, this can easily cause model over fitting. The proposed structure uses global average pooling to extract the mean value of each feature map, and directly inputs to the full connection layer. Furthermore, it establishes a quick connection between modules in residual structure, which effectively avoids the occurrence of over-fitting problems, when deepening the network depth. Table 1 details the output shape and parameter number of each layer of the proposed network.
We define the classification loss for modulation classification as:
L c 1 = 1 N i = 1 n ( y i ln a i + ( 1 y i ) ln 1 a i )
where y i represents the real labels of N modulation style samples in a batch size, and a i represents the predictive labels. L c 2 is the classification loss for SVM. Trainable parameters in ResNet can be optimized by the combination of L c 1 and L c 2 , as:
L = L c 1 + L c 2
The following, explains the role of residual connections. Define h ( · ) as a direct mapping of a residual unit, f ( · ) as an activation function, then the output of a residual unit is y l = h ( x l ) + F ( x l , W l ) , where x l represents the input of l layer, W l is the weight matrix of the residual module, F ( x l , W l ) represents the value of the residual part, then the output of the next layer is x l + 1 = f ( y l ) . For a deeper residual layer, L, its relationship with layer l, can be expressed as:
x l = x l + 1 + i = l L 1 F ( x l , W i )
From the above formula, it can be concluded that any L layer can be expressed as the sum of shallower l layers and the residual modules between them. Compared with conventional convolutional neural networks, L layer is expressed as the product of a shallow characteristic matrix. Applying the chain rule of derivatives, we can obtain the gradient of the loss function L , with respect to x l , as follows:
L x l = L x L x L x l = L x L 1 + x l i = l L 1 F ( x l , W i )
This means that the gradient of L layer can be directly transferred to any shallower l layer, in the process of back propagation of the network. Because of the additive structure of the residual module, i = l L 1 F ( x l , W i ) / x l cannot be kept at –1 all the time. The model is not prone to the problem of gradient disappearance, so for a better performance in recognition, this will be feasible to train a deeper network.

4.2. Joint Modeling Method

The joint modeling module is mainly divided into three feature sets, to form C = α w β q γ p , where α is the parameter vector, β is the indicator vector, γ is the modulation rule vector, and the superscripts w , q , and p represent the feature dimensions. The main function of the joint modeling module is to extract three feature sets and provide the input data of the SVM classifier.
The first step is to extract the parameter vector. Since the regulation of the signal has been completed in Section 4.1, only the range features will be extracted from the parameter vector. In this paper, the frequency characteristics of PDW in the whole pulse flow, are calculated by the histogram method. δ ( n ) is defined as a discrete sequence of N pulse values. The histogram method is used to obtain the statistical distribution, X ( k ) , of the parameters. The input vector is sampled on X ( k ) , in the interval of 10%, representing the overall information and frequency characteristics of the parameters. In this way, the parameter vector size of each sample is greatly compressed, while preserving the complete information as much as possible. By applying the same process to all parameters, an eigenmatrix, representing the range of parameters, can be obtained.
The next step, is to extract the feature vector, which is a comprehensive expression of radar performance, so it is necessary to use the original parameters and the modulation pattern information identified in Section 4.1. In this paper, radar performance is described by radar detection range, the radar ambiguity function, and ambiguity resolution capability. The performance index only considers the difference brought by the available parameters. For the parameter information that cannot be obtained, typical parameters are used instead. In view of the fact that radars use the reflection of electromagnetic waves to detect targets, the detection range of conventional radars is related to receiver performance, signal modulation, transmission power, and environmental noise [42]. Considering that MFR generally benefits from the gain brought by pulse accumulation, the radar equation can be written as follows:
R m a x = d p w P p k G A e σ n E i ( 4 π ) 2 S m i n L A L s y s 4
where, d p w represents the pulse width, P p k represents the peak power, G represents the antenna gain, A e represents the effective aperture of the antenna, σ represents the scattering sectional area, S m i n represents the reception threshold, L A and L s y s represent the propagation and system loss, respectively, and n E i represents the coherent accumulated gain of n pulses.
The radar ambiguity function is an important index in designing radar signals. It quantitatively describes the resolution of radar transmitted signals to targets, with different ranges and speeds. Define u ( t ) as the complex envelope of the transmitted signal. Assume that the time delay is d and d + τ , and the Doppler frequency shift is f and f + f d :
s 1 ( t ) = u ( t d ) e j 2 π ( f 0 + f ) ( t d )
s 2 ( t ) = u ( t ( d + τ ) ) e j 2 π ( f 0 + f + f d ) ( t ( d + τ ) )
Then the mean square deviation of two echoes in the radar matched filter response can be expressed as:
ε 2 = + u ( t d ) 2 d t + + u ( t ( d + τ ) ) 2 d t 2 W ( t )
W ( t ) = R e + u * ( t d ) u ( t ( d + τ ) ) e j 2 π ( ( t + d ) f d ( f 0 + f + f d ) τ ) d t
Order t ¯ = t ( d + τ ) , then the expression in the integral term in W ( t ) , can be simplified as:
X ( τ , f d ) = + u * ( t ¯ + τ ) u ( t ¯ ) e j 2 π f d t ¯ d t ¯
X ( τ , f d ) is the ambiguity function of the signal. The larger the value is, the smaller ε 2 is, and the harder it is to distinguish the two targets. Figure 6 shows the ambiguity function of the LFM signal.
We use the 4 dB width of X ( τ , 0 ) and X ( 0 , f d ) to represent their nominal resolutions. The range resolution and velocity resolution derived from the nominal resolutions Δ τ and Δ f d of delay τ and Doppler frequency f d , are:
Δ R = c Δ τ 2 = c + X ( τ , 0 ) 2 d τ 2 X ( 0 , 0 ) 2 = c 2 B
Δ V = c Δ f d 2 f 0 = c + X ( 0 , f d ) 2 d f d 2 f 0 X ( 0 , 0 ) 2 = c 2 f 0 T S P
where B is the effective bandwidth of the signal, f 0 is the carrier frequency, and T S P is the signal pulse accumulation time. Due to the lack of pulse information of the radar behavior pattern in public data, fuzzy function sampling and feature extraction cannot be carried out. Therefore, only theoretical feasibility is analyzed here, and the nominal resolution Δ τ , Δ f d is used as a substitute in the experiment.
In addition to the above indicators, radar ambiguity resolution capability is also an important indicator that we can refer to and obtain. Due to the influence of multiple echoes, and the insufficient sampling rate of low PRF in the Doppler base-band, the radar will have ambiguity in the range domain and speed domain, resulting in uncertainty in target detection. For dwell conversion, stagger, and other modulations, due to its high- and low-staggered PRF distribution, it can effectively solve the ambiguity in range and speed. In general, the joint modeling method, including parameter range, functional indicators, and modulation, can effectively simplify and represent the characteristic information of radar sequence.

4.3. Fixed Distribution Normalization

The pulse sample, after joint modeling, contains M radar features, and each feature has different units and orders of magnitude, which usually require standardization before feature comparison between modes. The conventional method, is to normalize based on the maximum value of the parameter feature, K. For the radar behavior pattern recognition, the test sample sequence may not contain all the behavior patterns. Therefore, normalization based on the maximum value of a certain type of feature may lead to different results for different sample sequences, which will lead to greater deviation in model testing. It is impossible to require an unknown and non-cooperative signal to have a complete pattern sequence. Therefore, this paper normalizes all samples based on a constant distribution. The standardized distribution of the mth feature, can be written as follows:
T m N 1 N i = 1 N p m i , 1 N i = 1 N p m i 1 N i = 1 N p m i
where n represents the number of samples of the feature, and p m i represents the value of the mth feature. This method provides a fixed standardized distribution from the training set, and the training and test samples are normalized according to this distribution. Experiments show that this method has the same recognition and classification ability for the samples with leaky patterns. Figure 7 shows the flow of constant threshold standardization.

4.4. Support Vector Machines for Mode Recognition

SVM is one of the most widely used kernel learning algorithms, and has achieved good results in the classification of signal, image, biology, and other fields [43]. The support vector machine performs classification by finding the optimal hyperplane in a given sample set. The boundary sample is defined as the support vector. The goal of the classification task is to maximize the distance between the support vector and the hyperplane [44]. Assuming that the sample set S = x i , y i , x i C , y i ± 1 , the optimization problem can be defined as:
min w 1 2 w w T + C i = 1 N ε i
where w is the weight vector of the decision plane, C is the penalty variable for error classification, and ε i is the relaxation variable, indicating the distance that a certain sample is allowed to cross the hyperplane. This is a convex optimization problem. According to the duality principle, the optimal solution can be obtained by using the Lagrange multiplier method. Since the dimensions of x i are α + β + γ , the feature needs to be mapped to a higher-dimensional space, for calculation by using the kernel method. The decision function of the classifier is defined as:
F ( x ) = s i g m u n x i s ρ * y i K ( x , x i ) + b *
where ρ * represents the Lagrange multiplier of each sample, K ( x , x i ) is the selected kernel function, s represents the support vector, and b * is the offset of the model. Different recognition problems require different kernel functions, and linear kernel functions are applicable to situations where the number of features is much larger than the number of samples; The Gaussian (RBF) kernel function is more suitable for the case where the number of samples is sufficient and the data is linearly indivisible. For radar signals, the sample size is much larger than the number of features, and the data meet the linear indivisibility condition, so it is more suitable for the RBF kernel function, the function can be written as:
K ( x , x i ) = e x p 1 2 σ 2 x x i 2
In the selection of SVM hyperparameters, the choice of the RBF kernel parameter, σ 2 , and penalty factor, C, is important in affecting the generalization ability and training error of the classifier. The larger the σ 2 is, the less sensitive it is to the training data, but the complexity of SVM decreases. If σ 2 becomes smaller, the training error decreases, but the support vector and the complexity increases. In this paper, the improved algorithm of SMO, proposed in reference [35], is used to optimize σ 2 and C. The main idea is similar to the coordinate ascending algorithm (CA). This method needs to optimize N Lagrange multipliers ρ 1 , ρ 2 ρ N in the convex optimization problem, decompose the programming problem with N parameters into multiple quadratic programming problems, and then obtain the optimal σ 2 and C through ρ * .

5. Experiments and Results

This section mainly conducts some simulation experiments and compares the performance of the proposed framework with other algorithms. The test is divided into two parts. In the first part, the convergence and generalization of different neural networks in the learning framework are compared, by taking the PRF modulation as an example. The second part tests the ability of different models to recognize radar behavior patterns in complex environments, and discusses the experimental results. The experiment was run on JetBrains PyCharm 2022 (Prague, Czezch Republic) with Lenovo Intel (R) Core (TM) CPU [email protected] GHz, GPU is NVIDIA GeForceRTX2060.

5.1. Dataset

Due to the confidentiality of radar technology, there is no publicly available dataset at present. The radar datasets used in most of the literature only show the range of parameters, which does not closely fit the actual data. Based on [37,39,45,46] and other public materials, this paper constructs two types of datasets, namely, radar pulse sequences set RPDWS-I and airborne multi-function radar behavior mode dataset RPDWS-II, with radar pulse sequences as the main body.
RPDWS-I mainly describes the radar modulation, including constant, jitter, slip, stagger, group change, periodic modulation and mixed modulation. The signal styles are shown in Table 2. RPDWS-II mainly describes the radar behavior modes, including speed search, range search, target tracking and search plus tracking. The modulation mainly focuses on PRI, RF, and PW. The signal style is shown in Table 3. Each signal sample is randomly generated within the given range, not limited to specific models.

5.2. Capability of Pulse Modulation Recognition Subnetwork

The experiment was conducted on the RPDWS-I dataset. To verify the effectiveness of the neural network in the learning framework, we trained according to 80% of the training set and 20% of the verification set, and the test set was completely independent of the training set and verification set. The experimental results were compared with the improved CNN network and full connection deep neural network (DNN), proposed in the literature [21], and the overall recognition accuracy was used as the evaluation standard.

5.2.1. In-Training Views

In order to study the influence of different neural network models and network layers on the recognition ability, we have built 6-layer, 12-layer, 18-layer, and 24-layer full connection, CNN, and ResNet networks. After many attempts, the hyperparameters of the optimized network are as follows. The network uses the Adam optimizer, as its structure is detailed in Section 4.1, with its learning rate reduced from 10 2 to 10 4 , at an interval of 40 epochs, size of the batch set to 128, L2 regularization used in the network, and regularization coefficient set to 10 4 . Each training is 160 rounds in total.
Each neural network is trained for ten times to ensure the stability of training. The training results are shown in Figure 8. The filled part represents the standard deviation of ten times of training. We can see that, after the convergence of the three models, the precision of the CNN and ResNet training sets and verification sets is almost the same, while the precision of the DNN network training set and verification set is quite different. This is because the DNN neurons are directly connected, without the structure of feature extraction, and the average parameter amount reaches an amazing 3G. Although it can fit the training data distribution to a certain extent, the over fitting is serious. From the perspective of model stability, the fluctuation between DNN and ResNet for ten times training is small, while the fluctuation of CNN is relatively serious, which means that CNN needs to be trained repeatedly to find the optimal result.
As shown in Table 4, the average precision of ResNet is 96.3%, much higher than the 79.5% of CNN and 58.3% of DNN. The networks with the highest precision of the three networks is the 12-layer DNN, 18-layer CNN, and 24-layer ResNet. The reason is, that when the number of network layers of DNN and CNN increases, the recognition accuracy has not been improved, but has declined to a certain extent. This is because there is no residual connection in the network, and the shallow training stops, due to the accumulation of small disturbances in deep layers, making the overall recognition rate unable to improve. From the training curve, when there are 40, 80, and 120 epochs, the curve shows obvious fluctuations. This phenomenon is meaningful, because the change in the learning rate on this node makes the model approach the optimal performance again due to the local oscillation. From the above training situation, the performance of ResNet is better than that of DNN and conventional CNN, while blindly increasing the number of network layers may not significantly improve the recognition accuracy, but may lead to performance degradation and the waste of computing resources.

5.2.2. In-Testing Views

In order to further verify the recognition capability of networks in complex environments, we tested three networks, DNN-12, CNN18-1D, and ResNet24-1D, which enjoy relatively good performance in the above analysis. In order to compare the migration performance of the models, we set up two experimental environments, one is to directly train on the original data, and the other is to take certain data enhancement measures, that is 0∼20% of the leaky pulses artificially added by steps. The test set is generated independently of the training set. On the basis of 10% measurement error, 0∼50% missing pulses was set every 5%, to verify the generalization ability of the model.
Similarly, each network selects five models with the best training effect, to ensure the stability of the test. The training results are shown in Figure 9. The filled area represents the standard deviation of the test accuracy. It can be seen that the performance of ResNet24-1D is better than that of the other two networks, under different leakage pulse ratios. Moreover, when the test data and training data have a more similar distribution, the accuracy of the test set is obviously higher. For example, in Figure 9a, when training on the original data with no leaky pulses, the accuracy rate was maintained at a high level. However, the accuracy dropped rapidly after the leaky pulses appeared. In Figure 9b, when training on 0∼20% enhanced datasets, the accuracy rate decreased slowly when the proportion of missed pulses was less than 20%, and ResNet24-1D could maintain a recognition rate of more than 90%.
When the difference between the distribution of the test set and the training set continues to increase, the performance decreases significantly, and some modulation patterns are almost unrecognized. From the results, the method in this paper is lower than the recognition rates of 95% and 96% proposed in [21,24], but it should be emphasized that the data transformation range and error of [21,24] are much smaller than the recognition environment used in this paper, and CNN18-1D, in Figure 9, is established with reference to the parameters given in [21]. It can be seen that the network structure of this paper has more advantages.
Figure 10 compares the confusion matrix of ResNet24-1D in different environments. It can be seen that the recognition performance is relatively stable in an ideal environment and 20% leaky pulses, while a large number of fluctuations occur at 40% leaky pulses. The reason is, that the increase in leaky pulses makes the data distribution of one modulation type tend to other types. In an ideal environment, the six conventional modulation styles are relatively stable, and the wrong classification is concentrated on the mixed style. This is because the mixed style has certain characteristics of other modulation styles, causing confusion in recognition. With 20% leaky pulses, although the absence of a certain pulse affects the distribution of different types of data, the original characteristic shape is basically maintained, so the accuracy is only slightly reduced.
With 40% leaky pulses, nearly half of the parameter values are wrong, causing serious changes to the data regularity. For example, leaky pulses cause the PRI value of adjacent pulses to increase to an integer multiple of the original value. In this way, a large number of constant signals and jittery signals are identified as stagger signals, and more complex signals, such as group changes and periods, tend to be recognized as mixed signals. Surprisingly, stagger, slip, and mixed signals are very little affected by the leaky pulses. This is because, even if the data regularities of stagger and mixed signals are destroyed, the more important thing is to change the signals from one stagger and mixed to another. Even if the PRI value of slip signals changes to an integer multiple within a slip period, the original regularities can still be restored, to a certain extent, during linear fitting of a slip period.

5.3. The Performance of the Proposed MSJR Learning Framework

In Section 5.2, we have verified that ResNet24-1D is applicable to the extraction of data distribution rules in the proposed framework. Based on the above work, this section will discuss the performance of the MSJR learning framework in radar behavior pattern recognition, and compare it with the recognition ability of traditional classifiers and deep learning algorithms. The experiment is conducted on the RPDWS-II dataset.
Considering the improvement of recognition performance brought about by the joint modeling method and feature extraction, we identify the leaky pulse signals with different proportions, based on three conditions: direct recognition, multi-source recognition, and recognition using the proposed learning framework. We choose support vector machine with linear kernel, support vector machine with Gaussian kernel, decision tree, random forest, naive Bayes, and other traditional classifiers, to compare with the ResNet deep learning network. Figure 11 shows the recognition accuracy curve of the classifier under various conditions. From Figure 11a–c, the improvement of recognition performance, after taking different measures, is recorded.
In Figure 11a, it is not difficult to see that when the single classifier is used to directly recognize the original radar data, the overall recognition accuracy is low. Even if there is no leaky pulse, the recognition rate is less than 70%. This is because, for different operating modes of the same type of radar, its operating frequency band and pulse arrival direction are relatively fixed, and the information’s amplitude is random. The key parameters for recognition, such as PRI and PW, overlap seriously in the data distribution, as shown in Figure 12. This undoubtedly increases the difficulty of recognition for the classification algorithm focusing on data distribution. However, it can be seen from the comparison of various classifiers, that the ResNet and SVM classifiers have stronger feature extraction ability than other classifiers.
After joint modeling, the overall recognition accuracy of the classifier is improved by about 15%, as shown in Figure 11b. At this time, the recognition ability of traditional classifiers is not much different, but surprisingly, the performance of ResNet is reduced by about 15%, compared with that of direct recognition. The reason is, that after joint modeling, the original data is greatly simplified, and only the feature data extracted from each information source is retained. The sample size is too small, and the difference between the data in the sample is too large, which makes the convolution layer unable to work effectively.
From the performance of each classifier in Figure 11c, when the proposed MSJR learning framework is used, the recognition performance is stable under the condition of 0∼30% missing pulses, and the recognition accuracy has been greatly improved. The recognition accuracy of the linear kernel SVM, Gaussian kernel SVM, decision tree, random forest, ResNet24-1D, and naive Bayes classifiers is 0.788 ± 0.004, 0.941 ± 0.006, 0.822 ± 0.004, 0.835 ± 0.005, 0.944 ± 0.006, and 0.745 ± 0.006, respectively, and the recognition ability of the ResNet24-1D and Gaussian kernel SVM is much higher than that the other classifiers. The reason for ResNet’s performance recovery is, that after the data distribution is standardized, the data in the samples are normalized to the same order of magnitude, which solves the problem of data loss.
Compared with the recognition curves of the three conditions, using the learning framework proposed in this paper, the recognition accuracy of the classifier is higher, and the recognition is more stable when the proportion of lost pulses increases. Surprisingly, under the condition of 10∼50% leaky pulses, the accuracy of ResNet and SVM in classifying radar behavior patterns, is even higher than that of modulation pattern recognition in the first stage. This means that the MSJR learning framework can still accurately identify the working mode in the case of some wrong modulation patterns. This may be that the classifier weakens the influence of modulation pattern recognition errors when extracting multiple information source features for classification, which is beneficial to pattern recognition tasks. It should be noted that the cost performance ratio of the classifier is also an important indicator. In the training above, the parameters of the ResNet24-1D model reached 120 M, and the training time was 2362 s, while the parameters of the SVM model with RBF core were only 4 M, and the training time was 13 s. Therefore, from the perspective of model availability, the SVM classifier avoids repeated construction and training of the large structure deep learning model, and has achieved similar results in feature classification, which may be more suitable for the MSJR learning framework.
In order to compare the recognition ability of the MSJR learning framework to each pattern, the pattern recognition confusion matrix, under the conditions of 20% and 40% missing pulses, is given, as shown in Figure 13. Although the accuracy of pattern recognition decreases with the increase in leaky pulses, it does not mean that all modes are affected. For example, the VS and STT modes are almost unaffected, while the RWS and TAS modes are seriously confused, and their performance degrades significantly, and the RWS mode can hardly be effectively recognized. This is because having 35% leaky pulses makes the range of parameters break the threshold of single waveform to composite waveform. In short, the damage of the waveform will lead to inefficiency in extracting characteristics, and ending up with wrong classifications. This also shows that if the range of radar parameters changes seriously, the performance of the MSJR learning framework is still limited. We will continue to investigate solutions in future research.
Finally, we test five kinds of ILSVRC champion network, in the same environment, to see how the MSJR framework improves performance compared to other similar state-of-the-art reported works. The classical convolutional network models, AlexNet, ConvNet, LeNet, ResNet, and VGGNet have been successfully applied in the field of radiation source identification. The accuracy under different levels of noise, the running time of a single epoch, and the model capacity are compared, and the results are shown in Table 5.
From the analysis of accuracy, AlexNet, ConvNet, and LeNet not only have lower accuracy than the proposed MSJR framework, but also show their shortcomings in the change in accuracy. Due to the lack of effective means to prevent over fitting, the accuracy of the three networks is the highest under the conditions of 10∼30% lost pulses, which is similar to the training set. However, it decreases under the ideal environment, which indicates that the model fails to effectively learn the signal characteristics. The performance of VGGNet decreases seriously under the condition of noise increase, which indicates that its generalization ability is insufficient and its adaptability is weak. ResNet achieved the best performance in the comparison of networks, but its overall performance was about 4% weaker than the proposed MSJR framework, due to the limitations of the purely data-driven approach. From the analysis of processing time and model capacity, AlexNet and LeNet have advantages in this aspect, but their accuracy cannot meet the requirements. ConvNet was inferior in every respect, and its overall performance was insufficient. Compared with the above three models, ResNet, VGGNet and the proposed MSJR framework have deeper network layers, more complex design and larger parameter quantity, so the processing time is slightly higher.
In summary, although the proposed MSJR network is slightly deficient in processing time and model capacity, its comprehensive performance is better than that of the other five networks, and its average accuracy is not only 4∼46% higher than that of other networks, but also has stronger generalization and robustness.

6. Conclusions

In this paper, a new learning framework, called MSJR learning framework, is designed for non-specific RMI. As an alternative to a simple neural network or classifier, the advantage is that the overall performance of each sub-module is improved, by the way of cascade training. Meanwhile, MSJR takes into account the impact of radar parameter range, functional indicators, and modulation style, on the behavior pattern, and can more accurately and efficiently extract the behavior features of different information sources within a wide range of overlapping parameters. In addition, the MSJR learning framework uses the ResNet neural network and the RBF core SVM, with the highest cost performance ratio, to realize modulation pattern recognition and pattern classification, respectively. The experimental results show that the model can maintain 90% recognition accuracy, stably, when the leaky pulses are less than 35%, which realizes effective recognition of radar behavior patterns. The stability and generalization of the proposed MSJR learning framework have also been quantitatively verified. Although the recognition performance of the model for RWS mode deteriorates in more severe environments, the MJRS learning framework is generally suitable for identifying non-specific radar behavior modes.
We plan to perform further work on the following three aspects in the future: (1) Working mode recognition under unsupervised or weakly supervised conditions. This makes the model have to find the effective characteristics of the signal by itself. The available methods include, sequence matching, envelope analysis, and prediction state description, etc. No substantial breakthrough has been made so far, but this may be closer to real conditions. (2) Working mode recognition in small samples, or no samples, namely zero-shot learning. The model has the ability to identify unknown signals from the visible training set, which is of great significance for reducing model cost and improving generalization ability. Common methods include, graph convolutional network and encoder-like model. (3) A radar pulse denoising algorithm. This is needed to meet the requirements of sparse space denoising. Available denoising networks include FFDNet, DNCNet, etc. This is helpful for further processing of the signal.

Author Contributions

J.P.: conceptualization, supervision, project administration, funding acquisition. J.X.: survey, model design, methodology, software, visualization, writing. Y.Z.: review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

National Natural Science Foundation of China (no. 62071476).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the reviewers and editors for their efforts and constructive comments to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ResNetResidual Neural Network
SVMSupport Vector Machine
MSJRMulti-Source Joint Recognition Framework
MFRMulti Function Radar
PDWPulse Description Word
CNNConvolutional Neural Network
DNNDeep Neural Network
VSVelocity Search
RWSRange while Search
STTSingle Target Tracking
TASTrace and Search
HPRFHigh Pulse Repetition Frequency
MPRFMedium Pulse Repetition Frequency
D&SDwell and Switch
RFRadio Frequency
PWPulse Width
PAPulse Amplitude
DOADirection of Arrival
TOATime of Arrival
PRIPulse Repetition Interval

References

  1. De Martino, A. Introduction to Modern EW Systems, 2nd ed.; Electronic Warfare Library, Artech House: Boston, MA, USA, 2018. [Google Scholar]
  2. Wiley, R.G. ELINT: The Interception and Analysis of Radar Signals; Artech House Radar Library, Artech House: Boston, MA, USA, 2006. [Google Scholar]
  3. Weber, M.E.; Cho, J.Y.; Thomas, H.G. Command and Control for Multifunction Phased Array Radar. IEEE Trans. Geosci. Remote Sens. 2017, 55, 5899–5912. [Google Scholar] [CrossRef]
  4. Mardia, H.K. New techniques for the deinterleaving of repetitive sequences. IEE Proc. F 1989, 136, 149–154. [Google Scholar]
  5. Jiang, Q. Network Radar Countermeasure Systems: Integrating Radar and Radar Countermeasures, 1st ed.; Springer: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  6. Rogers, J. ESM processor system for high pulse density radar environments. Proc. IEE Proc. Commun. Radar Signal Process. 1985, 132, 621–625. [Google Scholar] [CrossRef]
  7. Kauppi, J.P.; Martikainen, K.; Ruotsalainen, U. Hierarchical classification of dynamically varying radar pulse repetition interval modulation patterns. Neural Netw. 2010, 23, 1226–1237. [Google Scholar] [CrossRef]
  8. Lida, D.; Shilian, W.; Fanggang, W.; Wei, Z. Specific Emitter Identification via Convolutional Neural Networks. IEEE Commun. Lett. 2018, 22, 2591–2594. [Google Scholar]
  9. Nishiguchi, K.I.; Kobayashi, M. Improved algorithm for estimating pulse repetition intervals. IEEE Trans. Aerosp. Electron. Syst. 2000, 36, 407–421. [Google Scholar] [CrossRef]
  10. Han, L.; Gao, F.; Li, Z.; Dobre, O.A. Low Complexity Automatic Modulation Classification Based on Order-Statistics. IEEE Trans. Wirel. Commun. 2017, 16, 400–411. [Google Scholar] [CrossRef]
  11. Orlic, V.; Dukic, M. Multipath channel estimation algorithm for automatic modulation classification using sixth-order cumulants. Electron. Lett. 2010, 46, 1. [Google Scholar] [CrossRef] [Green Version]
  12. Lunden, J.; Koivunen, V. Automatic Radar Waveform Recognition. IEEE J. Sel. Top. Signal Process. 2007, 1, 124–136. [Google Scholar] [CrossRef]
  13. Shaokun, L.; Xiaopeng, Y.; Ping, L.; Xinhong, H.; Ke, W. Radar Emitter Recognition Based on SIFT Position and Scale Features. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 2062–2066. [Google Scholar]
  14. Zhang, M.; Liu, L.; Diao, M. LPI Radar Waveform Recognition Based on Time-Frequency Distribution. Sensors 2016, 16, 1682. [Google Scholar] [CrossRef] [Green Version]
  15. Xu, C.; Zhang, J.; Zhou, Q.; Chen, S. Recognition of radar signals based on AF grids and geometric shape constraint. Signal Process. 2019, 157, 30–44. [Google Scholar] [CrossRef]
  16. Nikita, V. Syntactic Modeling of Multi-Function Radars. Ph.D. Thesis, McMaster University, Hamilton, ON, Canada, 2005. [Google Scholar]
  17. Visnevski, N.; Krishnamurthy, V.; Wang, A.; Haykin, S. Syntactic Modeling and Signal Processing of Multifunction Radars: A Stochastic Context-Free Grammar Approach. Proc. IEEE 2007, 95, 1000–1025. [Google Scholar] [CrossRef]
  18. Zhou, Y.; Yu, X.; Cui, M. Radar Target Recognition based on Multiple Features Fusion with Dempster-Shafer Theory. In Proceedings of the IEEE 2011 10th International Conference on Electronic Measurement & Instruments, Chengdu, China, 16–19 August 2019; pp. 262–266. [Google Scholar]
  19. Li, D.; Gu, L.; Zhu, L. Identification and parameter estimation algorithm of radar signal subtle features. Phys. Commun. 2020, 42, 101140. [Google Scholar] [CrossRef]
  20. Wu, Z.; Yang, Z.; Sun, H.; Yin, Z.; Nallanathan, A. Hybrid radar emitter recognition based on rough k-means classifier and SVM. EURASIP J. Adv. Signal Process. 2012, 2012, 198. [Google Scholar] [CrossRef] [Green Version]
  21. Li, X.; Huang, Z.; Wang, F.; Wang, X.; Liu, T. Toward Convolutional Neural Networks on Pulse Repetition Interval Modulation Recognition. IEEE Commun. Lett. 2018, 22, 2286–2289. [Google Scholar] [CrossRef]
  22. Li, X.; Liu, Z.M.; Huang, Z. Denoising of Radar Pulse Streams with Autoencoders. IEEE Commun. Lett. 2020, 24, 797–801. [Google Scholar] [CrossRef]
  23. Lutao, L.; Xinyu, L. Radar signal recognition based on triplet convolutional neural network. EURASIP J. Adv. Signal Process. 2021, 2021, 112. [Google Scholar]
  24. Han, J.W.; Park, C.H. A Unified Method for Deinterleaving and PRI Modulation Recognition of Radar Pulses Based on Deep Neural Networks. IEEE Access 2021, 9, 89360–89375. [Google Scholar] [CrossRef]
  25. Dadgarnia, A.; Sadeghi, M.T. Automatic recognition of pulse repetition interval modulation using temporal convolutional network. IET Signal Process. 2021, 15, 633–648. [Google Scholar] [CrossRef]
  26. Li, Y.; Zhu, M.; Ma, Y.; Yang, J. Work modes recognition and boundary identification of MFR pulse sequences with a hierarchical seq2seq LSTM. IET Radar Sonar Navig. 2020, 14, 1343–1353. [Google Scholar] [CrossRef]
  27. Yu, H.h.; Yan, X.p.; Liu, S.k.; Li, P.; Hao, X.h. Radar emitter multi-label recognition based on residual network. Def. Technol. 2022, 18, 410–417. [Google Scholar]
  28. Lutao, L.; Xinyu, L. Unknown radar waveform recognition system via triplet convolution network and support vector machine. Digit. Signal Process. 2022, 123, 103439. [Google Scholar]
  29. Tao, X.; Shuo, Y.; Zhangmeng, L.; Fucheng, G. Radar Emitter Recognition Based on Parameter Set Clustering and Classification. Remote Sens. 2022, 14, 4468. [Google Scholar]
  30. Yang, Y.; Bijan, S.; Maria, R.; Masoud, M.; Nguyen, N.T.; Guang, Z. Vision-based concrete crack detection using a hybrid framework considering noise effect. J. Build. Eng. 2022, 61, 105246. [Google Scholar]
  31. Han, L.; Donghang, C.; Xiaojun, S.; Feng, W. Radar emitter recognition based on CNN and LSTM. In Proceedings of the 2021 International Conference on Neural Networks, Qingdao, China, 27–29 August 2021; Volume 11933, p. 119331T. [Google Scholar]
  32. Jifei, P.; Shengli, Z.; Lingsi, X.; Long, T.; Linqing, G. Embedding Soft Thresholding Function into Deep Learning Models for Noisy Radar Emitter Signal Recognition. Electronics 2022, 11, 2142. [Google Scholar]
  33. Santos, G.S.d.; Luvizotto, L.G.J.; Mariani, V.C.; Coelho, L.d.S. Least squares support vector machines with tuning based on chaotic differential evolution approach applied to the identification of a thermal process. Expert Syst. Appl. 2011, 39, 4805–4812. [Google Scholar] [CrossRef]
  34. Huang, Q.; Wang, C.; Ye, Y.; Wang, L.; Xie, N. Recognition of EEG based on Improved Black Widow Algorithm optimized SVM. Biomed. Signal Process. Control 2023, 81, 104454. [Google Scholar] [CrossRef]
  35. Gizen, M.; İnan, A.Ç. SVM-SMO-SGD: A hybrid-parallel support vector machine algorithm using sequential minimal optimization with stochastic gradient descent. Parallel Comput. 2022, 113, 102955. [Google Scholar]
  36. Qihang, Z.; Yan, L.; Zilin, Z.; Yunjie, L.; Shafei, W. Adaptive feature extraction and fine-grained modulation recognition of multi-function radar under small sample conditions. IET Radar Sonar Navig. 2022, 16, 1460–1469. [Google Scholar]
  37. Skolnik, M.I. Radar Handbook, 3rd ed.; McGraw-Hill: New York, NY, USA, 2008. [Google Scholar]
  38. Weichao, X.; Huadong, L.; Jisheng, D.; Yanzhou, Z. Spectrum sensing for cognitive radio based on Kendall’s tau in the presence of non-Gaussian impulsive noise. Digit. Signal Process. 2022, 123, 103443. [Google Scholar]
  39. Kun, C.; Jihong, S.; Yan, L.; Liyan, W.; Sheng, W. A novel segmentation approach for work mode boundary detection in MFR pulse sequence. Digit. Signal Process. 2022, 126, 103462. [Google Scholar]
  40. Liao, Y.; Chen, X. Multi-attribute overlapping radar working pattern recognition based on K-NN and SVM-BP. J. Supercomput. 2021, 77, 9642–9657. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Alabaster, C. Pulse Doppler Radar: Principles, Technology, Applications; SciTech Publishing: Edison, NJ, USA, 2012; p. xiv. 411p. [Google Scholar]
  43. Wang, S.; Gao, C.; Zhang, Q.; Dakulagi, V.; Zeng, H.; Zheng, G.; Bai, J.; Song, Y.; Cai, J.; Zong, B. Research and Experiment of Radar Signal Support Vector Clustering Sorting Based on Feature Extraction and Feature Selection. IEEE Access 2020, 8, 9332–93334. [Google Scholar] [CrossRef]
  44. Xu, J.; He, M.; Han, J.; Chen, C. A Comprehensive Estimation Method for Kernel Function of Radar Signal Classifier. Chin. J. Electron. 2015, 24, 218–222. [Google Scholar] [CrossRef]
  45. Feng, H.C.; Tang, B.; Wan, T. Radar pulse repetition interval modulation recognition with combined net and domain-adaptive few-shot learning. Digit. Signal Process. 2022, 127, 103562. [Google Scholar] [CrossRef]
  46. Hui, L.; Dong, J.W.; Dong, L.H.; Wei, C.T. Work Mode Identification of Airborne Phased Array Radar Based on the Combination of Multi-Level Modeling and Deep Learning. In Proceedings of the the 35th China Command and Control Conference, Chengdu, China, 27–29 July 2016; pp. 273–278. [Google Scholar]
Figure 1. Basic model of radar pulse.
Figure 1. Basic model of radar pulse.
Sensors 23 03123 g001
Figure 2. Conventional modulation styles in different environments. (a,b) respectively show 6 modulation styles under ideal conditions and 30% lost pulse condition.
Figure 2. Conventional modulation styles in different environments. (a,b) respectively show 6 modulation styles under ideal conditions and 30% lost pulse condition.
Sensors 23 03123 g002
Figure 3. Hierarchical model of radar pattern recognition task.
Figure 3. Hierarchical model of radar pattern recognition task.
Sensors 23 03123 g003
Figure 4. The architecture of the proposed MSJR. In the modeling module, Pulse parameter represents the modeling information of the original pulse parameters, Radar capability represents the calculated performance indicators such as typical detection range, resolution, ambiguity resolution, etc. Modulation style represents the pulse change rules identified by ResNet24-1D, such as constant, sliding, stagger, etc.
Figure 4. The architecture of the proposed MSJR. In the modeling module, Pulse parameter represents the modeling information of the original pulse parameters, Radar capability represents the calculated performance indicators such as typical detection range, resolution, ambiguity resolution, etc. Modulation style represents the pulse change rules identified by ResNet24-1D, such as constant, sliding, stagger, etc.
Sensors 23 03123 g004
Figure 5. ResNet24-1D deep neural network structure for pulse modulation pattern recognition. The three ResNet Units have the same architecture. Here we only use the convolution of one modulation characteristic graph to simplify the network and save training time. Before the full connection layer, a global average pooling was used to extract the important features of each channel.
Figure 5. ResNet24-1D deep neural network structure for pulse modulation pattern recognition. The three ResNet Units have the same architecture. Here we only use the convolution of one modulation characteristic graph to simplify the network and save training time. Before the full connection layer, a global average pooling was used to extract the important features of each channel.
Sensors 23 03123 g005
Figure 6. Ambiguity function of linear frequency modulation signal.The color in the figure does not affect the value of the fuzzy function, but only indicates whether it is directly on the Doppler plane.
Figure 6. Ambiguity function of linear frequency modulation signal.The color in the figure does not affect the value of the fuzzy function, but only indicates whether it is directly on the Doppler plane.
Sensors 23 03123 g006
Figure 7. Fixed distribution normalization. The fixed distribution extracted from the training set is used to translate and zoom all datasets.
Figure 7. Fixed distribution normalization. The fixed distribution extracted from the training set is used to translate and zoom all datasets.
Sensors 23 03123 g007
Figure 8. Training accuracy curve of deep learning network. The horizontal line represents the comparison between different networks, the vertical line represents the comparison between networks of different depths, and the filled area represents the standard deviation of ten trainings, to measure the stability of training.The dotted line indicates the reference line.
Figure 8. Training accuracy curve of deep learning network. The horizontal line represents the comparison between different networks, the vertical line represents the comparison between networks of different depths, and the filled area represents the standard deviation of ten trainings, to measure the stability of training.The dotted line indicates the reference line.
Sensors 23 03123 g008
Figure 9. Test accuracy curves of neural networks. Compared with (a) the step data enhancement work in (b) is adopted during training. The filled area represents the standard deviation of the test accuracy of the same model with the best training effect.
Figure 9. Test accuracy curves of neural networks. Compared with (a) the step data enhancement work in (b) is adopted during training. The filled area represents the standard deviation of the test accuracy of the same model with the best training effect.
Sensors 23 03123 g009
Figure 10. ResNet24-1D modulation mode recognition confusion matrix.
Figure 10. ResNet24-1D modulation mode recognition confusion matrix.
Sensors 23 03123 g010
Figure 11. Accuracy curve of mode recognition classifier. (a) Directly identified on the original data; (b) multiple information sources are used for recognition without feature extraction; (c) recognition using the complete MSJR learning framework.
Figure 11. Accuracy curve of mode recognition classifier. (a) Directly identified on the original data; (b) multiple information sources are used for recognition without feature extraction; (c) recognition using the complete MSJR learning framework.
Sensors 23 03123 g011
Figure 12. Overlapping of behavior mode parameters.
Figure 12. Overlapping of behavior mode parameters.
Sensors 23 03123 g012
Figure 13. Mode recognition confusion matrix using MSJR.
Figure 13. Mode recognition confusion matrix using MSJR.
Sensors 23 03123 g013
Table 1. Parameters of ResNet24-1D.
Table 1. Parameters of ResNet24-1D.
LayerOutput ShapeParameter
Input8000 × 1-
Conv1D4000 × 416
ResNet Unit12000 × 166160
ResNet Unit21000 × 3224,256
ResNet Unit3500 × 6494,592
Batch Normalization500 × 64256
GAP64-
Dense7455
Total parameters125,735
Trainable parameters123,935
Table 2. RPDWS-I dataset. The dataset contains six conventional modulation styles and randomly mixed modulation styles, and each pulse sequence is randomly combined from a given range.
Table 2. RPDWS-I dataset. The dataset contains six conventional modulation styles and randomly mixed modulation styles, and each pulse sequence is randomly combined from a given range.
TypeParameter
TotalNumber of pulses = 16,000
Scope of PRI = 3.3∼165 μs
PRI perturbation <1%
Sample size = 4000
ConstantPRI value = PRImean
StaggerNumber of PRI levels = 2∼8
JitteredPRI deviation = 10∼30%
Dwell and SwitchNumber of pulse groups = 2∼8
Length of pulse groups = 20∼1000
SlidingPRImax = (2∼6) PRImin
Sliding variable cycle = 20∼1000
Direction: up or down
WobbulatedWobbulated amplitude = (0.9∼1.1) PRImean
Cycle length = 20∼1000
HybridA mixture of any two of the above styles
Table 3. RPDWS-II dataset. Each behavior pattern sample is randomly combined from a given range. It should be noted that the test set is generated completely independent of the training set verification set, rather than being divided in the same set of data, to simulate the mismatch between the training set and the test set.
Table 3. RPDWS-II dataset. Each behavior pattern sample is randomly combined from a given range. It should be noted that the test set is generated completely independent of the training set verification set, rather than being divided in the same set of data, to simulate the mismatch between the training set and the test set.
Behavioral ModelWaveformPRI ValuePW ValueDuty RatioModulation
TotalNumber of pulse = 16,000
Scope of RF = 8500∼10,000 MHz
PA = PA/PAmax
Measurement error < 10%
Sample size = 6000
VSHPRF3.3∼10 μs1∼3 μs10∼30%Constant
RWSHPRF3.3∼10 μs1∼3 μs10∼30%Constant
MPRF50∼165 μs1∼20 μs1∼25%Jittered, Dwell and Switch
STTHPRF and MPRF3.3∼125 μs0.1∼20 μs0.1∼25%Constant, Stagger, Jittered,
Sliding, Wobbulated
TASHPRF and MPRF3.3∼165 μs0.1∼20 μs0.1∼25%Search waveform refer to RWS
Tracking waveform refer to STT
Table 4. Average recognition accuracy of neural network.
Table 4. Average recognition accuracy of neural network.
Number of LayersDNNCNNResNet
6-layer65.3%84.5%94.7%
12-layer87.1%80.1%95.9%
18-layer51.9%88.7%97.1%
24-layer49.8%64.6%97.4%
Average58.3%79.5%96.3%
Table 5. Network performance comparison. The comparison results of accuracy under different noise levels, processing time, and model capacity of AlexNet, ConvNet, LeNet, ResNet, MSJR, and VGGNet. The bold part indicates the optimal value.
Table 5. Network performance comparison. The comparison results of accuracy under different noise levels, processing time, and model capacity of AlexNet, ConvNet, LeNet, ResNet, MSJR, and VGGNet. The bold part indicates the optimal value.
ModelLost Pulse (%)Process
Time (s)
Model
Capacity
01020304050
AlexNet0.4070.5510.8070.5610.3170.2832.42520 K
ConvNet0.2960.3740.75250.5460.3970.41211.46954 K
LeNet0.4160.4590.7380.4810.3540.2524.16285 K
ResNet0.8900.8830.8700.8470.7990.66610.931110 K
MSJR0.9520.9420.9410.9290.7650.77014.501220 K
VGGNet0.8620.4790.2950.2830.2720.2589.151680 K
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pan, J.; Xiong, J.; Zhuo, Y. Working Mode Recognition of Non-Specific Radar Based on ResNet-SVM Learning Framework. Sensors 2023, 23, 3123. https://doi.org/10.3390/s23063123

AMA Style

Pan J, Xiong J, Zhuo Y. Working Mode Recognition of Non-Specific Radar Based on ResNet-SVM Learning Framework. Sensors. 2023; 23(6):3123. https://doi.org/10.3390/s23063123

Chicago/Turabian Style

Pan, Jifei, Jingwei Xiong, and Yihong Zhuo. 2023. "Working Mode Recognition of Non-Specific Radar Based on ResNet-SVM Learning Framework" Sensors 23, no. 6: 3123. https://doi.org/10.3390/s23063123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop