Next Article in Journal
Fast Online Coordinate Correction of a Multi-Sensor for Object Identification in Autonomous Vehicles
Previous Article in Journal
An Advanced Algorithm for Higher Network Navigation in Social Internet of Things Using Small-World Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep Learning-Based Satellite Target Recognition Method Using Radar Data

1
Graduate School, Space Engineering University, Beijing 101416, China
2
Space Engineering University, Beijing 101416, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(9), 2008; https://doi.org/10.3390/s19092008
Submission received: 27 March 2019 / Revised: 18 April 2019 / Accepted: 25 April 2019 / Published: 29 April 2019
(This article belongs to the Section Remote Sensors)

Abstract

:
A novel satellite target recognition method based on radar data partition and deep learning techniques is proposed in this paper. For the radar satellite recognition task, orbital altitude is introduced as a distinct and accessible feature to divide radar data. On this basis, we design a new distance metric for HRRPs called normalized angular distance divided by correlation coefficient (NADDCC), and a hierarchical clustering method based on this distance metric is applied to segment the radar observation angular domain. Using the above technology, the radar data partition is completed and multiple HRRP data clusters are obtained. To further mine the essential features in HRRPs, a GRU-SVM model is designed and firstly applied for radar HRRP target recognition. It consists of a multi-layer GRU neural network as a deep feature extractor and linear SVM as a classifier. By training, GRU neural network successfully extracts effective and highly distinguishable features of HRRPs, and feature visualization technology shows its advantages. Furthermore, the performance testing and comparison experiments also demonstrate that GRU neural network possesses better comprehensive performance for HRRP target recognition than LSTM neural network and conventional RNN, and the recognition performance of our method is almost better than that of other several common feature extraction methods or no data partition.

Graphical Abstract

1. Introduction

Space target recognition is a primary function of space surveillance information systems, and satellite recognition is of critical importance on this study, especially for observation satellites. However, few open research achievements have been reported. The difficulty of this problem is that satellites are simply too small or too far away for detailed information to be recognized, and it is relatively hard to obtain effective identification data. With the development of wideband radar, not only can we locate the target position, but also get other useful radar data of targets, such as high-resolution range profile (HRRP) and inverse synthetic aperture radar (ISAR) image [1,2]. An HRRP is the phasor sum of the time returns from different scatterers on the target located within a resolution cell [3], which represents the projection of the complex returned echoes from the target scattering centers onto the range axis [4]. It contains lots of geometric structure information about the target down range, such as scatterers distribution and target size. In addition, compared with ISAR image, HRRPs have the advantages of easy acquisition, storage and processing. That’s why radar HRRP target recognition has gained high attention from radar automatic target recognition (RATR) community [5,6,7]. In summary, this paper focuses on the satellite target recognition based on radar data and proposes a novel recognition method, which mainly consists of data partition and deep learning model.
Feature extraction and selection is a basic and crucial technology for radar target recognition research. It is significant to adopt reasonable and effective features to improve recognition performance. Currently, the targets of HRRP recognition research are basically about ground or aviation targets, such as tanks and airplanes [8,9,10,11,12,13,14,15]. Satellites, as an important space target, have different motion characteristics, one of which is that their motion must follow Kepler’s law. Besides, because the fuel carried by satellites is limited, the orbital maneuver range of most satellites is restricted, which makes satellite orbits relatively stable [16,17]. Orbit information is easily accessible and non-burdensome because target range and position measurement is a basic function of radar, therefore, orbit information is a distinct and accessible feature for recognizing the satellites whose eccentricity is very small, and could be introduced into radar data partition when multiple satellites need to be recognized. In addition, radar observation pitch and azimuth angles are also important and helpful information. Utilizing radar observation angles has many benefits for target recognition, such as reducing search range and computation, relaxing attitude sensitivity and improving recognition rate [18,19]. Radar observation angular domain division have been studied in [8,11,20,21]. Uniform frame segmentation method is used in [8,11] but has been proved simple and unreasonable [22]. Angular domain division method based on statistical characteristics is studied in [20], but it would generate many matching templates and need huge computation burden and storage requirement. Correlation coefficient is firstly introduced to solve this problem in [21] and could measure similarity between HRRPs to a certain extent. However, other information may also be helpful to measure similarity and could be applied, such as angular distance. Therefore, for the data partition, the proposed method will utilize orbit information as a powerful feature, and a hierarchical clustering method with a novel distance metric, namely normalized angular distance divided by correlation coefficient (NADDCC), will be applied to segment radar angular domain.
In addition to the above available satellite and radar information, it is still necessary to further mine information in HRRP data to raise the recognition accuracy. Many studies have been done on HRRP feature extraction and selection methods. In the early days, researchers often calculated FFT-magnitude, power spectrum and a variety of high-order spectrum of HRRP data, and used them as the features of classifier for target recognition [8,9,10]. Although these engineered features could play a part in target recognition, they are dependent on researchers’ experience and techniques. Other than the features of artificial selection, machine learning algorithms have been widely utilized to extract features based on high-dimensional HRRP data [11,12,13,14,15]. Principal component analysis (PCA) is applied to extract the complex HRRPs’ feature subspace within each target-aspect sector in the literature [11]. Dictionary learning is adopted to extract the features of HRRP data, which possess high noise-robust and discrimination [12,13]. Manifold learning is employed to reduce the feature dimensions of radar HRRP [14,15]. These methods can extract appropriate features in some cases, but they are all shallow architectures that may not represent the essence of radar HRRP. Thereby, how to automatically extract the deep abstract features, which can play an important role in target recognition, has become a significant research issue.
The deep learning theory [23] advanced by Hinton could solve the abovementioned problem effectively. Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction [24]. Because of its powerful recognition or prediction ability, deep learning methods have consistently been used in many applications, such as computer vision [25], speech recognition [26], networking and health care [27,28,29]. Some deep learning structures applied in several recent papers have been demonstrated useful for radar target recognition, such as autoencoder and its varieties [30,31], convolutional neural network (CNN) [32] and recurrent neural network (RNN) [33,34]. Due to the unique structure of RNN, it has been widely applied to process sequential data, such as action recognition [35], scene labelling [36], and language processing [37], and has achieved impressive results [38]. However, it is founded that simple conventional RNN could not learn on a wide range of dependencies, and gradients of distant time steps do not play a role in learning because of gradient vanishing. To solve this problem, long short-term memory (LSTM) and gated recurrent unit (GRU) architecture have been designed to learn long term reliance on sequential data. GRU is a variety of LSTM and retains the LSTM’s resistance to the problem of gradient vanishing. Moreover, the internal structure of GRU is simple and it requires less computation to update the hidden state, which makes the training speed faster [37,39,40]. So GRU neural network will be applied in this satellite recognition method to extract and select the deep abstract features of radar sequential HRRP data.
In this paper, a novel satellite target recognition method is proposed based on radar data. This method could make full use of radar data and apply deep learning technology to extract highly distinguishable features. Its features are summarized as follows:
(a)
Satellite orbital altitude is introduced as a distinct and accessible feature for radar data partition;
(b)
Radar observation angles information is fully utilized, and a hierarchical clustering method based on a new distance metric (NADDCC) is applied to segment angular domain to improve recognition rate;
(c)
A novel end-to-end GRU-SVM model is designed, which uses radar HRRP data as input and target class as output, and firstly applied to identify targets based on radar HRRP data. GRU neural network is applied as a deep features extractor and support vector machine (SVM) is used as a classifier in this model. Compared with some deep neural networks such as CNN, autoencoder and denoising autoencoder, and shallow learning algorithms such as PCA [11], dictionary learning [12,13] and manifold learning [14,15], the presented model can extract the deep abstract features of HRRPs and obtain better recognition results.
The rest of this paper is organized as follows: in Section 2, the information contained in satellite orbit and radar data is analysed. In Section 3, we introduce GRU, SVM and the construction of the GRU-SVM model, and present the overall flow chart of this recognition method. In Section 4, recognition results are provided and the performance under different recognition methods and conditions is compared and analysed. In Section 5, some conclusions are drawn.
Notations: To simplify the presentation, we define the following notations used in this paper. We use bold lower case letters to represent a vector, e.g., μ     C D , and use bold upper case letters to represent a matrix, e.g., M     C i × j . The acronyms used in this paper are summarized in Table 1 for the sake of readability.

2. Analysis of Satellite and Radar Information

2.1. Description and Preprocessing of HRRP

HRRP is the amplitude of echo summation for target scattering centers in each range cell of wideband radar. Figure 1 shows the illustration of a HRRP sample from a satellite target. High resolution radar operates in microwave frequency band, and the size of targets or their components is much larger than the wavelength of radar. In this case, the echo characteristics of targets can be calculated by using a simplified scattering center model [3,4,9,41,42]. Therefore, for complex targets such as a satellite, the projection of an object on radar line of sight can be divided into many range cells by high resolution radar. According to the scattering center model, the scatterers in different range cells will rotate in the same way when a satellite target rotates, which causes that the echo amplitudes between range cells have certain correlation. Besides, windowing processing of the returned echoes before getting HRRP data and multiple reflections phenomena of measured HRRP data would enhance the correlation between adjacent range cells. Therefore, there is a certain temporal correlation between range cells and radar HRRP could be seen as sequential data, which is suitable to be learned by RNN-based neural networks.
The radar signatures, which return from multiple scattering centers within the same range cell, will be coherently summed as a single signature for that range cell. According to related literature [3,9], suppose the transmitted signal is s ( t ) e j 2 π f c t , the n - th complex echo in the d - th range cell ( d = 1 ,   2 ,   ,   D ) in the baseband can be approximated as:
x ˜ d ( t ,   n )   s ( t ) i = 1 L d σ di e j { ( 4 π / λ ) [ R ( n ) + Δ γ di ( n ) ] }
where s ( t ) is the complex envelop which could be approximated as unchanged for all scatterers in one range cell. λ represents the wavelength of wideband radar and f c is the carrier frequency of radar signal. L d denotes the number of scatterers in the d - th range cell. σ di represents the intensity of the i - th scatterer in the d th range cell. R ( n ) is the radial distance between the radar and target reference center in the n - th echo. Δ γ di ( n ) is the radial displacement of the i - th scatterer of the d - th range cell in the n - th echo. Usually, s ( t ) is a rectangular pulse signal with unit intensity and could be omitted. After eliminating the initial phase of the n - th echo e - j ( 4 π / λ ) R ( n ) , the n - th HRRP can be defined as:
x ( n ) = [ x 1 ( n ) , x 2 ( n ) , , x D ( n ) ] = [ | i = 1 L 1 σ 1 i e j ϕ 1 i ( n ) | , | i = 1 L 2 σ 2 i e j ϕ 2 i ( n ) | , , | i = 1 L D σ Di e j ϕ Di ( n ) | ]
Several sensitivity issues of HRRP should be focused on when the HRRP target recognition task is carried out. The first one is time-shift sensitivity. To decrease the computation complexity, HRRP is only a portion of received radar echo extracted by a range window, which contains the target signal. Thus the position of the target signal in HRRP would change with the measurement. However, it would be better for feature learning that all the training samples meet a uniform parameter model. So we adopt envelope alignment method [43] as time-shift compensation technique in this paper, which is achieved based on the summation average of multiple HRRPs cross-correlations. The second one is amplitude-scale sensitivity. It is caused by the fact that many factors could influence the intensity of an HRRP, such as target distance, radar transmitting power, radar antenna gain, radar receiver gain and radar system losses. It makes that HRRPs measured by different radars or under different conditions would have different amplitude-scales. In order to deal with amplitude scale sensitivity, each HRRP is normalized by energy normalization method [3,9]. Suppose an HRRP is defined as x ( n )   =   [ x 1 ( n ) , x 2 ( n ) , , x D ( n ) ] , then its energy normalization preprocessing result x ˜ ( n ) is shown as follows:
x ˜ ( n ) = [ x ˜ 1 ( n ) ,   x ˜ 2 ( n ) ,     ,   x ˜ D ( n ) ] = [ x 1 ( n ) i = 1 D x i ( n ) 2 ,   x 2 ( n ) i = 1 D x i ( n ) 2 ,     ,   x D ( n ) i = 1 D x i ( n ) 2 ]
After the above preprocessing step, the HRRP sample examples of satellites are shown in Figure 2. The last and toughest one is called target-attitude sensitivity. This issue will be analyzed in detail in Section 2.3 and further alleviated.

2.2. Analysis and Statistics of Satellite Orbit

Currently, thousands of artificial satellites move around the Earth for the purposes of communication and navigation, information relay, missile warning, on-orbit service and so on. Among them, observation satellites are the key targets, which need to be identified for space surveillance information systems. Unlike ground or aviation targets, satellites’ motion must follow Kepler’s law, that is, they basically move along certain orbit. As shown in Figure 3a, satellite orbit is mainly described with the following parameters:
(1)
Semi-major axis ( a ): the distance from the apogee or perigee to orbital center which describes the size of satellite orbit;
(2)
Eccentricity ( e ): describes the shape of satellite orbit;
(3)
Inclination ( i ): the angle between the orbital plane and the equatorial plane which determines the accessible area of a satellite;
(4)
Longitude of ascending node ( Ω ): the right ascension at the intersection of the orbital plane and the equatorial plane.
(5)
Argument of perigee ( ω ): angular distance of orbit perigee and ascending node.
Based on the above parameters, the altitude range of a satellite can be computed by:
{ H max   =   a · ( 1   +   e )     R e H min   =   a · ( 1     e )     R e
where H max and H min respectively refer to the maximum and minimum altitude of a satellite. R e denotes the radius of Earth. In this way, the variation of satellite altitude difference with eccentricity can be calculated when the value of semi-major axis is given, as shown in Figure 3b. It can be seen that only when eccentricity is small enough can the orbital altitude information be used for identification, so it is still necessary to investigate the eccentricity distribution of current observation satellites. UCS Satellite Database [44] has made detailed statistics of satellites currently orbiting Earth. Eccentricity and apogee altitude statistical results of observation satellites are shown in Figure 3c,d. It shows that the eccentricity of most satellites is lower than 0.01, especially for optical or radar imaging satellites. Therefore, orbital altitude information is an available and useful feature for recognizing observation satellites.
For non-cooperative satellite targets, it is difficult to compute orbital altitude by orbital parameters because orbital parameters are not necessarily known or constant. However, it may become easy when radar is applied. For radar, distance measurement is a basic function and its observation angles are known. Satellite orbital altitude could be obtained by multiple coordinate transformation based on radar measurement data and the main coordinate transformation diagrams are shown in Figure 4. The detailed orbital altitude calculation process is shown in Appendix A.

2.3. Target-Attitude Sensitivity and Radar Observation Angular Domain Segmentation

Target-attitude sensitivity is one of the most difficult problems in radar HRRP target recognition research. According to the scattering center model [3,4,9,41,42], the variation of target attitude will lead to different range shifts for different scattering centers on the target, even within the attitude region where the scattering center structure remains unchanged (that is, without migration throuth resolution cells, MTRC). Specifically, for the HRRP of the m th returned echo, suppose it is x ( m )   =   [ x 1 ( m ) , x 2 ( m ) , , x D ( m ) ] , then the echo power of the n - th range cell could be computed by:
| x n ( m ) | 2   =   x n ( m ) x n * ( m )   =   i = 1 L n σ ni 2   +   2 i = 2 L n k = 1 i 1 σ ni σ nk cos [ θ nik ( m ) ]
where * represents complex conjugate operation. σ ni denotes the intensity of the i - th scatterer in the n - th range cell. i = 1 L n σ ni 2 is the conjugate product of the sub-echo for all scatterers in the n - th range cell, which represents the intensity sum of each scatterer and is relatively stable. However, cos [ θ nik ( m ) ] of the second item, which is called cross-term, will change with the variation of m , where θ nik ( m ) represents the phase difference of the i - th scatterer and the k - th scatterer of the n - th range cell in the m - th returned echo. Therefore, an HRRP, which is the amplitude of coherent sum of the complex returned echoes from scatterers in a range cell, can be changed substantially.
The target-attitude sensitivity problem makes it hard to recognize satellite targets base on HRRP data and needs to be focused on. It has been founded that average range profile [22] is helpful to improve attitude stability of HRRP, because the sum of the irrelevant echo power of the cross-terms will greatly weaken the effect of cross-items. As mentioned earlier, the correlation coefficient between average range profiles can be used to measure similarity of HRRPs [21]. For a set of HRRP samples { x ( 0 ) ,   x ( 1 ) ,   ,   x ( M 1 ) } , which are translationally aligned and without MTRC, we represent them as x ( i )   =   [ x 1 ( i ) , x 2 ( i ) , , x D ( i ) ] ,   i   =   0 , 1 , , M 1 , then its average range profile is [22] defined as:
μ   =   [ 1 M m = 0 M 1 | x 1 ( m ) | 2 ,   1 M m = 0 M 1 | x 2 ( m ) | 2 ,   ,   1 M m = 0 M 1 | x D ( m ) | 2 ]
Research also suggests that utilizing radar observation angles information is beneficial, such as reducing search range and computation, relaxing target-attitude sensitivity and improving recognition rate [18,19].
Through the above analysis, for the target-attitude sensitivity problem of HRRP, we propose a hierarchical clustering method with a novel distance metric, namely the normalized angular distance divided by correlation coefficient (NADDCC), to segment radar observation angular domain. This distance metric includes both angular distance information and correlation coefficient between HRRP average range profiles, which could measure similarity between HRRP average range profiles better. For average range profiles μ i = [ μ i ( 1 ) ,   μ i ( 2 ) ,   ,   μ i ( D ) ] whose observation azimuth and elevation angles are θ i and ε i , and μ j = [ μ j ( 1 ) ,   μ j ( 2 ) ,   ,   μ j ( D ) ] whose observation azimuth and elevation angles are θ j and ε j , their angular distance d angle ( μ i ,   μ j   ) and correlation coefficient ρ ( μ i ,   μ j   ) are defined as follows:
d angle ( μ i ,   μ j   )   =   ( θ i θ j ) 2   +   ( ε i - ε j ) 2
ρ ( μ i ,   μ j   )   =   μ i μ j T μ i 2 μ j 2   =   n = 1 D [ μ i ( n ) μ j ( n ) ] μ i 2 μ j 2
The larger the correlation coefficient and the smaller the angular distance, the higher the similarity of HRRP average range profiles. In addition, the distance metric applied for hierarchical clustering usually need to satisfy some necessary properties, such as non-negativity, identity and symmetry. In order to make the two distances work equally, they are normalized by means of dividing by their respective maximum. Taking into account the above considerations, the distance metric presented in this paper is designed as follows:
d ( μ i , μ j ) = d angle ( μ i ,   μ j   ) / max ( d angle ) ρ ( μ i ,   μ j   ) / max ( ρ ) = ( θ i θ j ) 2 + ( ε i ε j ) 2 · μ i 2 μ j 2 · max ( ρ ) n = 1 D [ μ i ( n ) μ j ( n ) ] · max ( d angle )
Hierarchical clustering algorithm adopts bottom-up aggregation strategy. Firstly, each HRRP average range profile is regarded as an initial cluster, and then two nearest clusters are found and merged at each step of algorithm operation. The process is repeated until the number of preset clusters is reached. The clustering process can be summarized as follows:
Hierarchical Clustering Algorithm
Input: HRRP average range profile sample set S   =   { μ 1 ,   μ 2 ,   ,   μ m } ;
            distance metric d avg ( C i ,   C j )   =   1 | C i | | C j | μ i C i μ j C j d ( μ i ,   μ j ) ;
            cluster number k .
Process: for j   =   1 ,   2 ,   ,   m do
                     C j   =   { μ j }
                end for
                for i   =   1 ,   2 ,   ,   m do
                     for j   =   i + 1 ,   ,   m do
                          M ( i ,   j )   =   d avg ( C i ,   C j ) ;
                          M ( j ,   i )   =   M ( i ,   j )
                     end for
                end for
                set the number of current clusters: q = m
                while q   >   k do
                     find the two nearest cluster C i * and C j * ;
                     merge C i * and C j * : C i *   =   C i *     C j * ;
                     for j   =   j * + 1 ,   j * + 2 ,   ,   q do
                          renumber cluster C j as C j - 1
                     end for
                     delete line j * and column j *   of matrix M
                     for j   =   i + 1 ,   ,   q - 1 do
                          M ( i * ,   j )   =   d avg ( C i * ,   C j ) ;
                          M ( j ,   i * )   =   M ( i * ,   j )
                     end for
                     q   =   q     1
                 end while
Output: clusters C   =   { C 1 ,   C 2 ,   ,   C k }

3. GRU-SVM Model

The designed GRU-SVM model is a combination of a GRU neural network as a deep feature extractor and SVM as a classifier. It makes the best of the advantages of the GRU deep neural network and SVM to extract the deep abstract features and complete an accurate classification. This model is described in detail below.

3.1. GRU

As mentioned above, GRU is designed in [37,45] to learn long term reliance on sequential data, and its overall performance is better than LSTM and simple conventional RNN [39,40]. Figure 5 shows a GRU model. There are only two gates in a GRU, namely the update gate z and the reset gate r . The update gate is utilized to modulate the previous information inside the unit. The larger the value of update gate, the more the status information of the previous moment insides. The reset door is used to control how much previous state information will be forgotten. The smaller the value of the reset gate, the more the previous state information is forgotten.
The update gaze z t and reset gate r t at the time t are defined as:
{ z t   =   σ ( W z x t   +   U z h t 1 ) r t   =   σ ( W r x t   +   U r h t 1 )
where W and U are weight matrices. x denotes input data. The hidden state h t and candidate hidden state h ˜ t in GRU are calculated respectively as follows:
{ h t   =   ( 1     z ) h t 1   +   z t h ˜ t h ˜ t   =   tan h ( W h x t   +   U t ( r t   *   h t 1 ) )
where * represents element-wise product. The σ ( · ) and tan h ( · ) are two different activation functions which can be defined as:
σ ( x )   =   1 1   +   e x tan h ( x )   =   1     e 2 x 1   +   e 2 x
In this paper, GRU is employed in the GRU neural network to extract effective features based on HRRP sequential data.

3.2. SVM

The support vector machine (SVM) was developed by Vapnik [46] for binary classification. Its objective is to find the optimal hyper-plane f ( w , x )   =   w   ·   x   +   b to separate two classes in a given dataset, where x is the feature vector. SVM learns the parameters w and b by solving the following constrained optimization problem:
min 1 p w T w   +   C i = 1 p ξ i . s . t   y i ( w   ·   x   +   b )     1     ξ i ξ i     0 ,   i   =   1 , ,   p
where w T w is the Manhattan norm, C is the penalty parameter, and   ξ is the cost function. The corresponding unconstrained optimization problem of Equation (18) is as follows:
min 1 p w T w   +   C i = 1 p max ( 0 ,   1   -   y i ( w T   x i   +   b ) )
where y is the actual label, and w T   x   +   b is the predictor function. This equation is known as L1-SVM, with the standard hinge loss. Its differential counterpart L2-SVM is given by the following equation:
min 1 p w 2 2   +   C i = 1 p max ( 0 ,   1     y i ( w T   x i   +   b ) ) 2
where w 2 is the Euclidean norm (also known as L2 norm), with the squared hinge loss.
Despite being intended for binary classification, SVM may be used for multi-classification as well. One approach to achieve this is the use of kernel tricks, which convert a linear model into a non-linear model by applying kernel functions. However, we just use LSVM instead of utilizing kernel tricks in this paper, because LSVM does not employ any feature extraction and transformation and can serve as a simple baseline for evaluating the quality of extracted features. A one-vs-one scheme is employed to achieve multi-classification in this paper, which establishes the binomial classifier for every two classification.

3.3. GRU-SVM Model Construction

In this paper, a novel end-to-end GRU-SVM model has been designed and firstly applied to recognize targets based on radar HRRP data. The structure of GRU-SVM model is shown in Figure 6. The composition of this model includes two parts: GRU neural network as a feature extractor and LSVM as a classifier. GRU neural network is constituted of input layer, four hidden layers and output layer, where GRU hidden layer and fully connected layer (dense layer) are included. And the input layer, two GRU hidden layers and a fully connected layer make up the encoder module, whose output are defined as the features extracted by GRU neural network. In order to make the training model more accurate, we apply the bidirectional scheme demonstrated in reference [47] in GRU hidden layers (see Bid-GRU layer in Figure 6). In addition, two fully connected layers have been employed after the encoder to extract good features. And the last layer, also called output layer, would output the satellite classifications by adopting softmax activation function. That is, GRU neural network is trained in a supervised way. The reasons why we choose to apply the softmax classifier to train GRU neural network are that it’s an excellent multi-class classifier and its common loss function, namely categorical cross-entropy loss, is more sensitive to classification output than the hinge loss of LSVM, which means that it is always optimizing the network parameters to reduce loss during training. Therefore, the utilization of softmax classifier may make the features extracted by the encoder more highly distinguishable. By training GRU neural network, the encoder will produce the length-fixed feature vectors which contain sufficient information for target recognition. LSVM classifier takes feature vectors as input and produce classification results. It mainly has two roles: one the one hand, LSVM classifier could have outstanding generalization performance for the testing data after it is trained with the extracted features and their corresponding labels; on the other hand, LSVM classifier is applied as a simple baseline for evaluating the quality of features extracted by different methods here, because it does not employ any feature extraction and transformation. However, the softmax classifier in this GRU neural network may be not suitable for comparing the quality of extracted features on account of the existence of nonlinear activation function. Linear φ ( · ) , relu ϕ ( · ) and softmax ψ ( · ) activation functions are employed in this model and defined as follows:
φ ( x )   =   i x i w i   +   b ϕ ( x )   =   max ( 0 ,   x ) ψ ( x ) j   =   e x j k = 1 K e x k
After the analysis and description of Section 2 and Section 3, the overall framework of this proposed satellite target recognition method is shown in Figure 7. It can be divided into three parts, namely training process, testing process and the methods and techniques used therein. Orbital altitude calculation and observation angular domain segmentation are applied to divide HRRP data, including training samples and testing samples. The detailed orbital altitude calculation process is shown in the Appendix A. On this foundation, GRU neural network is trained to get the deep abstract features based on divided HRRP training data. Then, the classification results of testing data could be obtained by the trained GRU neural network and LSVM.

4. Experimental Results and Discussion

In this section, test experiments will be carried out to obtain the performance of the proposed recognition method. After dividing training data and completing the training process of GRU-SVM model, the recognition accuracy of testing data set will be gotten according to the testing process in Figure 7. Furthermore, performance testing and comparison experiments of different conditions and recognition methods also have been done to better illustrate the advantages of this recognition method.

4.1. Data Generation and Partition

Considering the difficulty of obtaining satellite HRRP measured data, we utilize reliable simulation radar HRRP data from ten satellites that simulated by an X-band radar with a center frequency of 10 GHZ and a bandwidth of 1 GHz. The main parameters of radar and these satellites are listed in Table 2 and the detailed flow chart of radar data generation is shown in Figure 8. The observational relationship between satellite and radar is calculated based on their parameters. Then, radar cross section (RCS) and echo are computed and radar data is obtained. In our experiments, each satellite target has 70,000 HRRP samples and each HRPP is a 300-dimensional vector. The data preprocessing operation has been done as described in Section 2.1. 80% of radar data will be applied as training set and others as testing set. For training set, data partition would be done by orbital altitude calculation and observation angular domain segmentation. Figure 9 shows the data partition results. The orbital altitude of this ten satellites is divided into three ranges in this paper, namely H Sat 500 km , 500 < H Sat 1000 km and H Sat > 1000 km . Meanwhile, a hierarchical clustering method with a novel NADDCC distance metric has been implemented for HRRP data in each orbital altitude range to get clustering datasets. It should be noted that a set of radar observation angles is shared by continuous 700 radar HRRP data. Thus, we could be decide which cluster the test data belongs to when determining its altitude range and angles cluster.

4.2. Training Assessment of GRU Neural Network

After data partition, GRU neural network could be trained with these clustering datasets as input. In order to train the network faster and more accurately, we apply the following deep neural network training techniques in this paper:
(1)
20% of training data is used as validation set to adjust the hyper-parameters;
(2)
Drop-out is employed for the two GRU layers to avoid the problem of overfitting and set to 0.25;
(3)
Batch normalization is inserted after each layer to accelerate the training.
For multi-classification problems, recognition accuracy, categorical cross-entropy loss Loss CC and mean absolute error (MAE) loss Loss MAE are often applied to assess classification result. They are defined as follows:
accuracy   =   N / M Loss CC   =   i = 1 M j = 1 m y ij log y ^ ij Loss MAE   =   1 M i = 1 M | y i y ^ i |
where M is the total number of current training samples and N is the number of the samples predicted correctly. y ^ i denotes the prediction value of the ith sample and y i denotes expected value. m represents the class number and is usually greater than or equal to 3. In order to reduce the loss and improve accuracy in the training process, it is necessary to choose a suitable optimizer for GRU neural network. Adam is an excellent optimizer which combines the main advantages of the previous deep learning optimizers AdaGard and RMSProp. Thus, the Adam optimizer is used with the initial learning rate 1 × 10 3 . When evaluation indicator is not improving, the learning rate will decrease in multiple. The training records of 100 epochs are shown in Figure 10. It could be found that the training and validation accuracy raises with the increase of training epoch and converges to a high accuracy value. Correspondingly, the categorical cross-entropy loss and MAE of training and validation data decrease with the increase of training epoch. These results all confirm that GRU neural network is well trained for all training clustering datasets.

4.3. Recognition Results and Comparative Analysis

The performance of this recognition method could be obtained on the basis of data partition and GRU neural network training, and a full comparative analysis of the recognition results under different conditions and recognition methods is also made in this section.

4.3.1. Classification Results of this Recognition Method

After completing radar data partition and GRU neural network training, a series of trained neural network models can be obtained based on training datasets of different clusters. According to the overall recognition framework in Figure 7, the orbital altitude calculation and observation angles assignment of testing data will be carried out, which is for determining which training model to extract their deep abstract features. Then, classification testing would be implemented by LSVM based on these features and corresponding recognition results of testing data could be gotten. The confusion matrices of these data tested by all training models are shown in Figure 11 and the corresponding recognition accuracy are listed in Table 3. It can be seen that testing data could be correctly divided into its corresponding trained model by orbital altitude calculation and observation angles assignment. Meanwhile, these ten satellites can be well recognized and achieve a total accuracy of 99.2%.

4.3.2. Comparative Analysis of Classification Results

For the feature extraction based on radar HRRP data of satellites, it is unclear whether the GRU neural network performs better than the conventional RNN or LSTM neural networks or not. Therefore, we still need to conduct performance testing experiments for these three RNN-based neural networks. In this paper, we focus on the quality of features extracted by them and the training time they need, where feature quality could be reflected by the recognition accuracy after LSVM classifier. Radar HRRP data under different partition conditions are applied as the input of these neural networks. To better compare their training complexity, they are trained with the same training data scale, computing resources and network training parameters when the same radar HRRP data are used as input. The performance testing results are shown in Table 4.
These results demonstrate that the features extracted by GRU neural network are the most distinguishable, and thus GRU-SVM model achieve the maximum recognition accuracy. Although conventional RNN needs the least training time, it cannot learn a good presentation of the satellite HRRP sequential data because of the network structure limitation and vanishing gradient. The performance in satellite HRRP feature extraction of the LSTM neural network is inferior to that of GRU neural network, and LSTM needs the maximum training time owing to its complex network structure. Therefore, considering the above comparison results, we choose GRU neural network as the feature extractor of satellite HRRP sequential data and just need to compare the performance of GRU neural network and other common feature extraction methods later.
In order to demonstrate the effect and advantages of this proposed recognition method, recognition performance for testing data is compared when different methods are applied to recongnize satellites under different conditions. Four deep neural networks, namely GRU neural network, CNN, autoencoder (AE) and denoising autoencoder (DAE), and three shallow models, including PCA, dictionary learning (DL) and manifold learning (ML), serve as feature extractor and LSVM is used to classify these satellites based on these extracted features. To ensure fairness in performance comparison, the number of some layers in these neural networks should be as same as possible, such as GRU hidden layer, CNN layer and encoder/decoder layer; These seven feature extraction methods could all reduce the 300-dimensional HRRP samples to same dimension, for example 64 dimensions in this paper. Table 5 shows the detailed recognition accuracy of these seven methods under various conditions and the corresponding statistical comparable results are shown in Figure 12. We can make the following conclusions from these two charts:
(1)
Orbital altitude calculation and observation angles clustering technology are favourable for improving the recognition rate of satellites for all seven methods, which verifies the validity of radar data partition;
(2)
Compared with the latter six methods, GRU-SVM model has good recognition performance for these ten satellites. Therefore, its total recognition accuracy rate is almost highest among these seven methods no matter whether orbital altitude calculation or observation angles clustering is applied.
Although the classification results of LSVM could prove the feature extraction validity of GRU neural network in some aspects, it is still expected to further display the distribution of extracted features for these seven recognition methods. Therefore, dimension reduction visualization technology is employed in this paper, which can map high dimensional feature data to two or three dimensions. At this time, the distribution of extracted features can be seen intuitively. Figure 13 shows dimension reduction distribution of features extracted based on one cluster training data ( H Sat > 1000 km , Cluster = No . 3 ) for these seven methods. It can be found in Figure 13a that the GRU neural network has the best feature distribution result because the dimension reduction features of three satellites are separated from each other. However, other methods have more or less intersections between different satellite classes, especially the last five methods. It is confirmed that the features extracted by the GRU neural network are more effective and highly discriminative.

5. Conclusions

In this paper, a novel satellite target recognition method based on radar data is proposed. It mainly includes two modules of data partition and a deep classification model. Satellite orbital altitude is introduced as a distinct and accessible feature because it is easy to calculate based on radar data and has relative stability. A hierarchical clustering method based on a new NADDCC distance metric is utilized to segment the radar observation angular domain. These two technologies can effectively complete data partition. Then, a GRU-SVM model is designed for radar HRRP satellite recognition, which is comprised of a deep GRU neural network as a feature extractor and LSVM as a classifier. The GRU neural network training records and feature visualization results all confirm that this GRU neural network could extract more deep and abstract features and these features have better separability. Furthermore, performance testing and comparison experiments also demonstrate that GRU neural network possesses better comprehensive performance for HRRP target recognition than LSTM neural network and conventional RNN; data partition can improve the recognition rate of satellites, and the recognition performance of our satellite target recognition method is almost better than that of other several common feature extraction methods or no data partition.

Author Contributions

W.L. proposed this recognition framework for radar satellite, performed the experiment and completed this manuscript. Y.Z. and C.L. revised this manuscript. C.X. and Y.H. provided helpful advice.

Acknowledgments

The work in this paper has been supported by the National Natural Science Foundation of China (61304228).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Satellite Orbital Altitude Calculation Method

Satellite orbital altitude could be obtained by multiple coordinate transformation based on radar measurement data. The Figure 4 shows these coordinate transformation diagrams. It mainly involves the following coordinate systems:
(1)
Geocentric Cartesian coordinate system: its origin O E is situated at the center of the Earth. O E X E axis is located in the equatorial plane and points to the meridian at the Greenwich Observatory. O E Z E axis is perpendicular to the equatorial plane, coincident with Earth’s rotation axis and points to the north pole. O E Y E axis is located in the equatorial plane and its direction satisfies the right-handed rectangular coordinate system criterion.
(2)
Radar body Cartesian coordinate system: its origin S is situated at the radar station. SX S axis is located in the ground plane of the radar station and points to the south. SY S axis is located in the ground plane of the radar station and points to the east. SZ S axis passes the radar station, along its plumb line and points to the zenith.
The detailed coordinate transformation process is as follows:
(a) Conversion of radar polar coordinate system into radar body Cartesian coordinate system
Set the coordinates of satellite target in the radar polar coordinate system to ( ρ ,   θ ,   ε ) , where ρ is the slant distance from satellite to radar station. θ ,   ε represent the radar observation azimuth and elevation angle, respectively. These parameters could be obtained in the process of radar measurement. Then, the coordinates ( x 1 ,   y 1 ,   z 1 ) of the satellite in the radar body Cartesian coordinate system could be calculated by the following formula:
{ x 1   =   ρ · cos θ · cos ε y 1   =   ρ · sin θ · cos ε z 1   =   ρ · sin ε
(b) Conversion of geodetic coordinate system to geocentric Cartesian coordinate system
Set the geodetic coordinates of the radar station are ( λ S ,   φ S ,   H ) , where λ S , φ S , H represent the longitude, latitude and altitude of the radar station, respectively, which are known for certain radar station. Its corresponding coordinates ( x ,   y ,   z ) in the geocentric Cartesian coordinate system can be computed by the following formula:
{ x   =   ( N + H ) · cos λ S · cos φ S y   =   ( N + H ) · sin λ S · cos φ S z   =   ( N · ( 1 - e 2 ) + H ) · sin φ S
where N denotes the radius of prime vertical and is calculated by N = a / ( 1 - e 2 sin 2 φ S ) . a , e represent the semi-major axis and oblateness of earth.
(c) Conversion of radar body Cartesian coordinate system to geocentric Cartesian coordinate system
Set the coordinates of satellite in the radar body Cartesian coordinates system are ( x 1 ,   y 1 ,   z 1 ) , suppose its corresponding coordinates in the geocentric Cartesian coordinates system are ( x e ,   y e ,   z e ) , then the conversion formula is as:
[ x e y e z e ]   =   R T [ x 1 y 1 z 1 ]   +   [ x y z ]
where R T is rotation matrix and calculated as:
R T   =   [ sin φ S cos λ S - sin λ S cos φ S cos λ S sin φ S sin λ S cos λ S cos φ S sin λ S cos φ S 0 sin φ S ]
Once the coordinates of satellite in the geocentric Cartesian coordinate system are known, it is easy to calculate the altitude of the satellite. The calculation formula is as follows:
H sat   =   ( x e ) 2 + ( y e ) 2 + ( z e ) 2     R e

References

  1. Li, C.; Bao, W.M.; Xu, L.P.; Zhang, H. Clustered Multi-Task Learning for Automatic Radar Target Recognition. Sensors 2017, 17, 2218. [Google Scholar] [CrossRef] [PubMed]
  2. Zhang, L.; Qiao, Z.J.; Xing, M.D.; Li, Y.C.; Bao, Z. High-Resolution ISAR Imaging With Sparse Stepped-Frequency Waveforms. IEEE Trans. Geosci. Remote Sens. 2011, 49, 4630–4651. [Google Scholar] [CrossRef]
  3. Bogner, R.E. Radar target classification of commercial aircraft. IEEE Trans. Aerosp. Electron. Syst. 1996, 32, 598–606. [Google Scholar]
  4. Du, L.; Wang, P.H.; Liu, H.W.; Pan, M.; Chen, F.; Bao, Z. Bayesian Spatiotemporal Multitask Learning for Radar HRRP Target Recognition. IEEE Trans. Signal Process. 2011, 59, 3182–3196. [Google Scholar] [CrossRef]
  5. Eom, K.B.; Chellappa, R. Noncooperative Target Classification Using Hierarchical Modeling of High-Range Resolution Radar Signatures. IEEE Trans. Signal Process 1997, 45, 2318–2327. [Google Scholar] [CrossRef]
  6. Jiang, Y.; Li, Y.; Cai, J.J.; Wang, Y.H.; Xu, J. Rubust Automatic Target Recognition via HRRP Sequence Based on Scatterer Matching. Sensors 2018, 18, 593. [Google Scholar] [CrossRef]
  7. Guo, Y.; Xiao, H.T.; Kan, Y.Z.; Fu, Q. Learning using privileged information for HRRP-based radar target recognition. IET Signal Process. 2018, 12, 188–197. [Google Scholar] [CrossRef]
  8. Du, L.; Liu, H.W.; Bao, Z.; Xing, M.D. Radar HRRP target recognition based on higher order spectra. IEEE Trans. Signal Process. 2005, 53, 2359–2368. [Google Scholar]
  9. Du, L.; Liu, H.W.; Wang, P.H.; Feng, B.; Pan, M.; Bao, Z. Noise Robust Radar HRRP Target Recognition Based on Multitask Factor Analysis With Small Training Data Size. IEEE Trans. Signal Process. 2012, 60, 3546–3559. [Google Scholar]
  10. Feng, B.; Du, L.; Liu, H.W.; Li, F. Radar HRRP target recognition based on K-SVD algorithm. In Proceedings of the 2011 IEEE CIE International Conference on Radar, Chengdu, China, 24–27 October 2011; pp. 642–645. [Google Scholar]
  11. Zhang, X.D.; Shi, Y.; Bao, Z. A new feature vector using selected bispectra for signal classification with application in radar target recognition. IEEE Trans. Signal Process. 2001, 49, 1875–1885. [Google Scholar] [CrossRef]
  12. Liu, H.W.; Feng, B.; Chen, B.; Du, L. Radar high-resolution range profiles target recognition based on stable dictionary learning. IET Radar Sonar Navig. 2016, 10, 228–237. [Google Scholar] [CrossRef]
  13. Li, L.; Liu, Z. Noise-robust HRRP target recognition method via sparse-low-rank representation. Electron. Lett. 2017, 53, 1602–1604. [Google Scholar] [CrossRef]
  14. Jiang, Y.; Han, Y.B.; Sheng, W.X. Target recognition of radar HRRP using manifold learning with feature weighting. In Proceedings of the 2016 IEEE International Workshop on Electromagnetics: Applications and Student Innovation Competition, Nanjing, China, 16–18 May 2016; pp. 1–3. [Google Scholar]
  15. Yu, X.L.; Wang, X.G.; Liu, B.Y. Supervised kernel neighborhood preserving projections for radar target recognition. Signal Process. 2008, 88, 2335–2339. [Google Scholar] [CrossRef]
  16. Liu, L.; Zhu, W.; Huang, C. Satellite orbit-motion theories and their applications. I. Prog. Astron. 1988, 6, 42–51. [Google Scholar]
  17. Sidorenko, V.V.; Neishtadt, A.I. Investigation of the Stability of Long-Periodic Planar Motion of a Satellite in a Circular Orbit. Cheminform 2000, 38, 4687–4695. [Google Scholar]
  18. Li, H.J.; Yang, S.H. Using range profiles as features vectors to identify aerospace objects. IEEE Trans. Antennas Propag. 1993, 41, 261–268. [Google Scholar] [CrossRef]
  19. Du, L.; Liu, H.W.; Bao, Z. Using target aspect information to improve radar HRRP recognition performance. J. Syst. Eng. Electron. 2004, 26, 1041–1043. [Google Scholar]
  20. Yuan, L.; Liu, H.W.; Bao, Z. Adaptive learning of classifier parameters for radar high range resolution profiles recognition. J. Electron. Inf. Technol. 2007, 30, 198–202. [Google Scholar] [CrossRef]
  21. Liu, J.; Zhang, J.Y.; Du, L. A frame segmentation method for radar HRRPs based on correlation coefficient. J. Electron. Inf. Technol. 2008, 30, 2060–2064. [Google Scholar] [CrossRef]
  22. Du, L. Research on Target Recognition Method Based on Radar HRRP. Ph.D. Thesis, Xidian University, Xi’an, China, April 2007. [Google Scholar]
  23. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
  24. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  25. Athanasios, V.; Nikolaos, D.; Anastasios, D.; Eftychios, P. Deep Learning for Computer Vision: A Brief Review. Comput. Intell. Neurosci. 2018, 2018, 1–13. [Google Scholar]
  26. Zhang, Z.; Geiger, J.; Pohjalainen, J.; Mousa, A.E.D.; Jin, W.Y.; Schuller, B. Deep Learning for Environmentally Robust Speech Recognition. ACM Trans. Intell. Syst. Technol. 2018, 9, 1–28. [Google Scholar] [CrossRef]
  27. Aceto, G.; Ciuonzo, D.; Montieri, A.; Pescape, A. Mobile encrypted traffic classification using deep learning. In Proceedings of the 2018 Traffic Measurement and Analysis Conference (TMA), Vienna, Austria, 26–29 June 2018; pp. 1–8. [Google Scholar]
  28. Aceto, G.; Ciuonzo, D.; Montieri, A.; Pescape, A. Mobile Encrypted Traffic Classification Using Deep Learning: Experimental Evaluation, Lessons Learned, and Challenges. IEEE Trans. Netw. Serv. Manag. 2019. [Google Scholar] [CrossRef]
  29. Liang, Z.H.; Zhang, G.; Huang, J.X.; Hu, Q.V. Deep learning for healthcare decision making with EMRs. In Proceedings of the 2014 IEEE International Conference on Bioinformatics & Biomedicine (BIBM), Belfast, Northern Ireland, 2–5 November 2014; pp. 556–559. [Google Scholar]
  30. Zhao, F.X.; Liu, Y.X.; Huo, K.; Zhang, S.H.; Zhang, Z.S. Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine. Sensors 2018, 18, 173. [Google Scholar] [CrossRef]
  31. Feng, B.; Chen, B.; Liu, H.W. Radar HRRP target recognition with deep networks. Pattern Recognit. 2017, 61, 379–393. [Google Scholar] [CrossRef]
  32. Lundén, J.; Koivunen, V. Deep learning for HRRP-based target recognition in multistatic radar systems. In Proceedings of the 2016 IEEE Radar Conference, Philadelphia, PA, USA, 2–6 May 2016; pp. 1–6. [Google Scholar]
  33. Xu, B.; Chen, B.; Liu, H.W.; Jin, L. Attention-based Recurrent Neural Network Model for Radar High-resolution Range Profile Target Recognition. J. Electron. Inf. Technol. 2016, 38, 2988–2995. [Google Scholar]
  34. Jithesh, V.; Sagayaraj, M.J.; Srinivasa, K.G. LSTM recurrent neural networks for high resolution range profile based radar target classification. In Proceedings of the IEEE 2017 3rd International Conference on Computational Intelligence & Communication Technology, Ghaziabad, India, 9–10 February 2017; pp. 1–6. [Google Scholar]
  35. Donahue, J.; Hendricks, L.A.; Rohrbach, M.; Venugopalan, S.; Saenko, K.; Darrell, T. Long-Term Recurrent Convolutional Networks for Visual Recognition and Description. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 39, 677–691. [Google Scholar] [CrossRef]
  36. Byeon, W.; Breuel, T.M.; Raue, F.; Liwicki, M. Scene labeling with LSTM recurrent neural networks. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 8–10 June 2015; pp. 3547–3555. [Google Scholar]
  37. Cho, K.; Van Merrienboer, B.; Gulcehre, C.; Bougares, F.; Schwenk, H.; Bahdanau, D.; Bengio, Y. Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation. Comput. Sci. 2014. [Google Scholar] [CrossRef]
  38. Li, S.; Li, W.Q.; Cook, C.; Zhu, C.; Gao, Y.B. Independently Recurrent Neural Network (IndRNN): Building A Longer and Deeper RNN. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 5457–5466. [Google Scholar]
  39. Chung, J.; Gulcehre, C.; Cho, K.H.; Bengio, Y. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  40. Jozefowicz, R.; Zaremba, W.; Sutskever, I. An Empirical Exploration of Recurrent Network Architectures. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015. [Google Scholar]
  41. Du, L.; Liu, H.W.; Bao, Z. Radar HRRP Statistical Recognition: Parametric Model and Model Selection. IEEE Trans. Signal Process. 2008, 56, 1931–1944. [Google Scholar] [CrossRef]
  42. Xing, M.D.; Bao, Z.; Pei, B. Properties of high-resolution range profiles. Opt. Eng. 2002, 41, 493–504. [Google Scholar] [CrossRef]
  43. Li, Y.; Li, C.Y. Global correlation envelope alignment of high precision. In Proceedings of the 2007 Asian and Pacific Conference on Synthetic Aperture Radar, Huangshan, China, 5–9 November 2007; pp. 815–817. [Google Scholar]
  44. UCS Satellite Database. Available online: https://www.ucsusa.org/nuclear-weapons/space-weapons/satellite-database#.XCGbJHaFKx9 (accessed on 25 December 2018).
  45. Bahdanau, D.; Cho, K.; Bengio, Y. Neural machine translation by jointly learning to align and translate. Technical report. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  46. Cortes, C.; Vapnik, V. Support-vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  47. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
Figure 1. Illustration of a HRRP sample from a satellite.
Figure 1. Illustration of a HRRP sample from a satellite.
Sensors 19 02008 g001
Figure 2. HRRP sample examples of satellites. (a) HRRP sample example of satellite No.1; (b) HRRP sample example of satellite No.2.
Figure 2. HRRP sample examples of satellites. (a) HRRP sample example of satellite No.1; (b) HRRP sample example of satellite No.2.
Sensors 19 02008 g002
Figure 3. Analysis and statistical results of observation satellite orbit. (a) Satellite orbital parameters description; (b) Variation of altitude difference with eccentricity when semi-major axis is given; (c) Eccentricity statistical results; (d) Apogee altitude statistical results.
Figure 3. Analysis and statistical results of observation satellite orbit. (a) Satellite orbital parameters description; (b) Variation of altitude difference with eccentricity when semi-major axis is given; (c) Eccentricity statistical results; (d) Apogee altitude statistical results.
Sensors 19 02008 g003
Figure 4. Coordinate transformation diagrams. (a) Radar body coordinate system transforms into geocentric coordinate system; (b) Radar polar coordinate transforms into radar body Cartesian coordinate system.
Figure 4. Coordinate transformation diagrams. (a) Radar body coordinate system transforms into geocentric coordinate system; (b) Radar polar coordinate transforms into radar body Cartesian coordinate system.
Sensors 19 02008 g004
Figure 5. Gated recurrent unit.
Figure 5. Gated recurrent unit.
Sensors 19 02008 g005
Figure 6. The structure of GRU-SVM model.
Figure 6. The structure of GRU-SVM model.
Sensors 19 02008 g006
Figure 7. Overall framework of this satellite target recognition method. Note: Tr_S, Te_S and F_V represent the training samples, testing sample and feather vector, respectively. Their subscripts denote different data partition results, where the first subscript represents the partition results of orbital altitude and the second represents the partition results of hierarchical clustering.
Figure 7. Overall framework of this satellite target recognition method. Note: Tr_S, Te_S and F_V represent the training samples, testing sample and feather vector, respectively. Their subscripts denote different data partition results, where the first subscript represents the partition results of orbital altitude and the second represents the partition results of hierarchical clustering.
Sensors 19 02008 g007
Figure 8. Flow chart of data generation.
Figure 8. Flow chart of data generation.
Sensors 19 02008 g008
Figure 9. Partition results of training radar data. (a)   H Sat 500 km ; (b)   500 <   H Sat 1000 km ; (c)   H Sat > 1000 km .
Figure 9. Partition results of training radar data. (a)   H Sat 500 km ; (b)   500 <   H Sat 1000 km ; (c)   H Sat > 1000 km .
Sensors 19 02008 g009
Figure 10. Training records of 100 epochs based on different training dataset. (a) Training accuracy; (b) Validation accuracy; (c) Categorical crossentropy loss of training set; (d) Categorical crossentropy loss of validation set; (e) MAE loss of training dataset; (f) MAE loss of validation set.
Figure 10. Training records of 100 epochs based on different training dataset. (a) Training accuracy; (b) Validation accuracy; (c) Categorical crossentropy loss of training set; (d) Categorical crossentropy loss of validation set; (e) MAE loss of training dataset; (f) MAE loss of validation set.
Sensors 19 02008 g010aSensors 19 02008 g010b
Figure 11. Confusion matrices of untrained data tested by the following training model. H Sat 500 km : (a) Cluster = No . 1 ; (b)   Cluster = No . 2 ; (c) Cluster = No . 3 . 500 < H Sat 1000 km :(d) Cluster = No . 1 ; (e) Cluster = No . 2 ; (f) Cluster = No . 3 ; (g) Cluster = No . 4 . H Sat > 1000 km : (h) Cluster = No . 1 ; (i) Cluster = No . 2 ; (j) Cluster = No . 3 .
Figure 11. Confusion matrices of untrained data tested by the following training model. H Sat 500 km : (a) Cluster = No . 1 ; (b)   Cluster = No . 2 ; (c) Cluster = No . 3 . 500 < H Sat 1000 km :(d) Cluster = No . 1 ; (e) Cluster = No . 2 ; (f) Cluster = No . 3 ; (g) Cluster = No . 4 . H Sat > 1000 km : (h) Cluster = No . 1 ; (i) Cluster = No . 2 ; (j) Cluster = No . 3 .
Sensors 19 02008 g011
Figure 12. Statistical recognition result of seven methods under different conditions. 1 The first number represents whether to use orbital altitude calculation and the second represents whether to utilize observation angles clustering, for example ‘10′ denotes orbital altitude calculation has been applied but clustering is not applied.
Figure 12. Statistical recognition result of seven methods under different conditions. 1 The first number represents whether to use orbital altitude calculation and the second represents whether to utilize observation angles clustering, for example ‘10′ denotes orbital altitude calculation has been applied but clustering is not applied.
Sensors 19 02008 g012
Figure 13. Dimension reduction distribution of training data ( H Sat > 1000 km , Cluster = No . 3 ) for these seven methods. (a) GRU neural network; (b) CNN; (c) Autoencoder (AE); (d) Denoising autoencoder (DAE); (e) PCA; (f) Dictionary learning (DL); (g) Manifold learning (ML).
Figure 13. Dimension reduction distribution of training data ( H Sat > 1000 km , Cluster = No . 3 ) for these seven methods. (a) GRU neural network; (b) CNN; (c) Autoencoder (AE); (d) Denoising autoencoder (DAE); (e) PCA; (f) Dictionary learning (DL); (g) Manifold learning (ML).
Sensors 19 02008 g013
Table 1. List of the acronyms used in this paper.
Table 1. List of the acronyms used in this paper.
AcronymDefinition
GRUGated Recurrent Unit
HRRPHigh Resolution Range Profile
SVM (LSVM)(Linear) Support Vector Machine
RATRRadar Automatic Target Recognition
ISARInverse Synthetic Aperture Radar
NADDCCNormalized Angular Distance Divided by Correlation Coefficient
PCAPrincipal Component Analysis
CNNConvolutional Neural Network
RNNRecurrent Neural Network
LSTMLong Short-Term Memory
MTRCMigration Throuth Resolution Cells
RCSRadar Cross Section
MAEMean Absolute Error
AEAutoEncoder
DAEDenoising AutoEncoder
Table 2. Parameters of the radar and satellites.
Table 2. Parameters of the radar and satellites.
Radar ParametersCenter Frequency10 GHZ
Bandwidth1 GHZ
Longitude115°
Latitude30.5°
Altitude0
SatelliteApogee Range (km)EccentricityInclination (°)
No.1 390 ~ 425 8 . 1   ×   10 4 51.6
No.2 300 ~ 320 1 . 0   ×   10 7 51.0
No.3 310 ~ 322 1 . 1   ×   10 3 54.5
No.4 721 ~ 730 5 . 0   ×   10 3 57.0
No.5 630 ~ 643 2 . 1   ×   10 4 97.8
No.6 615 ~ 624 2 . 2   ×   10 4 97.9
No.7 670 ~ 682 1 . 1   ×   10 3 98.1
No.8 1045 ~ 1165 1 . 3   ×   10 3 63.4
No.9 1094 ~ 1112 3 . 5   ×   10 3 123.0
No.10 1347 ~ 1355 3 . 2   ×   10 4 58.0
Table 3. Recognition accuracy of this recognition method.
Table 3. Recognition accuracy of this recognition method.
Orbital Altitude (km) Cluster Recognition Rate of Satellites (%)Total Accuracy (%)
No.1No.2No.3No.4No.5No.6No.7No.8No.9No.10
( 0 ,   500 ] No.199.798.999.4/ 1//////99.2
No.299.799.699.5///////
No.397.090.595.7///////
( 500 , 1000 ] No.1///99.099.799.899.2///
No.2///99.499.799.798.8///
No.3///99.699.499.899.2///
No.4///10099.699.798.9///
( 1000 , ) No.1///////99.999.999.9
No.2///////98.599.998.9
No.3///////100100100
1 There is no such satellite in the corresponding cluster.
Table 4. Performance comparison of three RNN-based neural network.
Table 4. Performance comparison of three RNN-based neural network.
MethodOrbital Altitude (km)ClusterRecognition Rate of Satellites (%)Total Accuracy (%)
No.1No.2No.3No.4No.5No.6No.7No.8No.9No.10
GRU-SVM ( 0 ,   500 ] No.299.798.999.4/ 1//////99.6
RNN-SVM70.243.586.8///////66.9
LSTM-SVM96.793.797.7///////96.0
GRU-SVM ( 500 , 1000 ] No.1///99.099.799.899.2///97.8
RNN-SVM///99.182.987.580.7///86.9
LSTM-SVM///99.486.388.279.7/// 87.7
GRU-SVM ( 1000 , ) No.3///////100100100100
RNN-SVM///////97.693.090.093.3
LSTM-SVM///////96.199.493.196.0
GRU-SVM ( 1000 , ) × 2///////99.099.998.799.2
RNN-SVM///////86.793.955.378.7
LSTM-SVM///////97.499.896.697.9
GRU-SVM×No.299.697.492.098.194.695.089.097.899.697.195.8
RNN-SVM88.173.678.688.281.077.561.489.685.778.279.5
LSTM-SVM95.993.983.690.486.980.969.387.598.187.987.0
MethodOrbital Altitude (km)ClusterTraining Data ScaleEpochCPUGPUTraining Time
GRU ( 0 ,   500 ] No.255,440 HRRPs100E5-2630
32 G
NVIDIA
Quadro
P4000
21 h 36 min 16 s
RNN20 h 29 min 36 s
LSTM 43 h 59 min 28 s
GRU ( 500 , 1000 ] No.162,720 HRRPs100E5-2680
256 G
GTX
1080 Ti
26 h 32 min 11 s
RNN16 h 01 min 13 s
LSTM32 h 52 min 36 s
GRU ( 1000 , ) No.336,400 HRRPs100E5-2630
32 G
NVIDIA
Quadro
P4000
33 h 11 min 04 s
RNN19 h 43 min 53 s
LSTM30 h 51 min 39 s
GRU ( 1000 , ) ×84,000 HRRPs100E5-2680
256 G
GTX
1080 Ti
30 h 34 min 00 s
RNN17 h 17 min 48 s
LSTM36 h 08 min 35 s
GRU×No.286,800 HRRPs100E5-2680
256 G
GTX
1080 Ti
37 h 43 min 18 s
RNN22 h 08 min 34 s
LSTM42 h 37 min 34 s
1 There is no such satellite in the corresponding cluster. 2 This method or technology is not applied.
Table 5. The recognition accuracy of seven methods under different conditions.
Table 5. The recognition accuracy of seven methods under different conditions.
MethodOrbital Altitude (km) Cluster Recognition Rate of Satellite (%)Total Accuracy (%)
No.1No.2No.3No.4No.5No.6No.7No.8No.9No.10
GRU-SVM199.197.499.799.799.699.899.199.199.999.699.2
× 293.382.394.599.796.498.694.499.099.997.395.4
×97.282.889.497.789.392.484.893.899.590.791.7
××63.641.073.385.170.889.056.857.490.848.967.6
CNN-SVM98.993.897.499.695.095.488.499.799.197.496.6
×97.388.196.698.890.493.781.293.399.288.993.1
×97.983.683.398.585.992.778.594.398.791.890.4
××76.982.682.078.670.170.670.877.663.059.172.6
AE-SVM86.483.977.186.385.875.970.281.598.388.483.3
×81.085.079.076.083.070.064.072.093.082.077.9
×77.267.752.376.763.858.157.653.185.454.663.9
××69.057.043.064.049.043.040.039.078.043.052.0
DAE-SVM85.380.473.088.175.176.769.878.595.679.080.0
×81.074.072.077.070.069.062.071.092.069.073.6
×73.658.645.477.851.856.353.050.876.553.360.3
××64.046.039.066.039.047.040.038.070.042.049.0
PCA-SVM89.484.577.790.285.979.877.781.997.591.086.7
×83.077.077.076.082.070.066.074.093.082.077.6
×79.164.652.178.064.554.955.450.875.766.566.0
××71.048.038.066.051.044.046.038.082.051.054.0
DL-SVM87.480.075.289.476.476.074.478.596.985.081.2
×82.075.070.077.072.067.060.072.092.075.074.9
×74.453.847.377.355.753.156.551.977.257.460.6
××65.039.034.065.045.042.042.037.073.042.049.0
ML-SVM87.683.479.192.680.180.275.279.797.081.383.6
×83.078.074.082.074.073.067.072.093.073.077.3
×74.362.754.082.756.357.356.352.878.159.263.3
××68.047.043.074.044.047.044.038.073.046.053.0
1 This method or technology is applied. 2 This method or technology is not applied.

Share and Cite

MDPI and ACS Style

Lu, W.; Zhang, Y.; Xu, C.; Lin, C.; Huo, Y. A Deep Learning-Based Satellite Target Recognition Method Using Radar Data. Sensors 2019, 19, 2008. https://doi.org/10.3390/s19092008

AMA Style

Lu W, Zhang Y, Xu C, Lin C, Huo Y. A Deep Learning-Based Satellite Target Recognition Method Using Radar Data. Sensors. 2019; 19(9):2008. https://doi.org/10.3390/s19092008

Chicago/Turabian Style

Lu, Wang, Yasheng Zhang, Can Xu, Caiyong Lin, and Yurong Huo. 2019. "A Deep Learning-Based Satellite Target Recognition Method Using Radar Data" Sensors 19, no. 9: 2008. https://doi.org/10.3390/s19092008

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop