Next Article in Journal
Joint Estimation of SOC and Available Capacity of Power Lithium-Ion Battery
Next Article in Special Issue
Pipeline Leak Detection and Estimation Using Fuzzy PID Observer
Previous Article in Journal
Towards Human Stress and Activity Recognition: A Review and a First Approach Based on Low-Cost Wearables
Previous Article in Special Issue
Benchmarking GHG Emissions Forecasting Models for Global Climate Policy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Neural Networks and Deep Learning Techniques Applied to Radar Target Detection: A Review

School of Computer Science and Technology, University of Chinese Academy of Sciences, Beijing 101400, China
*
Authors to whom correspondence should be addressed.
Electronics 2022, 11(1), 156; https://doi.org/10.3390/electronics11010156
Submission received: 15 November 2021 / Revised: 20 December 2021 / Accepted: 24 December 2021 / Published: 4 January 2022
(This article belongs to the Special Issue Theory and Applications of Fuzzy Systems and Neural Networks)

Abstract

:
Radar target detection (RTD) is a fundamental but important process of the radar system, which is designed to differentiate and measure targets from a complex background. Deep learning methods have gained great attention currently and have turned out to be feasible solutions in radar signal processing. Compared with the conventional RTD methods, deep learning-based methods can extract features automatically and yield more accurate results. Applying deep learning to RTD is considered as a novel concept. In this paper, we review the applications of deep learning in the field of RTD and summarize the possible limitations. This work is timely due to the increasing number of research works published in recent years. We hope that this survey will provide guidelines for future studies and applications of deep learning in RTD and related areas of radar signal processing.

1. Introduction

Radar target detection (RTD) is widely used to determine whether there is a signal present in noise. Since radar signals reflected from targets are often immersed in complex backgrounds (e.g., noise, clutter, even jamming), traditional signal processing methods are often used to boost signal-to-noise ratio (SNR) [1], while constant false alarm rate (CFAR) is a useful method for detection in a noise environment based on hypothesis testing [2]. Traditional CFAR-based detection methods consider the models of target or environment as a stochastic process which is usually based on statistical theory [3]. However, due to the complex detection environment and diverse target types, finding targets in a complex scene is an extremely challenging task, and therefore a reliable and robust RTD method has been one of the key pursuits of research [4].
Deep learning is a rapidly developing technology which has dramatically brought a breakthrough in many fields such as image classification, natural language processing, speech recognition, etc. [5,6]. As a subset of machine learning, deep learning-based models attempt to extract features from large scale raw data automatically. The success of deep learning is mainly due to the availability of big data, the improvement of computational power, and the ability of data processing [7]. Various deep neural network (DNN) technologies have been successfully used, including deep neural networks (DNN), convolutional neural networks (CNN), recursive neural networks (RNN), deep belief networks (DBN), etc. [8].
Although deep learning technology has demonstrated an exciting trend over the past few years, its full potential for radar application has not yet been explored. In ref. [9], researchers grouped the radar application problems that can be solved by deep learning-based methods into three general categories: radar sensing, radar signal processing, and radar automatic target recognition (ATR), respectively, which are listed in Figure 1. Radar sensing and radar signal processing are the necessary prerequisites and procedures for radar ATR.
Deep learning methods have been fully applied in radar sensing. Wang et al. [10] applied CNNs to radar waveform recognition. According to [11,12], deep learning methods can solve the problem of high computational complexity of antenna parameter optimization. The design of array antennas has been addressed by artificial neural networks (ANN) [13]. Cognitive radar antenna selection has also been solved by deep learning methods [14,15]. Gao et al. [16] explored a feature extractor based on CNN and a stacked autoencoder (SAE) to recognize modulated signals of radar, including LFM, NLFM, BPSK, FRANK, COSTAS, P1, P2, P3 and P4, etc.
Automatic Target Recognition is the most straightforward and widely-used application of deep learning to radar, especially for synthetic aperture radar (SAR) [17,18,19,20]. Different “image-like” radar data have been well-studied for the purpose of feature extraction and classification of ATR. Deep learning models were built up on radar high-resolution range profiles (HRRP) for ATR [21,22,23]. ATR on radar Micro-Doppler signatures could also be considered in [24,25,26,27,28]. Deep learning can also be used for recognition of human activities using Micro-Doppler signatures, including hand gestures recognition [29,30,31], vital signs sensing [32,33,34], routine activities recognition [35,36,37,38,39], etc. Khalid et al. [40] investigated the use of radar Range-Doppler profiles for ATR with the Convolutional Long Short-Term Memory (CLSTM) model. A deep learning-based method on Range-Doppler radars is proposed to recognize human fall motions in [41].
Radar Signal Processing is an intermediate procedure between radar sensing and ATR. One of the most important purposes of radar signal processing is to detect targets. Machine learning-based classifiers algorithms, such as support vector machines (SVM) [42] and k-Nearest-Neighbour (kNN) [43], have been used for RTD [44,45]. As previously mentioned, RTD is a binary hypothesis testing which can be regarded as a binary classification problem, namely whether the target is present or absent. Based on this assumption, RTD can be regarded as an ATR application. As deep learning models have achieved good performance in ATR, it is feasible and reasonable to explore deep learning-based models in RTD. As one of the most reliable classifiers, ANN has been utilized to improve radar detection performance [46,47]. Many recent literatures utilize DNN to tackle RTD and also present performance improvements. Figure 2 illustrates the major developments and application of deep learning in RTD. Detailed descriptions of each algorithm will be presented later.
Applying deep learning technology to the field of RTD is a novel concept, yet by now, there is no paper which comprehensively summarizes and introduces the application and its development status. In this paper, we try to review the applications of ANN and deep learning methods in the classical radar system problem of target detection.
This review focuses on articles in online databases, e.g., IEEExplore, Open Science Elsevier, Scopus, Springer and Researchgate. Recent articles (available by July 2021) published in major journals of radar signal processing and major international conferences on artificial intelligence attract more attention; these include IEEE Signal Processing Magazine, IEEE Transactions on Antennas and Propagation, IEEE Transactions on Aerospace and Electronic Systems, IEEE Transactions on Geoscience and Remote Sensing, IEEE Geoscience and Remote Sensing Letters, IEEE Journal of Selected Topics in Signal Processing, IEEE Transactions on Signal Processing, ISPRS Journal of Photogrammetry and Remote Sensing, IEEE Radar Conference, International Conference on Radar, IEEE Conference on Computer Vision and Pattern Recognition, and IEEE International Conference on Computer Vision. In addition, a number of research papers from other sources are related to this topic and thus included in this review, and most of them were published between 2000 and 2020. These papers were selected by keywords in the titles and keywords of the studies, and they represent a wide range of: (a) methods from theoretical derivation to application research, (b) detection background from noise to clutter, (c) applications from maritime target detection to human motion detection, (d) data forms from radar received echoes to PPI images, and (e) comparative methods from neural networks to deep learning models. Lastly, only studies in the English language are included in the review.
The rest of this review is organized as follows: Section 2 introduces related work on RTD including traditional CFAR detectors, which are often used for performance comparisons. For completeness, in Section 3, we first recap on the basic methodology of ANN before applying it to practical applications, and we then recap the recent deep learning-based methods for. In Section 4, some open datasets are described as well as how to construct synthetic datasets, while Section 5 summarizes the research challenges and opportunities. Finally, conclusions are presented in Section 6.

2. Related Work on Radar Target Detection

2.1. Traditional Processing Methods for RTD

In a complex scenario of RTD, radar echo signals are often immersed in noise, jamming and clutter, etc., which is shown in Figure 3. During the typical pulse Doppler radar signal processing, the reflected radar signals are processed by a series of methods such as matched filtering, coherent accumulation, clutter suppression, CFAR detection, etc.
The received radar signals are sampled at a certain rate, and the pulses are compressed by the matched filter to obtain high range resolution and narrow pulse width. Moving target indication (MTI) is performed to suppress static clutter. Then, Doppler processing or coherent accumulation is applied over multiple pulses at each range unit to obtain the Range-Doppler spectrum. The reflected signal amplitude for each Range-Doppler is stored in separate cells, after which the CFAR detector is utilized to reveal one whose amplitude exceeds the threshold of detection. Thus, the corresponding information about velocity and position can be measured. The flowchart is presented in Figure 4.
One of the most important purposes of designing a radar detector is to distinguish targets from noise, clutter and jamming signals. A decision must be made at the end of the detector as to whether the radar echo contains the target or not. The classical method is to establish an adaptive detection threshold based on statistical models, which varies according to noise and clutter energy. In order to minimize the false alarm rate ( P f a ) and maximize the probability of detection ( P D ), the Neyman-Pearson criterion is utilized for decision-making [2]. The typical performance requirement of a radar system will require P D 0.8 and P f a 10 e 4 . This problem is commonly solved by applying a CFAR detector, which adaptively determines a local optimum threshold and maintains the P f a to be constant as a predetermined value [48].
Figure 5 depicts a general CFAR detector, which is described as a shift register of length 2 n + 1 . The input samples are sent into the detector cell by cell and the energy y in the cell under test (CUT) is estimated. The CFAR detector adjusts the statistic value z according to the variation of energy in 2 n reference cells. The energy of the CUT is compared to the statistical result z of the CFAR processor scaled by a constant scale factor α . Thus, the detection threshold is represented as α z . Determining whether there is a target according to (1):
{ y α z , t a r g e t   a b s e n t y > α z , t a r g e t   p r e s e n t
In order to find an adaptive detection threshold, many CFAR-based methods have been studied, which can adapt the threshold to the background changes, keeping a constant P f a . The diagram of typical CFAR processors is also showed in Figure 5 [49].
Compared with one-dimensional range detection, two-dimensional detection under phase-coherent accumulation includes range and Doppler detection, which is presented in Figure 6. x denotes the reference cells of range dimension, v denotes the reference cells of the Doppler dimension. The detection process of Doppler dimension is the same as that of range dimension.
CFAR-based detectors have been well-studied over the past years, however, there still remains a trade-off between the various CFAR technologies. The cell averaging CFAR (CA-CFAR) is the most widely-used method which has the highest detectability in homogeneous background [50]. However, it also exhibits severe performance degradation in the presence of an interfering target or an abrupt change in clutter background [51]. The greatest CFAR (GO-CFAR) and the smallest CFAR (SO-CFAR) detectors were developed to improve the detection performance under various non-ideal conditions such as nonhomogeneous clutter background and multiple target environments [52]. Rohling et al. [53] developed an ordered statistics CFAR (OS-CFAR) to control the detecting threshold even though other interfering signals occurred in reference cell. As far as we know, CFAR methods are still under investigation, such as the censored mean level detector CFAR (CMLD-CFAR) and trimmed mean CFAR (TM-CFAR) detector [54], which were introduced to improve the anti-inference performance. However, each of those CFAR algorithms is only limited to some specific cases, lacking in generality.

2.2. Deficiencies and Challenges in Conventional Approaches

RTD is a rather complex problem in practical applications. Although conventional methods have been working well in some conditions, deficiencies and challenges still exist. Currently, the difficult problems of RTD still mainly lie in high-resolution processing of targets, clutter suppression, anti-jamming technology, and ‘low-small-slow’ target detection (low glancing angle, small size, slow or stationary), etc. [4].
Conventional CFAR-based methods in radar systems mainly depend on statistical hypothesis. (1) In the actual RTD, only the specific type of target under the specific background has good detection performance, because the predefined parameters of the detector, such as margin, threshold, sizes of the guard and reference windows, will determine the detection accuracy. However, radar always works in a diverse scene. (2) Furthermore, traditional methods are computationally expensive and not flexible because they process inputs cell-by-cell and need to manually change the window size to adaptive targets of different resolutions. (3) Most importantly, in most cases, neither the target nor the environment (noise, clutter, interference) have known statistical models. It is difficult to find suitable parameters to design the radar detector, not to mention predict its performance accurately.
In brief, traditional statistical methods are no longer applicable to complex scenes and the selection of the optimal parameter set is extremely challenging. It seems reasonable and inevitable to develop a data-driven deep learning approach for RTD.

3. Deep Learning Methods for Radar Target Detection

3.1. Artificial Neural Networks and Deep Learning-Based Models for RTD

ANN is motivated by the biological structure of the human brain. Generally, a neural network consists of neuron, weight vector, bias, activation function, etc. Multi-Layer Perceptron (MLP) is a type of neural network which consists of several hidden layers with neurons in layers being interconnected to each other [55]. Neural networks can be utilized in RTD due to their learning ability. In fact, the problem of RTD can be considered as a problem of pattern recognition, which fits well to the possibilities that an ANN provides [56]. Several approaches have suggested that considering ANNs as non-linear detectors could improve the detection performance. A typical ANN detector is shown in Figure 7. The output of each neuron can be given as (2):
X k + 1 = f ( W k X k + b k )
where X k is the input vector of the k t h layer, W k is the weight vector of the k t h layer, b k is an element of bias vector of the k t h layer, f ( ) is the activation function. ANN detector produces y 1 if the CUT contains a value larger than the sum of all the reference cells scaled by weights, otherwise, it would output y 2 . Usually, a more complex ANN detector with multiple hidden layers would give better performance.
Using multiple layers to achieve more powerful generalization and abstractions, DNN outperforms all the machine learning methods. The traditional shallow neural network requires an empirical feature extraction process to decrease the networks’ computational load. If the features have to be extracted automatically by the network, a deep network with powerful training capability is necessary. Currently, development in computer power and memory have made it possible to train very large-scaled networks. Various sizes of layers are used to provide different degrees of abstraction and generalization [57].
The increased popularity of deep learning has brought an increase in research publications related to RTD with deep learning models over the past few years. Specially designed network frameworks for RTD in noise environments and clutter backgrounds will be included in this work; in addition, several different DNNs will be reviewed in this paper, and each of them perform well for RTD. The increase in the number of related publications confirms the valid and increasing motivation of the research community on RTD tasks.
Next, we will review the recent articles using deep learning that address RTD in online databases. Table 1 summarizes the related works on RTD where we present the type of tasks and main contributions of each study. A more detailed description of the detection methods is provided in Table 2, where we summarize all the related papers, particularly highlighting the architecture of networks and the type of input and dataset. Unfortunately, a direct comparison for all methods is not possible because they were evaluated on different datasets.

3.2. RTD in Noise Background

In the past decades, many approaches have been proposed to address the issue of RTD in diverse types of noise or clutter scenarios. Those approaches include complete descriptions of environmental statistics as well as statistical computing power. Recently, deep learning-based schemes were proposed to cope with the problem of RTD within noise backgrounds.
In practical application, the likelihood ratio can be obtained by sufficient statistics, which mainly depends on the probability density function of the noise, for example, Gaussian white noise. However, in a modern radar system, noise is usually non-Gaussian distribution, thus the likelihood ratio has a complex non-linear relation which makes it difficult to implement sufficient statistics [83,84]. Early in 1988, Gandhi and Kassam [85] analyzed the theoretical principle of CFAR processors in non-homogeneous backgrounds. This kind of processing can also be handled by ANN due to its ability to realize complicated nonlinear mappings on the data directly. In 1997, Gandhi and Ramaurti [58] were likely the first to employ neural networks to detect signals in non-Gaussian noise at some specified P f a . With a detailed theoretical derivation, it was noted that the ANN detector’s performance did not rely on the SNR, but actual relied on signal strength and noise common variations during training. By setting several non-Gaussian noise environments, the performance of the ANN detector has been shown to outperform the matched filter as well as the locally optimum detectors under some certain non-Gaussian noise environments. However, the computational power and the storage requirements are generally higher in the ANN detector.
In order to improve the detection performance of radar a in non-homogeneous environment, in 2015, Rohman et al. [59] presented a novel adaptive switch between the CA-CFAR and OS-CFAR detector by using the ANN structure with MLP which consisted of 2 hidden layers. The proposed architecture is illustrated in Figure 8. The inputs of ANN were calculated thresholds and the CUT value and the output was a preliminary threshold. Then, the nearest value between raw threshold and CA or OS-CFAR would be selected and utilized as the final threshold. The experiment’s results showed that the combined approach is capable of switching between CA-CFAR and OS-CFAR properly in homogenous and non-homogeneous environments based on the best detection performance.
Additionally, a competent radar system must provide a high P D with a low P f a , which is the major principle of employing standard CFAR detectors. Akhtar and Olsen [60] presented an ANN, which is trained under a CA-CFAR detector for a fluctuating target detection procedure with a noise background. The ANN detector would output positive outcomes conditionally if a real target exists at a CUT and CA-CFAR returns a positive detection or the network would not return positive results. A prominent benefit of the ANN detector is that the outcome may be regarded as a measure of P D , and not necessarily be either 0 or 1. It was also shown that such a scheme can obtain a slightly lower, or comparable target detection performance, but with a noticeably lower P f a than that of the traditional CA-CFAR detector.
Taking into consideration that ANNs are confirmed to be able to approximate the CFAR detector, Amores et al. [86] further argued that the ANN can improve the robustness of a radar detector. Results show that although the detection performance of the trained network tends to increase as the number of hidden neurons increases, MLPs with one hidden layer with 23 units can implement very robust detectors for SNR lower than 10 dB. For more than 23 hidden neurons, the performance improvement is trivial, while the associated computational cost continues growing.

3.3. RTD in Clutter Background

Compared with a noise background, target detection in clutter background is a more common but challenging problem, because the signal returned from the targets are severely immersed by the backscatter from the clutter. According to theoretical analysis, the detection performance of CFAR detection increases as the number of reference cells increases; as the number approaches infinity, the CFAR detector approaches the optimal detector [87]. However, the serious degradation in P D of the traditional CFAR detectors is due to a reduction in the number of available reference units. This decrease may be caused by high-resolution, the presence of interference signals, or clutter patches.
In the CFAR scheme, the need for a larger reference window, or more reference cells, results from the statistical requirements for the parameters which are used for representing the clutter background and the target fluctuation. The information loss caused by the window size reduction can be compensated for by introducing some extra parameters. Early in 1994, Amoozegar et al. [61] presented a NN-based CFAR detection method that provides a robust performance to compensate the loss of reference cells. The input layer contains nine specific statistical parameters which represent the features of target and clutter. A multi-layer feedforward neural network with sigmoid activation function was proposed to extract the features of target and clutter fluctuations. The results of experiments under diverse scenarios indicated that the NN-CFAR scheme consistently provides a superior robustness in responding to new environments and outperforms CA-CFAR detectors with small-sized reference windows.
To distinguish target from clutter, another meaningful method is to look for certain intrinsic features from the returned echoes to describe the difference between the targets and the clutter. Callaghan et al. [63] selected the magnitude of each Range-Doppler pixel as the feature from Range-Doppler maps and processed them by SVM and KNN algorithms to discriminate small maritime targets from sea clutter. Similarly, Li et al. [64] explored and extracted three practically discriminative features, namely the frequency peak to average ratio, the temporal Hurst exponent, and the temporal information entropy, from radar echo in time and frequency domains to construct feature space. An SVM-based detector which can flexibly control the P f a was designed for target detection within sea clutter. Experimental results showed that the detection probability of the proposed detector is obviously higher than that of the classical detectors under the condition of low signal-to-clutter ratio (SCR) and low P f a cases. All these statistical parameters which can depict the target and clutter characteristics may be embedded as the input of the NN-based detector in the literature [61].
However, because the characteristics of clutter highly rely on the actual environment and the parameters of radar, the above extracted parameters or features often become ineffective when the detection environment changes. Compared with machine learning methods which need to empirically select features, ANN is more suitable for extracting high-dimensional features, and has been adopted as a method of radar signal detection [88]. Following on from this work, Cheikh and Soltani [62] used different ANN architectures to assess the problem of RTD in a K-distributed clutter with thermal noise. They considered the MLP architecture with the genetic algorithm and back-propagation as the training methods, and the radial basis function (RBF) which has been largely used in signal processing was also adopted in the neuronal detector. A training set was used which can describe the clutter distribution well, and the architecture of the ANN-CFAR detector is depicted in Figure 9. The results show that the ANN-CFAR detector with MLP structure has better performance compared to the classical OS-CFAR and CA-CFAR detectors.
Previous work, such as [58,61,62], attempted to improve the target detection performance on the basis of traditional CFAR methods. One objective in [59,60,69] is to replace the CFAR-based detector completely with neural networks to optimize the process of detection. In 2019, Akhtar and Olsen [65] continued to propose a more general training strategy where the conventional GO-CFAR detectors mutually work to detect targets in K-distributed clutter. This process would also be transferred into an ANN with four hidden layers and 19 nodes in each layer, with tanh as the activity function. The training strategy is related to [60] and the ANN structure is similar to that of [62], which is shown in Figure 9. The complete training data includes 2000 independent range-doppler maps with the average SNR over all CPIs ranging from −40 dB to 75 dB, while the average SCR varied from −60 dB to 60 dB. The experimental results show that, at least for specifically trained scenes, the overall detection performance of the ANN can significantly outperform a GO-CFAR detector, namely augment on P D with reduction in P f a .
From the above work, these ANN-based methods distinguish targets from a noise or clutter background and deliver better and more robust results than conventional statistical approaches. In addition, the multilayer architecture shows better performance, particularly in a mixed clutter environment. However, classical ANNs also have a major limitation; that is, due to the small number of neurons and layers, satisfactory performance may not be achieved when dealing with classification and regression problems.

3.4. Deep Learning for RTD with Different Data Forms

DNNs, such as CNN, RNN, etc., operate in a similar way to ANNs, but with more hidden layers and neurons. DNNs are capable of learning complex relationships from different types of data, which are more suitable to deal with large datasets and complex training algorithms. A basic CNN architecture consists of an input layer, a convolution layer, a pooling layer, a fully connected layer, and an output layer. The convolutional operation could generate local feature maps, and the pooling operation plays a role in obtaining the translation invariant [9]. The connections between these layers are more sparse than classical ANN structures.
CNNs are favored for computer vision tasks due to the grid structure of the input data, and the sparseness and locality of interactions between layers [89]. Typical optical image target detection algorithms based on CNNs are grouped into two categories: regional proposal-based and regression-based. The former contains R-CNN [90], SPP-Net [91], Fast RCNN [92] and Faster R-CNN [93], the latter contains YOLO [94], SSD [95], etc. Recently, these CNN-based methods have been widely used in RTD, which present faster speed, higher detection and position accuracy compared with CFAR.
Radar echo is a one-dimensional discrete time sequence, but a radar input data usually includes data of multiple range cells for a certain time, which determines that there are diverse forms of data for the detection network. The accuracy of deep learning also scales with the data. To accomplish this goal, the received radar signals are reshaped as various images to fit the CNN input format. Therefore, in addition to the original radar received signals, most researchers tried to detect and measure targets hidden in the noise background and clutter for multi-dimensional information with multiple CNN models based on range-Doppler spectrums, pulse-range maps, time-frequency images, synthetic aperture radar (SAR) images and plane position indicator (PPI) images as the inputs respectively, listed in Figure 10.

3.4.1. Radar Received Echo

Many studies have concluded that traditional signal processing methods of radar, serving as preprocessing methods of training data, is beneficial to extract features and improve the performance of detection. But actually, as the conventional processing methods, matched filtering and coherent accumulation are essentially convolution computations, and DNN models can extract features from inputs automatically, so it is feasible to detect targets from original radar echoes with DNNs. Jiang et al. [49] proposed a model for RTD based on CNN, which works with radar echo signal directly and therefore avoids the process of conventional signal processing. The proposed model explores the time and frequency domain of radar echo, which is presented in Figure 11. The echo signal is a one-dimensional discrete complex sequence, so the input data should be constructed as a radar echo cube to fit the network. The main goal of RTD is not only to distinguish the target from noise, but also to predict position and velocity. The RD-Detection Net is used to measure range and velocity, while azimuth and elevation are predicted by the Angle-Detection Net. The CNN-based model presented better accuracy and performance of detection than traditional approaches.
As in previous analyses, it is feasible and reasonable to implement a complete “end-to-end” multi-task RTD learning scheme, and a DNN-based learning scheme can also be utilized for more complex RTD tasks. For specific or more complex tasks, the received radar signals are transformed from the raw echo data to diverse data forms using different transform methods.

3.4.2. Range-Doppler Spectrum

Following the idea of image processing with CNN, a range-Doppler spectrum can be considered as an “image”, then, the detector classifies the “image” as target absent or present, thus, the detection task could be treated as a classification task. Based on this idea, Wang et al. [66] designed a CNN target detector based on the range-Doppler spectrum and compared the proposed method with traditional CFAR detectors. The proposed network architecture presented in Figure 12 is an 8-layer CNN, including 2 convolutional layers, 2 ReLU layers, 2 max-pooling layers, and 2 fully connected layers. Different range-Doppler spectrum with multiple SNR values are constructed as the input of the CNN detector. The detector is actually a sliding window detector that slides over the Range-Doppler spectrum with a fixed window, and further decides whether the spectrum contains targets or just noise.
A deep learning method for automotive radar detection with non-image-like Range-Doppler data is proposed in [67], Brodeski et al. described a CNN model to detect and localize targets in the Range-Doppler-Azimuth-Elevation space. The training data was collected during the calibration process and augmented with raw radar data. Radar signals are transformed to the Range-Doppler domain by Doppler processing as the input to the network. When comparing with the conventional CA-CFAR method, the proposed approach outperformed the classical detection method while keeping the real-time abilities. Their method for data construction was inspired by a simulation-based method for synthetic automotive scenes generation in [68].

3.4.3. Pulse-Range Maps

Radar echoes also can be processed as pulse-range two-dimensional images for training and testing. Therefore, the problem of RTD can also be transformed into the target detection and location problem in Pulse-Range images. Pan et al. [69] introduced a CNN model for small marine target detection in strong sea clutter. A modified Faster R-CNN is utilized to extract the features of small targets and sea clutter, then the extracted features are utilized to detect and position the target in the pulse-range image. Figure 13 presents the DNN architecture for RTD based on Faster R-CNN with pulse-range images. This method proved to be easier to use in locating small targets from sea clutter and is able to obtain high P D obviously, which overcomes the weakness of the traditional CFAR detection methods.
We can find that classic object detection methods based on CNNs, especially Faster R-CNN and SSD, which have been widely employed to RTD. Girshick et al. [92] transformed the detection task into a classification task and proposed Fast R-CNN structure. Ren et al. [93] depicted a new region proposal network (RPN) to implement the end-to-end target detection based on Faster R-CNN and RPN with shared convolutional features [96]. The target detection process of Fast R-CNN is performed in two steps; RPN proposes regions where are easy to find the target. Each region uses anchors to provide its possible course position, and then classifies the proposed target and fine-tunes its position by regression. The structure of Faster R-CNN is presented in Figure 14, and this architecture has been widely used as the basic framework for RTD, not only in Pulse-Range images, but also in SAR images and PPI images.

3.4.4. SAR Images

Some radar image signals, such as SAR data, can be inputted as images. Target detection based on SAR images is a key step in ATR, because SAR can provide high-resolution radar images of a wide range of scenarios including all-weather and all-day. However, as mentioned earlier, SAR imaging technology is another application branch of radar signal processing, and SAR images are widely-used in ATR. RTD tasks utilizing SAR images, as the input in this review will mainly focus on target feature extraction, target detection and location.
CNN is a common deep learning architecture for target detection of SAR images. Wang et al. [74] presented a method and ideas of CNN in the research of target detection in SAR images, laying a foundation for future research. Yang et al. [75] adopted a DNN regression method for SAR images target detection based on the improved structure of YOLO [94]. This network is effective for extracting features with low resolution and complex composition. Zheng et al. [76] proposed a multi-feature target detection method in SAR imagers aiming at obtaining the target’s actual position, which is presented in Figure 15. The CNN model was applied to capture deep features while the ANN was adopted to analyze hand-crafted features. Then, two sub-channel features are concatenated together in the main channel and the experimental results showed that the multi-features-based method outperformed other methods.
In addition, Faster R-CNN [93] was modified to improve detection performance in SAR images. Kang et al. [70] introduced a CNN-based detection method combining features and pixels in which the Faster R-CNN framework was modified by the traditional CFAR detection algorithm for small-sized targets detection, which is presented in Figure 16. For targets with clearer shape and structure, Faster R-CNN gave higher classification scores, while for those small-sized targets with smaller bounding boxes, the classification score was relatively low. CFAR was employed because this detector relies on the amplitude of the pixel rather than the shape or structure of the target. The combination of detectors based on deep features and pixels can improve the multi-scale target’s detection performance. An SAR image is different from an optical image in that it reflects the electromagnetic characteristics of the target. Making full use of electromagnetic characteristics in feature extraction for RTD would help improve detection performance. Zhang et al. [77] fused electromagnetic and geometrical characteristics and involved the fused features in Faster R-CNN, and the architecture is presented in Figure 17. in their work, convolutional layers were utilized to extract geometric features of SAR images, just like the process of optical images. Meanwhile, SVM was applied to extract electromagnetic characteristics from the complex data.
SSD is a single DNN designed to detect targets in images, which consists of the basic network and the auxiliary structure [95]. Zhao et al. [78] presented a SSD-based approach to deploy a maritime target detection network model of SAR on an embedded device. As shown in Figure 18, a truncated VGG-16 [97] is adopted as the backbone. The auxiliary structure generates the following key features for detection: convolution predictors, multi-scale feature maps, aspect ratios, and default boxes. Experiments based on the Gaofen-3 spaceborne SAR dataset showed that this approach has practicability and expansibility. Ma et al. [98] proposed a modified SSD model and designed a complete workflow for different targets detection in large-scale GF-3 SAR images.

3.4.5. PPI Images

Although there are different visualizations of radar displays, such as P-display and A-display, they are all processed through radar signal processing and have a the similar process. Radar PPI images indicate all or part of the data of range, azimuth elevation or height. Mou et al. [71] proposed an improved Faster R-CNN method for marine target detection using radar PPI images. VGG16 and ResNet101 were used as backbone network models to extract target features. They modified Faster R-CNN in 4 aspects: (1) using the focal loss [99] instead of classification loss to overcome the deficiency of sample imbalance. (2) ROI Pooling is replaced by precise ROI Pooling [100] to reduce the precision loss in the process of scale unification and enhance the accuracy of pooling. (3) NMS is replaced by soft-NMS [101] to boost missing detection. (4) ReLU is replaced by ELUs [102] to speed up convergence and avoid gradient disappearance. Experimental results proved that compared with the traditional Faster R-CNN, the modified approach based on Faster R-CNN has better detection performance in stability and accuracy.

3.4.6. Time-Frequency Images

The development of micro-Doppler technology also provides a valid method for target detection [103,104]. Targets have the characteristics of micro-motion; due to the micro-motion of the targets, the amplitude and the phase of electromagnetic wave scattered by the moving parts change periodically or irregularly. This is, other words, the micro-motion signature induced by the high-speed moving parts of the targets, known as micro-Doppler.
The extraction of target micro-motion features is always one of the challenges in RTD. A common method for signal time-frequency analysis is Wigner-Ville distribution (WVD). Risueno et al. [72] introduced a WVD-CNN detector for RTD using less free weights than the conventional MLP scheme. Su et al. [73] investigated CNN-based methods (LeNet [105] and GoogLeNet [106] models) for maritime targets detection under different polarization and sea states. In this work, short-time Fourier transform (STFT) is adopted to convert the radar signal (IPIX measured data) to two-dimensional time-frequency images of the target and the clutter. According to the experiment results, LeNet is more efficient in echo signal preprocessing, while GoogLeNet has better detection performance under different polarization and sea states modes in P D and P f a .

3.5. Summarization of Different Structures for RTD

ANN-based models could be used as a kind of detectors to differentiate targets from noise or clutter with high detection performance, especially in a time-varying and non-homogeneous environment, the parameters of the background will change. The robustness of the detector is improved by training the network for different scenarios corresponding to different noise and clutter distribution parameters. Therefore, ANN-based detector is more robust and could provide a higher detection performance than the classical CFAR detector.
DNN-based models could be designed for different RTD tasks with different input forms. The purpose of RTD is to identify whether the radar echo undertested contains a signal from the target or only contains the noise, but more importantly, to obtain multi-dimensional information of position and motion. DNNs turned out to be qualified and excellent in RTD, and more complex architecture with different training ways for more complicated RTD scenes. In addition, various preprocessing methods of radar signal processing could extract features effectively, which could help to improve detection performance. A fusion of the traditional detection methods and the new concepts of deep learning methods has become a promising trend and solution in RTD application.

4. Summary of Datasets and Performance Evaluation

The effectiveness of deep learning-based methods depends largely on the available quantity of training data. The availability of a labelled dataset is regarded as a prerequisite for applying deep learning methods to a certain application. Although several publicly datasets are available for image processing, speech recognition, nature language processing, etc., there are very few ones for radar (except SAR images), not to mention the sampled radar data for RTD based on deep learning. In this review, we summarize IPIX and CSIR experimental datasets which have been used for RTD by some researchers. Although these two measured datasets have been used for RTD, they are not generally public because they are domain-specific. Therefore, we collect and summarize the related radar experimental parameters of these two datasets, which could help you to to understand the existing datasets or construct simulation datasets.

4.1. Dataset Descriptions

4.1.1. IPIX Database

The IPIX database is a widely-used database for sea-surface small target detection, sea clutter characteristic analysis and modeling. It was collected and maintained by Prof. Haykin’s research group at McMaster University. The IPIX database contains two data sets, one was collected by the popular intelligent pixel (IPIX) processing radar under the staring mode in the city of Dartmouth, Canada in 1993 [64,107], and the other came from IPIX radar in the Grimsby area of Canada in 1998 [108]. The Dartmouth database in 1993 contains 339 data sets which cover a wide range of wave and wind conditions. About 14 data files are particularly useful to test algorithms aiming at detecting small objects in sea clutter. All these target data files have a weak target in one of their range bins and are more than 2 min long, which (a subset) is available in [109]. The Grimsby experiment upgraded the quantization bits and measured sea clutter data from different range resolutions; in this way, weak clutter signals and strong targets can be observed simultaneously without large quantization errors or clipping. About 222 datasets in the Grimsby database focus specifically on the floating targets of various sizes, and the actual Grimsby data files are available in [110]. However, the related target information and auxiliary sea state information have yet to be released [111]. The parameters of IPIX radar, experimental parameters and sea state information are summarized in Table 3.
There are limitations to using the IPIX database. These two datasets only cover restricted environmental conditions and limited relative positions, and related important auxiliary information was not recorded detailed, especially the measured data from 1998 [111]. The other limitation is that the radar echoes are only reflected from motionless floating objects, not mobile ships, so it is inevitable for researchers to add extra simulated target echo data. Li et al. [64] adopted the IPIX database to extract features for surface small target detection. Chen et al. [112] validated the proposed micromotion target detection method with a simulated IPIX dataset. Su et al. [72] used IPIX measured target signal and sea clutter data to conduct CNN training and testing schemes for maritime target detection. What is worth mentioning is that the results obtained in [72] are significantly worse than the simulated data. Besides the diversity of the measurements themselves, the low speed and acceleration of the targets in IPIX radar can also lead to degradation.

4.1.2. CSIR Database

The CSIR dataset was collected from two kinds of sea clutter and ship target echo measurement trials which were conducted by the Defense, Peace, Safety, and Security Unit of Council for Scientific and Industrial Research (CSIR) in the southwest coastline of South Africa. The first trial was conducted at the Overberg Test Range (OTR) near Arniston in July 2006 with the Fynment radar [113,114]. The second measurement trial was conducted on 4 November 2007 with an experimental monopulse radar deployed on top of Signal Hill in Cape Town [115,116]. Radar parameters and experimental parameters are demonstrated in Table 4.
The CSIR dataset contains a large amount of sea clutter and target echo data, covering multiple parameter combination (various transmitted waveforms, azimuths and distances) under different environment conditions. The radar operating parameters, marine environment parameters, GPS auxiliary data and the types of cooperative target (inflatable boat, motor yacht and fishing boat) could make up the limitations of the IPIX database. A detailed trial designing scheme and trial data records are also maintained, which could provide reference and guidance for similar trials [108]. Currently, many research institutes are using the CSIR dataset in related research. In Pan’s experiment [69], the dataset TFC15_008, collected from the Fynment radar was used to demonstrate that the DNN approach can easily detect small targets from sea clutter with low SCR and accurately locate the position of the targets compared with the traditional CFAR detection methods. Chen et al. [112] employed the CSIR database to validate the detection performance of marine maneuvering targets with translational motion.
In fact, from Table 2 in Section 2, we can conclude that most research teams construct and label radar datasets by simulation methods for deep RTD, besides SAR images, as the actual radar data was hard to obtain. The simulation parameters of radar waveform can also be referred from Table 3 and Table 4.

4.2. Data Preprocessing and Construction

To boost target detection performance, appropriate preprocessing is necessary. According to the summarization of all related papers, radar received echoes, Pulse-Range maps, Range-Doppler spectrum and Time-Frequency images are widely-used in RTD.
All of them are transformed from raw radar echo by a train of signal processing methods or mathematical operations. In the signal processing process of a pulsed coherent radar, the received signals are processed by pulse compression, coherent accumulation, and clutter suppression. Therefore, to make a better comparison, we will discuss these inputs that are related to deep learning regarding RTD from the literature. Figure 19 shows the flow of the preprocessing of datasets for RTD.

4.2.1. Radar Received Echoes and Radar Cube

A typical pulsed radar system is considered in which a transmitted waveform S ( t ) is emitted at a certain interval. The received echoes S r ( t , k ) , which includes signals of a target and noise, are modulated by amplitudes, time delays, doppler frequency and then sampled at a given rate as (3):
S r ( t , k ) = n σ n S ( t τ n ) e j φ n , k + W ( t )
where t = 1 , 2 , , R (fast time). In the incoming radar echoes, σ n denotes the reflectivity amplitude which can be calculated from the radar function, and τ n is the time delay of target n , W ( t ) is white Gaussian noise, an independent complex Gaussian random variable with mean zero. e j φ n , k denotes the doppler shift for each target, for a target with a constant velocity, it can be defined as (4):
φ n , k = φ n , k 1 + 2 π f c 2 ν n c T r
where k = 1 , , M (slow time), M is the total number of pulses in a coherent processing interval (CPI), assuming φ n , 0 = 0 and ν n is the radial velocity of target n , f c is the radar carrier frequency, T r is the pulse repetition interval and c is the propagation velocity.
Simulations in a setting with fluctuating targets model and noise background are carried out to construct the radar received echoes for RTD. An example of how different targets stand out in different SNR is presented in Figure 20. The emitted pulse is assumed to be a unit vector which does not incorporate any beneficial compression gains or additional antenna gains. The targets are assumed to fluctuate slowly and follow the standard Swerling distribution, where the value of σ n varies randomly. It is assumed to consider the impact of path propagation and other environmental effects indirectly. By time and frequency domain analysis of radar received echoes, targets or environment characteristics can be obtained.
Using raw data as the input to perform RTD training without any preprocessing method is a complete “end-to-end” task, and the radar echo cube enables the network to exploit the temporal and spatial correlation simultaneously [49], which is presented in Figure 21. Prashant et al. [58] used radar received echoes to detect signals in a non-Gaussian noise environment. However, for raw data, it may require a deeper network to extract target features; in other words, the increasing complex detection environment makes the “end-to-end” tasks more challenging.

4.2.2. Pulse-Range Maps

Radar emits multiple pulses in the CPI, and the reflected signals are integrated in a coherent or incoherent way. The axes of the original echo data can be labelled as ‘fast’ and ‘slow’ time, where fast time is utilized to calculate distance and slow time is used to calculate doppler velocity (see in Figure 22). Pulse compression is applied to the echo signal via a standard matched filtering operation, using the reference signal over the fast time, which is calculated as (5):
S ^ r ( t , k ) = S * ( t ) S r ( t , k )
The operation of pulse compression through matched filter can obtain a narrow pulse width and a high resolution of range profile, but has no impact on the radar detection range. Therefore, the problem of RTD can be converted into target detection and location in Pulse-Range maps. The upper part of Figure 22 shows an example of the Pulse-Range map. Akhtar et al. [60] used Pulse-Range maps to detect targets in a noisy background, and in [69], the Pulse-Range maps were applied to ranging and detecting targets from sea clutter.

4.2.3. Range-Doppler Maps

Subsequent to gathering all pulses, doppler processing is applied over multiple pulses by applying fast Fourier transform (FFT) over a slow time at each range. The slow time domain of S ^ r ( t , k ) is multiplied by a window function firstly and then FFT is used to yield a range-Doppler spectrum as (6):
S ^ r ( t , ω ) = F ( w i n × S ^ r ( t , k ) )
where w i n is a windowing function, F denotes the discrete Fourier transform (DFT). Following DFT, targets with a steady pace will appear concentrated in doppler. An example of doppler processing after matched filtering is shown in Figure 22. The Range-Doppler map in Figure 22 shows three targets at different doppler bins, a close target at range 180 samples having a positive velocity, which indicates that the target is moving towards the radar (shown as orange grid in Figure 22), and 2 moving targets at the same range 250 samples but different doppler bins (shown as blue and green grids respectively). Multiple frames are extracted from the raw data, producing a sequence of Range-Doppler frames.
Range-Doppler maps with targets absent and present is presented in Figure 23. The target echo occupies multiple cells, forming a mountain-like shape in the spectrum. If the SNR is low, the two targets at the range of 250 samples are not obvious in the spectrum. Researchers often consider the Range-Doppler map as an image, and use a classifier to classify the image as target absent or present. Target present examples with different SNR can be generated By setting a different noise power. In Refs. [62,65,66,67], range-Doppler maps were utilized in various RTD tasks.

4.2.4. Time-Frequency Images

The time-frequency analysis method is an effective and powerful tool in analyzing time-varying non-stationary signals because of the time-varying characteristics of target micro-motion. After demodulation and pulse compression are applied on the radar echo, STFT is often adopted to convert the radar echo to two-dimensional time-frequency images, as shown in (7):
S T F T ( t , f ) = + S ^ r ( δ , k ) g * ( δ t ) e j 2 π f δ d δ
where g ( t ) is a narrow windowing function such as a Hamming window function. Since the micro-motion characteristics of targets differ from background noise or clutter, time-frequency images are utilized to build training and testing datasets. Su et al. [72] adopted a time-frequency analysis method to analyze micro-motion characteristics, thereby differentiating targets from clutter.

4.3. Performance Evaluation

RTD is a fundamental process for separating targets of interest from background noise. A primary goal of RTD is to satisfy two very contradictory requirements: acquire a high P D with a low P f a . Comparing the P D under the same P f a with the DNN-based detector and CFAR detector respectively is the common metric of detection performance [59,70]. P D and P f a are defined as (8) and (9):
P D = N t d N t o t a l _ t a r g e t s
P f a = N f d N d e _ t a r g e t s
where N t d is the total number of truly detected targets, N t o t a l _ t a r g e t s denotes the total number of targets in the sample, N f d is the total number of falsely detected targets (mistaken for targets), N d e _ t a r g e t s is the total number of non-targets in the sample.

4.4. Summarization of Dataset and Preprocessing

The actual radar data is not widely accessible at present because of the particularity of RTD tasks. Although a few of the existing datasets, such as, IPIX and CSIR, whose data have been used for sea surface targets detection and sea clutter characteristic analysis, most recent research on RTD still relies on synthetic or simulation data. In this section, we describe how to construct radar received echoes and preprocess three types of data which are widely used in RTD. Three types of data processed in radar signal processing are applied in different target detection tasks with respective characteristics. Since pulse compression could obtain a high resolution of the range profile, pulse-range maps are mainly used in radar ranging tasks. Range-Doppler maps are more commonly used in position and velocity measurement in complex scenarios with background noise or clutter because doppler processing could improve SNR effectively. The time-frequency analysis method is efficient and especially suitable for targets with distinct micro-motion characteristics. An example would be maritime target detection and recognition, which can make full use of the micro-motion information of the target and clutter.

5. Discussion

Although deep learning-based approaches have had some successful applications in the field of RTD, the following challenges remain:

5.1. Dataset Deficiency

One of the difficulties in applying deep learning-based methods to RTD is the lack of publicly available labelled data. The difference with other applications is the high cost of radar data collection. Currently, radar system simulation modeling is a method to solve the problem, but it is computationally demanding and extremely challenging to generate alternative data because multiple factors need to be considered, e.g., multiple attenuation, discrete cells and multipath reflections. In other words, even if all the above issues are avoided, relying on mathematical models for simulation is inevitable, which may introduce the problem of model fidelity. Similar inaccuracies have emerged in [72].
We still hope to get enough data for training from real-world radar systems, which will certainly make the detection model more reliable. In order to further advance the research of RTD in the absence of realistic radar data, the following approaches can be considered: (1) Data augmentation, which has been used in [67,68]; (2) Developing more robust deep learning-based algorithms with insufficient training data, such as generative adversarial networks (GAN) [117]; (3) Establishing more advanced RTD frameworks seems to be a new trend; (4) Developing learning-based methods, such as meta learning [118,119] and transfer learning [120,121], which can overcome the limitation of insufficient data and insensitivity to the changes of the radar detection environment.

5.2. Varied Models in Complex Tasks

Although the application of deep learning technology in RTD has made remarkable progress, the existing literatures on ANNs and DNNs of RTD are still relatively sparse and not mature enough. For instance, a common aspect found in many of the papers cited above is the moderate size of ANN, where the networks proposed in [58,61] contain only one hidden layer. In addition, although many other architectures have been proposed in DNN, only CNN is widely used in RTD. Therefore, there is plenty of room to explore various DNN in RTD. Signal processing system that can efficiently suppress target RCS fluctuations, noise, clutter, and jamming is always considered as one of the key directions of radar research and development. How to effectively distinguish targets from strong active jammed signals is still under investigation. How to verify other deep learning architectures can perform better in RTD also needs to be considered. These also imply that newer, more diverse but practical powerful learning schemes for more complex tasks are required urgently.
The advance in computing power allows the training of large-scaled deep learning models on massive data. Remarkable progress in deep learning algorithms and great advances in radar system would benefit each other. It is worth exploring replacing the entire radar signal processing by deep learning methods.

5.3. Integrated Training Methods

Much of the existing research shows that the deep fusion of traditional signal processing methods and deep learning-based schemes in RTD application is an evident trend [122]. For one thing, as a kind of data preprocessing method, traditional radar signal processing methods, such as pulse compression, coherent accumulation, STFT, etc., are beneficial to enhance features, thus improving the detection performance. For another, according to the previous review, neural networks combined with typical radar signal processing approaches (e.g., CFAR) as a learning strategy can help improve detection performance. In addition, we believe that models based on deep learning can provide an “end-to-end” framework for integrating perception, processing and decision making. Besides, studies of simulation process optimization, performance judgment criteria and other basic problems, such as model interpretability and generalization ability, are still under investigation.

6. Conclusions

Research efforts in artificial neural network and deep learning models in RTD has been discussed in this review. Various architectures of networks for various application schemes have been investigated. The results obtained have shown that deep learning-based detectors perform better than the traditional processing methods to some certain degree in some specific cases. Although the study of deep learning in the field of RTD is at the initial stage and still faces some challenges, there is no doubt that the research and usage of deep learning-based methods in RTD will contribute great improvement in the future. We wish to take a step towards making deep learning more applicable to the related fields of radar signal processing.

Author Contributions

Conceptualization, W.J. and Y.L.; methodology, W.J. and Y.R.; investigation, W.J. and Y.R.; resources, Y.L. and J.L.; writing—original draft preparation, W.J.; writing—review and editing, Y.L.; supervision, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

This work was supported in part by the National Natural Science Foundation of China under Grant 71671178, 62176247, 62036007, 62050175, and 62102057. It was also supported by the Fundamental Research Funds for the Central Universities and the Science and Technology Research Program of Chongqing Municipal Education Commission (KJQN-202100627).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Richards, M.A. Fundamentals of Radar Signal Processing, 1st ed.; McGraw-Hill: New York, NY, USA, 2005. [Google Scholar]
  2. Kay, S.M. Fundamentals of Statistical Signal Processing: Detection Theory; Prentice Hall PTR: Hoboken, NJ, USA, 1998; Volume 2. [Google Scholar]
  3. Melvin, W.L.; Scheer, J.A. (Eds.) Principles of Modern Radar; SciTech Publishing: Raleigh, NC, USA, 2013. [Google Scholar]
  4. Long, T.; Liang, Z.; Liu, Q. Advanced technology of high-resolution radar: Target detection, tracking, imaging, and recognition. Sci. China Inf. Sci. 2019, 62, 1–26. [Google Scholar] [CrossRef] [Green Version]
  5. LeCun, Y.; Bengio, Y.; Hinton, G. Deep Learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  6. Bengio, Y.; Courville, A.; Vincent, P. Representation Learning: A Review and New Perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef] [PubMed]
  7. Xisto, L.T.; Sergio, L.A.; Nathan, I. Artificial Neural Networks and Machine Learning techniques applied to Ground Penetrating Radar: A review. Appl. Comput. Inf. 2018, 17, 296–308. [Google Scholar] [CrossRef]
  8. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; The MIT Press: Cambridge, MA, USA, 2016. [Google Scholar]
  9. Mason, E.; Yonel, B.; Yazici, B. Deep Learning for Radar. In Proceedings of the 2017 IEEE Radar Conference, Seattle, WA, USA, 8–12 May 2017. [Google Scholar] [CrossRef]
  10. Wang, C.; Wang, J.; Zhang, X. Automatic Radar Waveform Recognition based on Time-frequency Analysis and Convolutional Neural Network. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2437–2441. [Google Scholar]
  11. Singhal, M.; Saini, G. Optimization of Antenna Parameters Using Artificial Neural Network: A Review. Int. J. Comput. Trends Technol. 2017, 44, 64–73. [Google Scholar] [CrossRef]
  12. Kim, Y.; Keely, S.; Ghosh, J.; Ling, H. Application of Artificial Neural Networks to Broadband Antenna Design Based on a Parametric Frequency Model. IEEE Trans. Antennas Propag. 2007, 55, 669–674. [Google Scholar] [CrossRef]
  13. Rawat, A.; Yadav, R.; Shrivastava, S. Neural network applications in smart antenna arrays: A review. Int. J. Electron. Commun. 2012, 66, 903–912. [Google Scholar] [CrossRef]
  14. Elbir, A.M.; Mishra, K.V.; Eldar, Y.C. Cognitive radar antenna selection via deep learning. IET Radar Sonar Navig. 2019, 13, 871–880. [Google Scholar] [CrossRef] [Green Version]
  15. Haykin, S. Cognitive radar: A way of the future. IEEE Signal Process. Mag. 2006, 23, 30–40. [Google Scholar] [CrossRef]
  16. Gao, J.; Lu, Y.; Qi, J.; Shen, L. A Radar Signal Recognition System Based on Non-Negative Matrix Factorization Network and Improved Artificial Bee Colony Algorithm. IEEE Access 2019, 7, 117612–117626. [Google Scholar] [CrossRef]
  17. Chen, S.; Wang, H. SAR Target Recognition based on Deep Learning. In Proceedings of the International Conference on Data Science and Advanced Analytics (DSAA), Shanghai, China, 30 October–1 November 2014. [Google Scholar]
  18. Ding, J.; Chen, B.; Liu, H.; Huang, M. Convolutional Neural Network with Data Augmentation for SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2016, 13, 364–368. [Google Scholar] [CrossRef]
  19. Geng, J.; Fan, J.; Wang, H.; Ma, X.; Li, B.; Chen, F. High-resolution SAR Image Classification via Deep Convolutional Autoencoders. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2351–2355. [Google Scholar] [CrossRef]
  20. Liu, H.; Feng, B.; Chen, B.; Du, L. Radar high-resolution range profiles target recognition based on stable dictionary learning. IET Radar Sonar Navig. 2016, 10, 228–237. [Google Scholar] [CrossRef]
  21. Liao, K.; Si, J.; Zhu, F.; He, X. Radar HRRP Target Recognition Based on Concatenated Deep Neural Networks. IEEE Access 2018, 6, 29211–29218. [Google Scholar] [CrossRef]
  22. Feng, B.; Chen, B.; Liu, H. Radar HRRP target recognition with deep networks. Pattern Recognit. 2017, 61, 379–393. [Google Scholar] [CrossRef]
  23. Yu, S.; Xie, Y. Application of a Convolutional Autoencoder to Half Space Radar HRRP Recognition. In Proceedings of the 2018 International Conference on Wavelet Analysis and Pattern Recognition, Chengdu, China, 15–18 July 2018. [Google Scholar]
  24. Seyfioglu, M.S.; Gurbuz, S.Z. Deep Neural Network Initialization Methods for Micro-Doppler Classification with Low Training Sample Support. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2462–2466. [Google Scholar] [CrossRef]
  25. Kouemou, G. Radar Target Classification Technologies. In Radar Technology; InTech: Rijeka, Croatia, 2010. [Google Scholar]
  26. Lombacher, J.; Hahn, M.; Dickmann, J.; Wöhler, C. Potential of Radar for Static Object Classification using Deep Learning Methods. In Proceedings of the 2016 IEEE MTT-S International Conference on Microwaves for Intty (ICMIM), San Diego, CA, USA, 19–20 May 2016. [Google Scholar]
  27. Ningyuan, S.; Xiaolong, C.; Jian, G.; Xiaoqian, M.; Ningbo, L. Detection and Classification of Marine Target with Micro-motion Based on CNNs. J. Radar 2018, 7, 565–574. [Google Scholar] [CrossRef]
  28. Mendis, G.J.; Wei, J.; Madanayake, A. Deep Learning Cognitive Radar for Micro UAS Detection and Classification. In Proceedings of the 2017 Cognitive Communications for Aerospace Applications Workshop (CCAA), Cleveland, OH, USA, 27–28 June 2017; pp. 1–5. [Google Scholar]
  29. Kim, Y.; Moon, T. Human Detection and Activity Classification Based on Micro-Doppler Signatures Using Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2016, 13, 8–12. [Google Scholar] [CrossRef]
  30. Jiang, W.; Ren, Y.; Liu, Y.; Wang, Z.; Wang, X. Recognition of Dynamic Hand Gesture Based on Mm-wave FMCW Radar Micro-Doppler Signatures. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Toronto, ON, Canada, 6–11 June 2021. [Google Scholar]
  31. Dong, X.; Zhao, Z.; Wang, Y.; Zeng, T.; Wang, J.; Sui, Y. FMCW Radar-Based Hand Gesture Recognition using Spatiotemporal Deformable and Context-Aware Convolutional 5D Feature Representation. IEEE Trans. Geosci. Remote Sens. 2021. [Google Scholar] [CrossRef]
  32. Lohman, B.; Boric-Lubecke, O.; Lubecke, V.M.; Ong, P.W.; Sondhi, M.M. A Digital Signal Processor for Doppler Radar Sensing of Vital Signs. IEEE Eng. Med. Biol. Mag. 2001, 4, 3359–3362. [Google Scholar] [CrossRef] [Green Version]
  33. Wang, K.; Zeng, Z.; Sun, J. Through-Wall Detection of the Moving Paths and Vital Signs of Human Beings. IEEE Geosci. Remote Sens. Lett. 2018, 16, 717–721. [Google Scholar] [CrossRef]
  34. Qiu, L.; Jin, T.; Lu, B.; Zhou, Z. An Isophase-Based Life Signal Extraction in Through-the-Wall Radar. IEEE Geosci. Remote Sens. Lett. 2016, 14, 193–197. [Google Scholar] [CrossRef]
  35. Cardillo, E.; Li, C.; Caddemi, A. Millimeter-Wave Radar Cane: A Blind People Aid with Moving Human Recognition Capabilities. IEEE J. Electromagn. RF Microw. Med. Biol. 2021. [Google Scholar] [CrossRef]
  36. Chen, H.; Ye, W. Classification of Human Activity Based on Radar Signal Using 1-D Convolutional Neural Network. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1178–1182. [Google Scholar] [CrossRef]
  37. Barkat, M. On Adaptive Cell-averaging CFAR Radar Signal Detection. Ph.D. Thesis, Department of Electrical and Computer Engineering, Syracuse University, New York, NY, USA, 1987. [Google Scholar]
  38. Seyfioğlu, M.; Özbayoğlu, A.; Gürbüz, S. Deep Convolutional Autoencoder for Radar-based Classification of Similar Aided and Unaided Human Activities. IEEE Trans. Aerosp. Electron. Syst. 2018, 54, 1709–1723. [Google Scholar] [CrossRef]
  39. Park, J.; Rios, J.; Moon, T.; Kim, Y. Micro-Doppler based Classification of Human Activities on Water via Transfer Learning of Convolutional Neural Networks. Sensors 2016, 16, 19–90. [Google Scholar] [CrossRef] [PubMed]
  40. Khalid, H.; Pollin, S.; Rykunov, M.; Bourdoux, A.; Sahli, H. Convolutional Long Short-Term Memory Networks for Doppler-Radar based Target Classification. In Proceedings of the 2019 IEEE Radar Conference, Boston, MA, USA, 22–26 April 2019. [Google Scholar] [CrossRef]
  41. Jokanovic, B.; Amin, M. Fall Detection Using Deep Learning in Range-Doppler Radars. IEEE Trans. Aerosp. Electron. Syst. 2017, 54, 180–189. [Google Scholar] [CrossRef]
  42. Cortes, C.; Vapnik, V.N. Support Vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  43. Short, R.; Fukunaga, K. The optimal distance measure for nearest neighbor classification. IEEE Trans. Inf. Theory 1981, 27, 622–627. [Google Scholar] [CrossRef]
  44. Mountrakis, G.; Im, J.; Ogole, C. Support vector machines in remote sensing: A review. ISPRS J. Photogramm. Remote Sens. 2011, 66, 247–259. [Google Scholar] [CrossRef]
  45. Youngwook, K. Application of Machine Learning to Antenna Design and Radar Signal Processing: A Review. In Proceedings of the 2018 International Symposium on Antennas and Propagation (ISAP), Busan, Korea, 23–26 October 2018. [Google Scholar]
  46. del Rey-Maestre, N.; Jarabo-Amores, M.P.; Mata-Moya, D.; Humanes, J.; Hoyo, P. Machine Learning Techniques for Coherent CFAR Detection Based on Statistical Modeling of UHF Passive Ground Clutter. IEEE J. Sel. Top. Signal Proc. 2018, 12, 104–118. [Google Scholar] [CrossRef]
  47. Mata-Moya, D.; Del-Rey-Maestre, N.; Peláez-Sánchez, V.M.; Jarabo-Amores, M.-P.; Martín-De-Nicolás, J. MLP-CFAR for improving coherent radar detectors robustness in variable scenarios. Expert Syst. Appl. 2015, 42, 4878–4891. [Google Scholar] [CrossRef]
  48. Cao, T.; Palmer, J.; Berry, P. False Alarm Control of CFAR Algorithms with Experimental Bistatic Radar Data. In Proceedings of the IEEE Radar Conference, Arlington, VA, USA, 10–14 May 2010; pp. 156–161. [Google Scholar]
  49. Jiang, W.; Ren, Y.; Liu, Y.; Leng, J. A Method of Radar Target Detection Based on Convolutional Neural Network. Neural Comput. Appl. 2021. [Google Scholar] [CrossRef]
  50. Sciotti, M.; Lombardo, P. Performance Evaluation of Radar Detection Schemes based on CA-CFAR Against K-distributed Clutter. In Proceedings of the 2001 CIE International Conference on Radar, Beijing, China, 15–18 October 2001; pp. 345–349. [Google Scholar]
  51. Gandhi, P.; Kassam, S. Optimality of the cell averaging CFAR detector. IEEE Trans. Inf. Theory 1994, 40, 1226–1228. [Google Scholar] [CrossRef]
  52. Rohling, H. Radar CFAR Thresholding in Clutter and Multiple Target Situations. IEEE Trans. Aerosp. Electron. Syst. 1983, 19, 608–621. [Google Scholar] [CrossRef]
  53. Rohling, H. New CFAR-Processor Based on An Ordered Statistic. In Proceedings of the IEEE 1985 International Radar Conference, Arlington, VA, USA, 6–9 May 1985. [Google Scholar]
  54. Ritcey, J.A. Censored Mean-level Detector Analysis. IEEE Trans. Aerosp. Electron. Syst. 1986, AES-22, 443–454. [Google Scholar] [CrossRef]
  55. Reed, R.; Marks, R.J. Neural Smithing: Supervised Learning in Feedforward Artificial Neural Networks; MIT Press: Cambridge, MA, USA, 1999. [Google Scholar]
  56. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006. [Google Scholar]
  57. Google’s AI Brain. Available online: http://research.google.com/teams/brain/ (accessed on 30 January 2017).
  58. Gandhi, P.P.; Ramamurti, V. Neural networks for signal detection in non-Gaussian noise. IEEE Trans. Signal Process. 1997, 45, 2846–2851. [Google Scholar] [CrossRef]
  59. Rohman, B.P.A.; Kurniawan, D.; Miftahushudur, M.T. Switching CA/OS CFAR using Neural Network for Radar Target Detection in Non-Homogeneous Environment. In Proceedings of the 2015 International Electronics Symposium (IES), Surabaya, Indonesia, 29–30 September 2015. [Google Scholar]
  60. Akhtar, J.; Olsen, K.E. A Neural Network Target Detector with Partial CA-CFAR Supervised Training. In Proceedings of the 2018 International Conference on Radar (RADAR), Brisbane, QLD, Australia, 27–31 August 2018. [Google Scholar]
  61. Amoozegar, F.; Sundareshan, M.K. A Robust Neural Network Scheme for CFAR Processing for Target Detection in Clutter Environment. In Proceedings of the American Control Conference, Baltimore, MD, USA, 29 June 1994. [Google Scholar]
  62. Cheikh, K.; Soltani, F. Application of Neural Networks to Radar Signal Detection in K-distributed Clutter. IEE Proc. Radar Sonar Navig. 2006, 153, 460–466. [Google Scholar] [CrossRef]
  63. Callaghan, D.; Burger, J.; Mishra, A.K. A Machine Learning Approach to Radar Sea Clutter Suppression. In Proceedings of the IEEE Radar Conference, Seattle, WA, USA, 8–12 May 2017. [Google Scholar] [CrossRef]
  64. Li, Y.; Xie, P.; Tang, Z.; Jiang, T.; Qi, P. SVM-Based Sea-Surface Small Target Detection: A False-Alarm-Rate-Controllable Approach. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1225–1229. [Google Scholar] [CrossRef] [Green Version]
  65. Akhtar, J.; Olsen, K.E. GO-CFAR Trained Neural Network Target Detectors. In Proceedings of the 2019 IEEE Radar Conference, Boston, MA, USA, 22–26 April 2019. [Google Scholar] [CrossRef]
  66. Wang, L.; Tang, J.; Liao, Q. A Study on Radar Target Detection Based on Deep Neural Networks. IEEE Sens. Lett. 2019, 3, 1–4. [Google Scholar] [CrossRef]
  67. Brodeski, D.; Bilik, I.; Giryes, R. Deep Radar Detector. In Proceedings of the 2019 IEEE Radar Conference, Boston, MA, USA, 22–26 April 2019. [Google Scholar] [CrossRef]
  68. Wheeler, T.A.; Holder, M.; Winner, H.; Kochenderfer, M.J. Deep Stochastic Radar Models. In Proceedings of the 28th IEEE Intelligent Vehicles Symposium, Los Angeles, CA, USA, 11–14 June 2017; Available online: Github.com/tawheeler/2017_iv_deep_radar (accessed on 20 December 2021).
  69. Pan, M.; Chen, J.; Wang, S.; Dong, Z. A Novel Approach for Marine Small Target Detection Based on Deep Learning. In Proceedings of the IEEE 4th International Conference on Signal and Image Processing, Wuxi, China, 19–21 July 2019; pp. 395–399. [Google Scholar] [CrossRef]
  70. Kang, M.; Leng, X.; Lin, Z.; Ji, K. A Modified Faster R-CNN based on CFAR Algorithm for SAR Ship Detection. In Proceedings of the 2017 International Workshop on Remote Sensing with Intelligent Processing (RSIP), Shanghai, China, 18–21 May 2017. [Google Scholar] [CrossRef]
  71. Mou, X.; Chen, X.; Guan, J. Marine Target Detection Based on Improved Faster R-CNN for Navigation Radar PPI Images. In Proceedings of the 2019 International Conference on Control, Automation and Information Sciences, Chengdu, China, 23–26 October 2019. [Google Scholar]
  72. Risueno, G.; Haykin, S.; Oliver, R. Convolutional Neural Networks for Radar Detection; Springer: Berlin/Heidelberg, Germany, 2002; pp. 1150–1155. [Google Scholar]
  73. Su, N.; Chen, X.; Guan, J.; Li, Y. Deep CNN-Based Radar Detection for Real Maritime Target Under Different Sea States and Polarizations. In Cognitive Systems and Signal Processing; Springer: Singapore, 2019; Volume 1006. [Google Scholar] [CrossRef]
  74. Wang, Y.; Zhang, Y.; Qu, H.; Tian, Q. Target Detection and Recognition Based on Convolutional Neural Network for SAR Image. In Proceedings of the 2018 11th International Congress on Image and Signal Processing, Biomedical Engineering and Informatics, Beijing, China, 13–15 October 2018. [Google Scholar]
  75. Yang, T.; Zhu, J.; Liu, J. SAR Image Target Detection and Recognition based on Deep Network. In Proceedings of the 2019 SAR in Big Data Era, Beijing, China, 5–6 August 2019. [Google Scholar] [CrossRef]
  76. Zheng, T.; Wang, J.; Lei, P. Deep Learning based Target Detection Method with Multi-features in SAR Imagery. In Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 26–29 November 2019. [Google Scholar]
  77. Zhang, J.; Xing, M.; Sun, G. A Fast Target Detection Method for SAR Image Based on Electromagnetic Characteristics. In Proceedings of the 2018 China International SAR Symposium (CISS), Shanghai, China, 10–12 October 2018. [Google Scholar]
  78. Zhao, C.; Wang, P.; Wang, J.; Men, Z. A Marine Target Detector based on CNN and Embedded Device for GF-3 Images. In Proceedings of the 2019 6th Asia-Pacific Conference on Synthetic Aperture Radar (APSAR), Xiamen, China, 26–29 November 2019. [Google Scholar]
  79. Moller, M.F. A Scaled Conjugate Gradient Algorithm for Fast Supervised Learning. Neural Netw. 1993, 6, 525–533. [Google Scholar] [CrossRef]
  80. Robbins, H.; Monro, S. A Stochastic Approximation Method. Ann. Math. Stat. 1951, 22, 400–407. [Google Scholar] [CrossRef]
  81. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  82. Zeiler, M.D. Adadelta: An Adaptive Learning Rate Method. Comput. Sci. 2012. Available online: https://arxiv.org/abs/1212.5701v1 (accessed on 20 December 2021).
  83. Kassam, S.A. Signal Detection in Non-Gaussian Noise; Springer: New York, NY, USA, 1988. [Google Scholar]
  84. Garth, L.M.; Poor, H.V. Detection of Non-Gaussian Signals: A Paradigm for Modern Statistical Signal Processing. Proc. IEEE 1994, 7, 1061–1095. [Google Scholar] [CrossRef]
  85. Gandhi, P.P.; Kassam, S.A. Analysis of CFAR processors in nonhomogeneous background. IEEE Trans. Aerosp. Electron. Syst. 1988, 24, 427–445. [Google Scholar] [CrossRef]
  86. Jarabo-Amores, P.; Mata-Moya, D.; Zurera, M.R.; Nieto-Borge, J.C.; L’opez-Ferreras, F. A Neural Network Approach to Improve Radar Detector Robustness. In Proceedings of the 14th European Signal Processing Conference, Florence, Italy, 4–8 September 2006. [Google Scholar]
  87. Galvez, N.; Pasciaroni, J.; Agamennoni, O.; Cousseau, J. Radar Signal Detector Implemented with Artificial Neural Networks. In Proceedings of the XIX Congreso Argentino de Control Automatico; 2004. Available online: https://www.researchgate.net/publication/236222896_RADAR_SIGNAL_DETECTOR_IMPLEMENTED_WITH_ARTIFICIAL_NEURAL_NETWORKS (accessed on 20 December 2021).
  88. Wang, J.; Zheng, T.; Lei, P.; Wei, S. Study on Deep Learning in Radar. Radars 2018, 7, 395–411. [Google Scholar]
  89. Huang, G.; Liu, Z.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; Volume 1, pp. 2261–2269. [Google Scholar]
  90. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar] [CrossRef] [Green Version]
  91. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [Green Version]
  92. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar] [CrossRef]
  93. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Network. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [Green Version]
  94. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  95. Liu, W.; Anguelov, D.; Erhan, D. SSD: Single Shot Multi-box Detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  96. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 386–397. [Google Scholar]
  97. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  98. Ma, M.; Chen, J.; Liu, W.; Yang, W. Ship Classification and Detection Based on CNN Using GF-3 SAR Images. Remote Sens. 2018, 10, 2043. [Google Scholar] [CrossRef] [Green Version]
  99. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  100. Borui, J.; Ruixuan, L.; Jiayuan, M.; Tete, X.; Yuning, J. Acquisition of Localization Confidence for Accurate Object Detection. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  101. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS Improving Object Detection with One Line of Code. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5562–5570. [Google Scholar]
  102. Djork-Arné, C.; Unterthiner, T.; Hochreiter, S. Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs). Comput. Sci. 2015. Available online: https://arxiv.org/abs/1511.07289v1 (accessed on 20 December 2021).
  103. Chen, X.; Guan, J.; He, Y. Applications and Prospect of Micro-motion Theory in the Detection of Sea Surface Target. J. Radars 2013, 2, 123–134. [Google Scholar] [CrossRef]
  104. Chen, X.; Guan, J.; Bao, Z.; He, Y. Detection and Extraction of Target with Micro-Motion in Spiky Sea Clutter via Short-time Fractional Fourier Transform. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1002–1018. [Google Scholar] [CrossRef]
  105. LeCun, Y.; Bottou, L. Gradient-based Learning Applied to Document Recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  106. Szegedy, C. Going Deeper with Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  107. Drosopoulos, A. Description of the OHGR Database; Defense Research Establishment: Ottawa, ON, Canada, 1994; Available online: https://www.researchgate.net/publication/235127506_Description_of_the_OHGR_database (accessed on 20 December 2021).
  108. Ding, H.; Liu, N.; Dong, Y. Overview and Prospects of Radar Sea Clutter Measurement Experiments. J. Radars 2019, 8, 281–302. [Google Scholar]
  109. IPIX Database in 1993: A Subset of the Dartmouth Database. Available online: http://soma.ece.mcmaster.ca/ipix/dartmouth/datasets.html (accessed on 20 August 2021).
  110. The Grimsby Database. Available online: http://soma.mcmaster.ca/ipix.php (accessed on 20 August 2021).
  111. Yan, L.; Sun, P.; Yi, L.; Han, N.; Tang, J. Modeling of Compound-gaussian Sea Clutter based on an Inverse Gaussian Distribution. J. Radars 2013, 2, 461–465. [Google Scholar] [CrossRef]
  112. Chen, X.; Guan, J.; Huang, Y. Radon-linear Canonical Ambiguity Function-based Detection and Estimation Method for Marine Target with Micromotion. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2225–2240. [Google Scholar] [CrossRef]
  113. Inggs, M.; Griffiths, H.; Fioranelli, F.; Ritchie, M.; Woodbridge, K. Multistatic Radar: System Requirements and Experimental Validation. In Proceedings of the 2014 IEEE International Radar Conference, Lille, France, 13–17 October 2014; pp. 1–6. [Google Scholar]
  114. Herselman, P.; Baker, C. Analysis of Calibrated Sea Clutter and Boat Reflectivity Data at C- and X-band in South African Coastal Waters. In Proceedings of the IET International Conference on Radar Systems, Edinburgh, UK, 15–18 October 2007. [Google Scholar] [CrossRef] [Green Version]
  115. De Wind, H.J.; Cilliers, J.C.; Herselman, P.L. DataWare: Sea Clutter and Small Boat Radar Reflectivity Database. IEEE Signal Process. 2010, 27, 145–148. [Google Scholar] [CrossRef]
  116. Herselman, P.; Barker, C.; de Wind, H.J. Analysis of X-band Calibrated Sea Clutter and Small Boat Reflectivity at Medium-to-low Grazing Angles. Int. J. Navig. Obs. 2008, 2008, 347518. [Google Scholar] [CrossRef] [Green Version]
  117. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B. Generative Adversarial Networks. Adv. Neural Inf. Processing Syst. 2014, 2672–2680. [Google Scholar] [CrossRef]
  118. Hospedales, T.; Antoniou, A.; Micaelli, P.; Storkey, A. Meta-Learning in Neural Networks: A Survey. arXiv 2020, arXiv:2004.05439. [Google Scholar] [CrossRef]
  119. Ma, L.; Liu, Y.; Zhang, X.; Ye, Y.; Yin, G.; Johnson, B.A. Deep Learning in Remote Sensing Applications: A Meta-analysis and Review. ISPRS J. Photogramm. Remote Sens. 2019, 152, 166–177. [Google Scholar] [CrossRef]
  120. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How Transferable are Features in Deep Neural Networks. Adv. Neural Inf. Proc. Syst. 2014, 2, 3320–3328. [Google Scholar]
  121. Oquab, M.; Bottou, L.; Laptev, I. Learning and Transferring Mid-Level Image Representations using Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar] [CrossRef] [Green Version]
  122. Evans, M.C.; Ruf, C.S. Toward the Detection and Imaging of Ocean Microplastics with a Spaceborne Radar. IEEE Trans. Geosci. Remote Sens. 2021, 1–9. [Google Scholar] [CrossRef]
Figure 1. Radar tasks that can benefit from deep learning.
Figure 1. Radar tasks that can benefit from deep learning.
Electronics 11 00156 g001
Figure 2. Major milestone of deep learning in RTD.
Figure 2. Major milestone of deep learning in RTD.
Electronics 11 00156 g002
Figure 3. A schematic of RTD in a complex scene.
Figure 3. A schematic of RTD in a complex scene.
Electronics 11 00156 g003
Figure 4. A signal processing diagram of the Pulse Doppler radar.
Figure 4. A signal processing diagram of the Pulse Doppler radar.
Electronics 11 00156 g004
Figure 5. The diagram of typical CFAR detectors.
Figure 5. The diagram of typical CFAR detectors.
Electronics 11 00156 g005
Figure 6. Schematic diagram of CFAR detection area, (a) One-dimensional detection area, (b) Two-dimensional detection area.
Figure 6. Schematic diagram of CFAR detection area, (a) One-dimensional detection area, (b) Two-dimensional detection area.
Electronics 11 00156 g006
Figure 7. A typical architecture of a multi-layer ANN detector.
Figure 7. A typical architecture of a multi-layer ANN detector.
Electronics 11 00156 g007
Figure 8. Design of ANN in switching CA/OS CFAR.
Figure 8. Design of ANN in switching CA/OS CFAR.
Electronics 11 00156 g008
Figure 9. Artificial neural network detector structure.
Figure 9. Artificial neural network detector structure.
Electronics 11 00156 g009
Figure 10. Input type of multiple CNN-based detectors, (a) Radar received echo; (b) Range-Doppler spectrum; (c) Pulse-Range image; (d) Time-Frequency image; (e) SAR image; (f) PPI image.
Figure 10. Input type of multiple CNN-based detectors, (a) Radar received echo; (b) Range-Doppler spectrum; (c) Pulse-Range image; (d) Time-Frequency image; (e) SAR image; (f) PPI image.
Electronics 11 00156 g010
Figure 11. Structure of a CNN detector for RTD with radar echo data.
Figure 11. Structure of a CNN detector for RTD with radar echo data.
Electronics 11 00156 g011
Figure 12. Structure of a CNN detector for RTD with Range-Doppler spectrum.
Figure 12. Structure of a CNN detector for RTD with Range-Doppler spectrum.
Electronics 11 00156 g012
Figure 13. Faster R-CNN-based model for RTD with Pulse-Range images.
Figure 13. Faster R-CNN-based model for RTD with Pulse-Range images.
Electronics 11 00156 g013
Figure 14. Basic structure of the Faster R-CNN.
Figure 14. Basic structure of the Faster R-CNN.
Electronics 11 00156 g014
Figure 15. Structure of target detection model with multi-features in SAR images.
Figure 15. Structure of target detection model with multi-features in SAR images.
Electronics 11 00156 g015
Figure 16. The flowchart of Faster R-CNN-based detector in SAR images.
Figure 16. The flowchart of Faster R-CNN-based detector in SAR images.
Electronics 11 00156 g016
Figure 17. The structure of fused features of echo data and SAR images.
Figure 17. The structure of fused features of echo data and SAR images.
Electronics 11 00156 g017
Figure 18. The structure of fused features in Faster R-CNN.
Figure 18. The structure of fused features in Faster R-CNN.
Electronics 11 00156 g018
Figure 19. Preprocessing diagram of the dataset for the radar detection system.
Figure 19. Preprocessing diagram of the dataset for the radar detection system.
Electronics 11 00156 g019
Figure 20. Example of simulated radar received echoes with different SNR, left: SNR = 10 dB, right: SNR = 0 dB.
Figure 20. Example of simulated radar received echoes with different SNR, left: SNR = 10 dB, right: SNR = 0 dB.
Electronics 11 00156 g020
Figure 21. Data preprocessing for a radar echo cube.
Figure 21. Data preprocessing for a radar echo cube.
Electronics 11 00156 g021
Figure 22. Schematic diagram of a pulse compression and Doppler processing.
Figure 22. Schematic diagram of a pulse compression and Doppler processing.
Electronics 11 00156 g022
Figure 23. Target present and Target absent in Range-Doppler Map.
Figure 23. Target present and Target absent in Range-Doppler Map.
Electronics 11 00156 g023
Table 1. Summary of RTD tasks and methods in recent literature.
Table 1. Summary of RTD tasks and methods in recent literature.
TaskSignal Processing MethodDetection MethodMain ContributionsRef.Year
Detecting signal in non-Gaussian noise——NNNN is employed to detect signals in non-Gaussian noise.[58]1997
Detecting targets in non-homogeneous noiseCA-CFAR and OS-CFARANNANN is employed to improve target detection performance by switching CA/OS CFAR.[59]2015
Detecting targets in noisy backgroundPulse compression and CA-CFARANNANN detector combined with CA-CFAR detector to offer a lower false alarm rate than CFAR.[60]2018
Detecting targets in clutter environmentStatistical parameters of target and clutter fluctuationsNNA NN-CFAR detection scheme is presented to offer a robust performance in the face of loss of reference cells.[61]1994
Detecting signal in K-distributed clutterCFARANNANN-CFAR detector with MLP and RBF architecture is employed to detect signals in K-distributed clutter.[62]2006
Detecting targets in sea clutterPulse compression and Doppler processingSVM&KNNSVM and KNN are used for suppression of sea clutter.[63]2017
Detecting targets in sea clutter——SVMSVM-based detector can flexibly control the false alarm rate.[64]2019
Detecting targets embedded in clutterDoppler processing and GO-CFARANNANN detector combined with GO-CFAR detector to offer a higher detection performance than CFAR.[65]2019
Detecting target absent or presentPulse compression and Doppler processingCNNThe classical CFAR detector is replaced by a CNN detector.[66]2019
Target detection in 4D spaceDoppler processingCNNCNN is used to detect and localize in the 4D space of range, Doppler, azimuth and elevation.[67]2019
Predicting targets’ location and power distributions——AutoencoderConstruct input information and employ DNNs as radar models to learn conditional probability end-to-end from data.[68]2017
Ranging and detecting target from sea clutterPulse compressionFaster R-CNNFaster R-CNN is applied to achieve the target detection and localization with low SCR.[69]2019
Detecting target absent or presentRDA and CFARFaster R-CNNFaster R-CNN combined with CFAR to detect small-sized targets.[70]2017
Detecting targets in sea clutter——Faster R-CNNImproved Faster R-CNN are used for target detection in navigation radar PPI images. [71]2019
Detecting targets in clutterWVDCNNWVD-CNN detector is used for clutter analysis.[72]2002
Detecting target or no-target under different sea statesPulse compression and STFTCNNCNNs are used for the detection of target micro-Doppler.[73]2019
RDA: Range Doppler Algorithm; WVD: Wigner-Ville Distribution; STFT: Short Time Fourier Transform; detection in 4D space: 4 dimension information of range, velocity, azimuth and elevation.
Table 2. Summary of RTD models in the literature.
Table 2. Summary of RTD models in the literature.
Arch.NameAuthorModelInputActivation FunctionPoolingRegularizationOptimizationOther
Resources
Ref.
ANNNNGandhi et al.1 hidden layerRadar received signalsSigmoidNone————Simulated data[58]
ANNRohman et al.2 hidden layersCA, OS and CUT dataLog-sigmoidNone————Simulated data[59]
ANNAkhtar et al.2 hidden layers, 32 nodes NNPulse-Range mapsHyperbolic tangentNone——SCGSimulated data[60]
NN-CFARAmoozegar et al.2 hidden layers9 statistical parametersSigmoidNone————Simulated data[61]
ANNCheikh et al.1 hidden layerRange-Doppler mapsSigmoidNone————Simulated data[62]
ANNAkhtar et al.4 hidden layers, 19 nodes in each layerRange-Doppler mapsHyperbolic tangentNone——SCGSimulated data[65]
DNNCNNWang et al.8 layers CNNRange-Doppler mapsReLUMax——SGDSimulated data[66]
RD-Net +
Ang-Net
Brodeski et al.CNN-basedRange-Doppler mapsReLUMaxDropoutAdamCollected data + Augmented data[67]
Faster R-CNNPan et al.RPN + CNN + RoIPulse-Range imagesReLUMaxSmooth L1
Dropout
GDCSIR dataset[69]
CNNWang et al.5 layers CNNSAR imagesSigmoidAverage————MSTAR dataset[74]
GoogLeNetYang et al.36 layers CNNSAR imagesReLUMaxDropout——MSTAR dataset[75]
ANN + CNNZheng et al.CNN-basedSAR imagesReLU————SGDSentinel-1 dataset 1[76]
Faster R-CNNKang et al.RPN + CNN + RoISAR imagesReLUMax————Sentinel-1 dataset 1[70]
Faster R-CNNZhang et al.RPN + CNN + RoIEcho data and SAR imagesReLUMax————Simulated data + MSTAR dataset[77]
SSDZhao et al.VGG16-basedSAR images————————Gaofen-3 dataset[78]
Faster R-CNNMou et al.RPN + CNN + RoIPPI imagesELUsMax——AdamCollected data[71]
LeNet&GoogLeNetSu et al.7 layers & 22 layers CNNTime-Frequency imagesSigmoid&ReLUMax & AverageDropoutSGDIPIX measured data[73]
DNNWheeler et al.AutoencoderSpatial raster and object listReLU————AdaCollected and generated data[68]
RD-Net +
Ang-Net
Jiang et al.CNN-basedEcho dataSoft-maxMaxSmooth L1AdamSimulated data[49]
ReLU: Rectified Linear Units, ELUs: Exponential Linear Units, SCG: Scale conjugate gradient [79], GD: gradient descent, SGD: stochastic gradient descent [80], Adam: Adaptive Moment Estimation [81], Ada: Adadelta [82], CA: CA-CFAR, OS: OS-CFAR, CUT: Cell under test, RBF: Radial Basis Function. 1 https://scihub.copernicus.eu/ (accessed on 30 December 2021).
Table 3. IPIX radar and experimental parameters.
Table 3. IPIX radar and experimental parameters.
IPIX Parameter
Radar parametersTX frequency (GHz)9.39Width of beam (°)0.9 pencil beam
Peak power (kW)8Antenna gain(dB)45.7
Pulse width(us)0.2Sampled range resolution (m)15
PRF (kHz)0.8/1Polarization modeHH;HV;VH;VV
Experiment
Summary
Year1993 11998 2
Longitude63°25.41′ W79°35′54.6″ W
Latitude44°36.72′ N43°12′41.0″ N
Height (m)3020
Duration (s)131——
Distance resolution (m)303~60
Target TypeSpherical buoyant apparatusFloating boat
Target range (m)2660/5525/2655——
Target Azimuth (°)128/130/170——
Quantization bits810
Environmental
parameters
SCR (dB)0~6——
Observation directionUpwind——
Significant wave height (m)1.0/1.5/2.1——
Douglas sea state2/3/4——
1 http://soma.ece.mcmaster.ca/ipix/dartmouth/datasets.html (accessed on 30 December 2021). 2 http://soma.mcmaster.ca/ipix.php (accessed on 30 December 2021).
Table 4. CSIR trial radar and experimental parameters.
Table 4. CSIR trial radar and experimental parameters.
Items Fynment RadarMonopulse Radar
Radar parametersTX frequency (GHz)6.5~17.58.8
Peak power (kW)2——
PRF (kHz)0~30Adjustable
Width of beam (°)≤2——
Antenna gain(dB)≥30——
Sampled range resolution (m)15/4515
Gates1~96——
Experiment summaryYear20062007
Setup siteOverberg Test RangeSignal Hill
Longitude20°17′17.46″ E18°23′53.76″ E
Latitude34°36′56.52″ S33°55′15.62″ S
Height (m)67294
Grazing angle (°)0.3~30.3~10
Duration (s)169.149.17
Maximum target range (km)1560
Target azimuth (°)90° N~225° N240°N~20° N
Environmental parametersMean wind speed (m/s)0~10.30~20.58
Wind direction180° N~270° N130° N~140° N,320° N~330° N
Significant wave height (m)1~3.81~6
Swell direction135° N~180° N230° N~270° N
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jiang, W.; Ren, Y.; Liu, Y.; Leng, J. Artificial Neural Networks and Deep Learning Techniques Applied to Radar Target Detection: A Review. Electronics 2022, 11, 156. https://doi.org/10.3390/electronics11010156

AMA Style

Jiang W, Ren Y, Liu Y, Leng J. Artificial Neural Networks and Deep Learning Techniques Applied to Radar Target Detection: A Review. Electronics. 2022; 11(1):156. https://doi.org/10.3390/electronics11010156

Chicago/Turabian Style

Jiang, Wen, Yihui Ren, Ying Liu, and Jiaxu Leng. 2022. "Artificial Neural Networks and Deep Learning Techniques Applied to Radar Target Detection: A Review" Electronics 11, no. 1: 156. https://doi.org/10.3390/electronics11010156

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop