Next Article in Journal
A Power Supply Technology for a Low-Power Online Monitoring Sensor Based on Electric Field Induction
Next Article in Special Issue
Robust Noise Suppression Technique for a LADAR System via Eigenvalue-Based Adaptive Filtering
Previous Article in Journal
A Monte Carlo Emissivity Model for Wind-Roughened Sea Surface
Previous Article in Special Issue
A Successive Approximation Time-to-Digital Converter with Single Set of Delay Lines for Time Interval Measurements
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Flying Small Target Detection for Anti-UAV Based on a Gaussian Mixture Model in a Compressive Sensing Domain

1
School of Computer Science, Shenyang Aerospace University, Shenyang 110136, China
2
School of Automation Science and Electrical Engineering, Beihang University, Beijing 100191, China
3
School of Electronic and Information Engineering, Shenyang Aerospace University, Shenyang 110136, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(9), 2168; https://doi.org/10.3390/s19092168
Submission received: 5 April 2019 / Revised: 2 May 2019 / Accepted: 6 May 2019 / Published: 10 May 2019
(This article belongs to the Special Issue Signal Processing for Intelligent Sensor Systems)

Abstract

:
Addressing the problems of visual surveillance for anti-UAV, a new flying small target detection method is proposed based on Gaussian mixture background modeling in a compressive sensing domain and low-rank and sparse matrix decomposition of local image. First of all, images captured by stationary visual sensors are broken into patches and the candidate patches which perhaps contain targets are identified by using a Gaussian mixture background model in a compressive sensing domain. Subsequently, the candidate patches within a finite time period are separated into background images and target images by low-rank and sparse matrix decomposition. Finally, flying small target detection is achieved over separated target images by threshold segmentation. The experiment results using visible and infrared image sequences of flying UAV demonstrate that the proposed methods have effective detection performance and outperform the baseline methods in precision and recall evaluation.

1. Introduction

In recent years, the Unmanned Aerial Vehicles (UAV) industry has been booming at a rate and scale and has gained emerging interest in both civilian and military applications. Especially, the consumer micro-UAV, such as drones, is growing exponentially and joins in the airspace setting for commercial and individual needs, such as aviation videography, courier services, traffic surveillance and land surveying [1]. However, as the rapid spread of micro-UAV, many new security threats and social problems, including disturbing aviation safety of commercial aircraft, breaching no-fly security in sensitive areas and invading public privacy, frequently occur and appear to be continually  increasing [2].
In spite of the laws and regulations being enforced and no-fly zone control being built in micro-UAV to restrain operators’ behavior, there still exists potentially dangers due to the unauthorized flying with spontaneous. As a consequence, there is a critical need for an anti-UAV system to detect a micro-UAV that enters a specific sensitive airspace in time. The traditional technology of anti-UAV is radar which works by means of active illumination with electromagnetic waves [3]. In the low-attitude airspace context, the radar signals of targets immerse in complex background clutter, e.g., trees, buildings [4,5]. Moreover, potentially dangerous targets should be detected when they are still far away, which means they may still be very small. Thereupon, the cost performance of expensive radar for micro-UAV detection is greatly reduced.
Fortunately, vision sensors are low-cost, passive and robust to jamming; hence, motivating their use for micro-UAV detection, even though they have a limited range compared to radars. As a matter of fact, there still exists many challenges to detect a micro-UAV with a visual method. This is because it not only occupies a small portion of the field of view from a far distance and flies against complex backgrounds under various weather and lighting conditions, but also appears changing size and diverse shapes in fully 3D dimensional environment.
Over the past few years, the problems of anti-UAV have been attracting more and more attention, but only a few studies of visual detection are studied. A non-technical way of anti-UAV was gathering information about potentially illegal or dangerous activity of UAV via crowd sourcing from their mobile phone by submitting a short online DroneALERT incident report [6]. The drone surveillance technologies and the existing anti-drone systems are comprehensively reviewed [2], and an anti-drone system called ADS-ZJU is developed by utilizing audio, video and RF technologies [7]. Sosnowski et al. presented a concept of short-range air defence systems to protect against drone threats with various optoelectronic sensors [8]. Inspired by different optical properties possessed by small drones with the sky as a background, a combination of infrared polarization filters was suggested to identify drones [9]. In addition, multiple local features of aircraft was used to compose a feature clustering criterion for vision based aircraft tracking [3]. Furthermore, some researchers have attempted to detect and track flying small UAVs from a single moving camera. Li et al. classified candidate targets by using their spatio-temporal traits on the basis of background motions estimation and distinctive points identification, and then a Kalman filter was used to boost the tracking performance [10]. Rozantsev et al. achieved object-centric motion stabilization of image patches by a regression-based approach, and  then detected flying objects by combining both appearance and motion cues as spatio-temporal image cubes [11,12]. In addition to these, the research of vision-based sense-and-avoid for micro-UAV could provide a reference [13,14].
In this study, multiple visual sensors are distributed around the surveillance area, and no-blind area coverage for surveillance is realized in a cooperative manner, so as to build an intelligent visual surveillance system for discovering the approaching UAV in time [15,16]. However, due to the optical properties and image resolution of visual sensors, it is limited to detect a target as far as possible. In  that case, the size of a target is often very small and with only a few pixels because of a long imaging distance, as well as the energy of the target received by an imaging sensor is very low and with dim spotlike features because of atmospheric effects [17,18,19]. Therefore, an efficient flying small target detection algorithm is crucial to give a warning as early as possible.
In this paper, we concentrate on flying small target detection for anti-UAV from a stationary camera. Making use of the advantages of compressive sensing which can efficiently compress a higher-dimensional data to lower-dimensional one and perfectly reconstruct from a small number of linear measurements with high probability, the image background modeling for flying small target detection is no longer over each pixel but over each linear measurement. Subsequently, the images in sequence are converted into patches to construct a background model for better enforcing the image neighboring correlation, and then the candidate patches which perhaps contain targets that are identified by using background models of each patch. Moreover, with the characteristics of background scene slight change and target relative salient movement within a finite time period, a matrix composed of images in sequence is separated into background images and target images by low-rank and sparse matrix decomposition. As a Consequence, flying small target detection for anti-UAV is achieved over a separated target image.
The remainder of this paper is organized as follows. Firstly, a new Gaussian mixture model in a compressive sensing domain is designed in Section 2. Section 3 presents the approach to flying small target detection composed by candidate patches identification, local foreground-background images separation and comprehensive scheme of the proposed method. Section 4 contains experiments to demonstrate the effectiveness of the proposed method and compare its performance with baseline methods. Finally, conclusions are drawn in Section 5.
In order to clearly depict this work, the major notations used throughout the rest of this paper are briefly summarized in Table 1.

2. Gaussian Mixture Modeling in a Compressive Sensing Domain

Traditionally, background modeling has been attracting significant attention in recent years for target detection in the foreground. To tackle this problem, the Gaussian mixture model (GMM) is proposed as a statistical approach by training multiple models of probability density functions over each pixel [20,21]. If a variable x is Gaussian distribution with mean μ and variance σ 2 , its probability density function (PDF) is defined as
N ( x ; μ , σ 2 ) = 1 2 π σ 2 e ( x μ ) 2 2 σ 2
However, the gaussian distribution is a single modal distribution, which can not provide a good approximation for the data distribution of multimodal states. Fortunately, according to previous literatures, GMM is considered to fit any probability distribution of a variable, that is to say, any variable can be regarded as with probability of multiple specific gaussian distributions [22]. Consequently, the GMM is composed by multiple Gaussian distributions, and its PDF is written as the linear sum of PDFs of the Gaussian distributions
p ( x ) = i = 1 s p ( i ) p ( x | i ) = i = 1 s w i N ( x ; μ i , σ i 2 )
where s is the number of Gaussian distribution, w i , μ i and σ i 2 are the weight, mean and variance of the i t h Gaussian distribution, respectively. In general, s is set to 3 5 , and the Gaussian distributions are independent to each other [23,24]. In this study, the flying target with salient movement is the foreground, while the background scene changes are relatively slight. Thereupon, a  Gaussian distribution will catch the pattern of follow-up compressive sensing measurements.
The pixel-based background modeling algorithms establish models for each pixel of an image and classify a new pixel as foreground if its PDF value is below a threshold [25]. Within this general paradigm, these algorithms require high computational complexity and energy consumption as well as large storage. As a new sampling theory, compressive sensing (CS) breaks through the Shannon/Nyquist sampling theorem which states that the sampling rate must be at least twice the highest frequency [26]. Compressive sensing is able to collect and keep sufficient information of a given signal by a small number of linear measurements. Correspondingly, perfect reconstruction or robust approximation of signal relies on effective sparse signal representation and appropriate recovery algorithms [27]. Therefore, background modeling can be performed on these linear measurements, as they are a low-dimensional compressive representations of the given signal.
An image is stacked to form a vector x = [ x 1 , x 2 , , x n ] T R n . The linear projection of the image vector on the measurement matrix Φ R m × n ( m n ) is defined as
y = Φ x
where y = [ y 1 , y 2 , , y m ] T R m represents the low dimensional CS measurement vector.
However, image itself may not be sparse but can be sparsely represented in some transform domain, such as DCT, Fourier, wavelet and over-complete dictionary. Assume that x R n can be sparsely represented in terms of a basis of vectors { ψ i } i = 1 q with ψ i R n , then forming the sparse basis matrix Ψ = [ ψ 1 , ψ 2 , , ψ q ] R n × q by stacking the vectors as columns, so x can express as
x = i = 1 q ψ i θ i = Ψ Θ
where Θ = [ θ 1 , θ 2 , , θ q ] T R q is the weighting coefficients vector of x in the Ψ domain [28].
Based on the above discussion, compressive sensing of an image can be described as linear projection process [29]
y = Φ Ψ Θ = A Θ
where A : = Φ Ψ is referred to as sensing matrix. In order for the well-conditioned reconstruction of a K-sparse signal x in the specific transform domain, the sensing matrix A R m × q should satisfies the restricted isometry property (RIP) with order 2 K [30].
Definition 1.
A matrix A satisfies the restricted isometry property (RIP) of order K if there exists a δ K ( 0 , 1 ) such that
( 1 δ K ) Θ 2 2 A Θ 2 2 ( 1 + δ K ) Θ 2 2 ,
for all Θ { Θ : Θ 0 K } , where Θ 0 : = ( i | θ i 0 ) is the zero pseudo-norm [31].
Unfortunately, the task of holding that the sensing matrix satisfies the RIP is an NP-hard problem, but an alternative approach is to make sure that the measurement matrix Φ is incoherent with the sparse basis matrix Ψ [29]. The greater incoherence of the measurement/sparsity pair ( Φ , Ψ ) , the smaller number of the linear measurements needed. In a Gaussian random matrix, the elements are independent and identically distributed (i.i.d.) random variables from a zero-mean, 1 / n -variance Gaussian density. Then, a Gaussian random matrix is incoherent with most of orthogonal sparse basis matrix, and is chosen as measurement matrix of compressive sensing to collect information of random projections
Φ = [ ϕ 1 , ϕ 2 , , ϕ m ] T
where ϕ i R n , and  ϕ i is Gaussian distribution with n-dimension N ( 0 , I ) .
With these considerations in mind, we replace the pixels with a small set of CS measurements to establish a background model for flying small target detection [32], employing the low dimensional CS measurement vector y = [ y 1 , y 2 , , y m ] T instead of the high dimensional image vector x = [ x 1 , x 2 , , x n ] T to establish Gaussian mixture model by Equation (2), and this process is so-called Gaussian mixture modeling in a compressive sensing domain (GMM_CS) [33].
When a new image is captured at time t during model training, a series of linear projections with Gaussian random matrix is operated over the image vector by Equation (3). Subsequently, each  element in CS measurement vector y is identified with previous Gaussian mixture model by
| y k t μ k , i t 1 | < ξ σ k , i t 1 ( i = 1 , 2 , , s )
where ξ is a constant. If there exists a matched Gaussian distribution, the corresponding mean and variance should be updated by utilizing current CS measurement
μ k t = ( 1 β ) μ k t 1 + β y k t
( σ k t ) 2 = ( 1 β ) ( σ k t 1 ) 2 + β ( y k t μ k t ) 2
where β ( 0 , 1 ) represents the learning rate of mean and variance of the Gaussian distribution. Then, the weights of s Gaussian distributions are updated simultaneously
w k , i t = ( 1 ( 1 ) τ α ) w k , i t 1
where α ( 0 , 1 ) denotes the learning rate of weight of the Gaussian distribution, τ { 0 , 1 } will set to 1 for the matched Gaussian distribution and to 0 for mismatched one. If there no matched Gaussian distribution, the mean and variance of the Gaussian distribution with minimum weight will be updated, and then the weights of s Gaussian distributions are normalized and let i = 1 s w k , i = 1 .

3. Approach to Flying Small Target Detection by Local Image Separation

3.1. Candidate Patches Identification Based on GMM_CS

From a general viewpoint, images captured by imaging systems for fly small target detection have large sizes, thus, when we treat a two-dimensional image by vectorizing it into a one-dimensional vector, the corresponding vector is with high dimensions [34]. As a consequence, there are two negative influences. On one hand, a fairly large measurement matrix is needed for CS which makes the storage and computations of a Gaussian ensembly very difficult [35]. On the other hand, the linear projection process over the high dimensional vector of whole image reckons without the image neighboring correlation.
In addressing the problems above, we break the whole image into overlapping or non-overlapping patches by patch transform [36]. Assume a new image is represented by a matrix D , the non-overlapping patch transform will convert the matrix into several block matrices with r rows and c columns
D = D 1 , 1 D 1 , 2 D 1 , c D 2 , 1 D 2 , 2 D 2 , c D r , 1 D r , 2 D r , c
where D i , j denotes the i t h row, j t h column image patch represented by a block matrix. For simplicity, each train image and test image are divided into non-overlapping patches for experiment in this study. In spite of the small size of the image patch leading to low computational complexity for one image patch, more image patches which perhaps maintain a flying small target, and subsequently, high computational complexity will be induced in the following process of target detection. On account of these considerations, we designate the size of image patch to refer to the roughly maximum size of flying target and division into integer patches.
Subsequently, the background modeling and target detection procedures will be applied over image patches. The background modeling is to establish a Gaussian mixture model for each linear measurement of image patches generated from a group of training images. After that, the obtained background model based on GMM_CS can be employed to identify the candidate patches which perhaps maintain a flying small target. Thereinto, the identification of an image patch maintains a flying small target, operates in a compressive sensing domain, and includes two steps. The first step is the comparison between each linear measurement of input image patch with the obtained background model by Equation (7); the second step is to compute the statistical ratio of matched linear measurements to the total linear measurements, which is expressed as
δ = 1 m k = 1 m J k
where δ ( 0 , 1 ] is regard as a threshold for identifying current image patch is maintain a flying small target or not, m is the total linear measurements of current image patch, and J is defined as
J k = 1 , if | y k μ k , i | < ξ σ k , i 0 , otherwise ( i = 1 , 2 , , s )
Based on the aforementioned discussion, a demonstration of candidate patches identification based on GMM_CS is depicted in Figure 1.

3.2. Local Target-Background Patches Separation

Along with continuously captured scene images of the imaging system, a group of consecutive candidate patches extracted form several adjacent images is designated and vectorized as columns to compose a data matrix G R d × l , where d is the total number of pixels in each candidate patch, and l is the number of adjacent images. Making use of the characteristics of background scene slight change and target relative salient movement within a finite time period, the data matrix G will be decomposed to a low-rank matrix L and a sparse matrix S via robust principal component analysis (RPCA) [37,38], where the low-rank matrix L can be regard as a composition of the separated background patches B 1 B l , and the sparse matrix S can be regard as a composition of the separated target patches T 1 T l , as shown in Figure 2.
Compared with the conventional principal component analysis which can offer low-rank representation of a given data matrix corrupted by small i.i.d. Gaussian noise, the RPCA can work well with respect to grossly corrupted observations, and then the low-rank and sparse matrix decomposition can be achieved by many kinds of optimization techniques [39], in which the optimization problem can be formulated as following
min L , S ( rank ( L ) + λ vec ( S ) 0 ) subject to G = L + S
where rank ( · ) is a function to solve the rank of matrix, vec ( · ) is vectorization for a matrix, · 0 is the 0 -norm for vector which denotes the number of non-zero entries, and λ is a weight parameter for tradeoff the low-rank matrix L and sparse matrix S.
Conventionally, the optimization problem Equation (14) can be exactly solved via convex optimization, in which the rank function is relaxed to the nuclear norm · * as its convex envelope, and the 0 -norm is relaxed to the 1 -norm as its convex envelope under a certain condition. Consequently, the optimization problem is converted to minimize a combination of the nuclear norm and the 1 -norm [40].
In this study, considering that each column of the data matrix G is the vectorized image and the images are obtained from a static camera observing a fixed scene, the rank of the expected low-rank matrix L can be approximately estimated. On the other hand, the pixel number of foreground target is specific in a short time, so the sparseness of the expected sparse matrix S is also appreciable. Therefore, the optimization problem Equation (14) can be recast into a combination of the rank-R matrix and K-sparse matrix approximation problem
min L , S vec ( G ( L + S ) ) 2 subject to rank ( L ) R , vec ( S ) 0 K
where · 2 is the 2 -norm for vector.
To solve the NP-hard problem, a  pursuits algorithm dubbed SpaRCS for sparse and low-rank decomposition via compressive sensing is adopted [41]. Note that, in this study, the linear operator A : R d × l R N is vectorization of a matrix, and its adjoining operator A * : R N R d × l denotes rearrangement from a vector to a matrix. The low-rank matrix L and the sparse matrix S are iteratively estimated by constructing two bases and attempting matrix recovery restricted to the bases. At each iteration, five steps: supports identification, supports merging, least squares estimation, supports pruning and residue updating are respectively processed to update its estimates of L and S.
On one hand, the low-rank matrix L is approximated by a set of up to R rank-1 matrices arranged from its singular vectors, and the set of entire singular vectors is defined as
O = { ( u i , v i ) : ( U , Σ , V ) = svd ( · ) }
where u i and v i are respectively the i t h column vector of matrices U and V, and svd ( · ) is a function to singular value decomposition (SVD) of a given matrix. Further assume that a subset of singular vectors Ω O , and let F Ω ( · ) denotes the projection operator onto the subspace spanned by Ω , then  the estimation of low-rank matrix can be converted to a optimization problem about Frobenius norm of the projection matrix [42].
On the other hand, the sparse matrix S is approximated by the largest K-term support set of the given matrix, and the set of total practicable terms is defined as
Q = { ( i , j ) : i = 1 , 2 , d ; j = 1 , 2 , l }
where ( i , j ) denotes the position of support term. Further assume that a subset of support terms Ω Q , then let H Ω R d × l denotes a 0-1 matrix with entity
h i j = 1 , if ( i , j ) Ω 0 , otherwise ( i = 1 , 2 , d ; j = 1 , 2 , l )
Subsequently, the estimation of sparse matrix can be converted to a optimization problem about Frobenius norm of dot product matrix [43].

3.3. Comprehensive Scheme and Algorithm Implementation

Based on the aforementioned discussion, a new approach to flying small target detection for anti-UAV based on a Gaussian mixture model in a compressive sensing domain is designed. The  approach includes two operation stages: background model training and flying small target detecting, in which the flying small target detecting stage is composed by three process parts: candidate patches identification, local foreground–background image separation and threshold segmentation over reconstructed sparse images. The comprehensive scheme of the proposed approach is depicted in Figure 3.
Correspondingly, the synthetic integrated detection approach to flying small target in image sequence can be presented based on all above discussed modules, and outlined as Algorithm 1.
Algorithm 1 Approach to flying small target detection based on GMM_CS.
Input: Flying small target image sequence
Output: Detection results of each image in sequence
  1:
Initialize N t r a i n , N d e t e c t , s, m, ξ , α , β , R, K, ϵ
  2:
for i = 1 to N t r a i n do
  3:
 Break image into patches by Equation (11) and compressive sensing by Equation (3)
  4:
 Make GMM_CS for each patch by Equations (7)–(10)
  5:
end for
  6:
for i = 1 to N d e t e c t do
  7:
 Break image into patches by Equation (11) and compressive sensing by Equation (3)
  8:
 Identify candidate patches by Equations (7), (12), (13)
  9:
for j = 1 to N p a t c h do
 10:
  Designate successive patches and compose a data matrix G
 11:
  repeat
 12:
   Set k = 0 , L ^ 0 = 0 , S ^ 0 = 0 , Ω ^ L = { } , Ω ^ S = { } , P 0 = G
 13:
   Update indicator k = k + 1
 14:
   Identify supports: Ω L = arg max Ω O { F Ω ( P k 1 ) F : | Ω | 2 R }
 15:
   Merging: Ω ˜ L = Ω ^ L Ω L
 16:
   Estimating: L ˜ k = arg min L { G L S ^ k 1 F : L span ( Ω ˜ L ) }
 17:
   Pruning: Ω ^ L = arg max Ω O { F Ω ( L ˜ k ) F : | Ω | R } , L ^ k = F Ω ^ L ( L ˜ k )
 18:
   Update residue: P ˜ k = G L ^ k S ^ k 1
 19:
   Identify supports: Ω S = arg max Ω Q { H Ω · P ˜ k F : | Ω | 2 K }
 20:
   Merging: Ω ˜ S = Ω ^ S Ω S
 21:
   Estimating: S ˜ k = arg min S { G L ^ k S F : S = H Ω · P ˜ k , Ω Ω ˜ S }
 22:
   Pruning: Ω ^ S = arg max Ω Q { H Ω · S ˜ k F : | Ω | K } , S ^ k = H Ω ^ S · S ˜ k
 23:
   Update residue: P k = G L ^ k S ^ k
 24:
  until G L ^ k S ^ k F / G F ϵ
 25:
end for
 26:
 Detect target over reconstructed sparse images by threshold segmentation
 27:
end for

4. Experimental Results and Analysis

4.1. Experiment Setting and Evaluation Metrics

In order to validate the proposed detection method of flying small target, two experiments are conducted on visible and infrared image sequences of flying UAVs derived from two challenging datasets. The image sequence for the first experiment has a 189-frame visible image with resolution 640 × 480 pixels acquired by a CCD visible light camera, selected from the video named ‘video_37.avi’ in the UAV dataset which was collected by Rozantsev et al. [11]. The video with 14 s and 25 frames per seconds is the only one drone video captured from a stationary camera. Due to some discontinuous frames being without a ground-truth, we choose the labeled 189-frame visible image of this video for the experiment. The image sequence for the second experiment has a 337-frame infrared image with resolution 640 × 480 pixels acquired by a longwave FLIR A655sc camera, selected from the sequence named ‘quadrocopter2’ in the LTIR dataset which was used in the Visual Object Tracking (VOT) challenge [44]. The image sequence is captured from a stationary camera. Due to the moving foreground of flying target not being salient with no obvious changes, we choose 337 frames from the original 1010 frames by a three-frame interval for the experiment.
Additionally, the true positives (TP), false positives (FP) and false negatives (FN) of flying UAVs are counted, and then two most important evaluation metrics: the recall and precision are calculated to quantitatively evaluate the performance of the proposed flying small target detection method, and they are defined as follows:
R e c a l l = T P T P + F N
P r e c i s i o n = T P T P + F P
In the evaluation, we regarded the overlap score between detected bounding boxes r d and manually labeled ground-truth bounding boxes r t that are greater than a certain threshold as correct detections and boxes that less than a certain threshold as misdetection, and the overlap score is defined as
S c o r e = | r d r t | | r d r t |
where ∩ and ∪ represent the intersection and union of two regions, respectively. | · | denotes the number of pixels in the region.
All the experiments are carried out on a general computer with 8-GB memory and 2.2-GHz Intel Core i5-5200U processor.

4.2. Performance Evaluation and Comparative Analysis

To test the effectiveness and practicality of the proposed method, its performance is evaluated by the aforementioned metrics and compared with three baseline methods. One is a well-studied background subtraction method based on an improved adaptive Gaussian mixture model (GMMv1) [45], another is a superior foreground and background separation method named Grassmannian robust adaptive subspace tracking algorithm (GRASTA) [46]. the third one is a state-of-the-art moving target detection method by detecting contiguous outliers in the low-rank representation (DECOLOR) [47].
The GMMv1 method is performed in a background subtraction library named BGSLibrary developed by Sobral [48], and its parameter setting in the library is adopted. The GRASTA and DECOLOR methods are implemented as open source and provided in this paper, and 20 frames visible or infrared image are introduced to train the initial subspace of GRASTA method, while the parameters of DECOLOR method employ the default values provided by the author.
The proposed method is implemented and executed on a Matlab R2012a software platform. In the following experiments, the number of Gaussian distribution is s = 3 . Correspondingly, the learning rate of weight is α = 0.2 and the learning rate of mean and variance is β = 0.2 for each Gaussian distribution. In the procedure of CS measurement, the compressive ratio is m = 0.5 n . Moreover, the rank of the expected low-rank matrix R is set to 1, and the sparseness of the expected sparse matrix K is set to 0.1 times the number of pixels in one image patch. For identifying a image patch with a flying small targets, the threshold of δ is set to 0.25 m and 0.75 m for the visible and infrared image sequences, respectively. In addition, 20 frames of visible or infrared images are introduced to train the proposed GMM_CS, and all images are broken into non-overlapping patches with a resolution of 80 × 60 and 64 × 48 pixels for the visible and infrared image sequences, respectively.
In the first experiment, a visible image sequences of a flying drone is tested to detect using three baseline methods and the proposed method. Due to the changing flight attitude, near or far distance and variable lighting conditions, the shape and size of the drone is in flux and is challenging for detection. As an intuitively illustration, the flying small target detection results of four in-between frames are given as shown in Figure 4. The yellow number marked in the lower left quarter of image is the serial number of the image frame, and the target region of groundtruth, detected region by GMMv1, GRASTA, DECOLOR and the proposed method in this study are marked by a red bounding box. These results show that the proposed method can effectively detect the actual flying small target and generate less false alarm targets relative to the baseline methods.
In order to objectively quantitate the evaluation, the recall and precision are statistically calculated over the whole visual small target sequences. The thresholds of overlap score for target success detection affirmation are choose from 0.05 to 1 with step 0.05, and the result curves are plotted in Figure 5. It can be observed visually that the higher threshold of overlap score leads to lower recall and precision. Note that the proposed method outperforms the baseline methods in precision evaluation while they have approximately equal recall evaluation.
In the second experiment, a longwave infrared image sequences of flying quadrocopter is tested to detect by three baseline methods and the proposed method. Due to the sequence coming from a surveillance scenario with long imaging distance, the flying quadrocopter presents not only small size but also dim intensity even immersing in complex background. so it is rarely perfectly visible by the naked eye, not to mention automatic detection. For clear performance demonstration, the flying small target detection results in four in-between frames that are given as shown in Figure 6. The yellow number marked in the lower left quarter of the image is the serial number of the image frame, and the target region of groundtruth, detected region by GMMv1, GRASTA, DECOLOR and the proposed method in this study are marked by a red bounding box. We can find that the actual flying small target still can be detected in the challenging context in spite of being about some false alarm targets. Especially, the detection performance of the proposed method is not inferior to any baseline methods.
Consequently, the recall and precision are statistically calculated over the whole infrared small target sequences. The thresholds of overlap score for target success detection affirmation are choose from 0.05 to 1 with step 0.05, and the resultant curves are plotted in Figure 7. It is easy to observed that the proposed method has better recall and precision evaluation when the thresholds of overlap score is less than 0.45, and has similar recall and precision evaluation when the thresholds of overlap score is more than 0.45.

5. Conclusions

In this paper, a new Gaussian mixture model in a compressive sensing domain is designed, and then a flying small target detection method for anti-UAV is proposed. The method not only makes use of the advantages of compressive sensing to reduce data quantity of background modeling, but also achieves flying small target detection on a handful of candidate patches by low-rank and sparse matrix decomposition. The results of experiments show that the proposed method can effectively detect flying small targets with less false alarms, and has outstanding recall and precision performance on both visible and infrared image sequences. In future studies, we will aim to focus on distributed deployment of multiple visual sensors for barrier coverage of surveillance, as well as multi-view stereo vision for flying small target detection by multi-sensor cooperation.

Author Contributions

C.W. proposed the idea of this paper and conducted the theoretical analysis and experimental work. T.W. proposed many useful suggestions in the process of theoretical analysis. E.W. marked groundtruth images. E.S. and Z.L. took part in experiments. C.W. wrote the paper, and T.W. revised it.

Funding

This research was supported by the National Natural Science Foundation of China (Grant Nos. 61703287, 61503017 and 61571309), the Liaoning Provincial Natural Science Foundation of China (Grant No. 20180550664), the Scientific Research Program of Liaoning Provincial Education Department of China (Grant Nos. L201726 and L201702), the Science and Technology Plan Projects of Shenyang City of China (Grant No. 18-013-0-24) and the Doctoral Scientific Research Foundation of Shenyang Aerospace University (Grant No. 17YB16).

Acknowledgments

The authors would like to thank the reviewers and editors for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, T.; Qin, R.; Chen, Y.; Snoussi, H.; Choi, C. A reinforcement learning approach for UAV target searching and tracking. Multimedia Tools Appl. 2019, 78, 4347–4364. [Google Scholar] [CrossRef]
  2. Ganti, S.R.; Kim, Y. Implementation of detection and tracking mechanism for small UAS. In Proceedings of the 2016 International Conference on Unmanned Aircraft Systems (ICUAS), Arlington, VA, USA, 7–10 June 2016; pp. 1254–1260. [Google Scholar]
  3. Mian, A.S. Realtime visual tracking of aircrafts. In Proceedings of the Digital Image Computing: Techniques and Applications (DICTA’08), Canberra, ACT, Australia, 1–3 December 2008; pp. 351–356. [Google Scholar]
  4. Farlik, J.; Kratky, M.; Casar, J.; Stary, V. Radar cross section and detection of small unmanned aerial vehicles. In Proceedings of the 2016 17th International Conference on Mechatronics-Mechatronika (ME), Prague, Czech Republic, 7–9 December 2016; pp. 1–7. [Google Scholar]
  5. Hoffmann, F.; Ritchie, M.; Fioranelli, F.; Charlish, A.; Griffiths, H. Micro-Doppler based detection and tracking of UAVs with multistatic radar. In Proceedings of the 2016 IEEE Radar Conference (RadarConf), Philadelphia, PA, USA, 2–6 May 2016; pp. 1–6. [Google Scholar]
  6. Tapsall, T.B. Using crowd sourcing to combat potentially illegal or dangerous UAV operations. In Unmanned/ Unattended Sensors and Sensor Networks XII; International Society for Optics and Photonics: Edinburgh, UK, 2016; Volume 9986, p. 998605. [Google Scholar]
  7. Shi, X.; Yang, C.; Xie, W.; Chao, L.; Shi, Z.; Chen, J. Anti-Drone System with Multiple Surveillance Technologies: Architecture, Implementation, and Challenges. IEEE Commun. Mag. 2018, 56, 68–74. [Google Scholar] [CrossRef]
  8. Sosnowski, T.; Bieszczad, G.; Madura, H.; Kastek, M. Thermovision system for flying objects detection. In Proceedings of the 2018 Baltic URSI Symposium (URSI), Poznan, Poland, 15–17 May 2018; pp. 141–144. [Google Scholar]
  9. Hwang, K.; Har, D. Identifying Small Drones Using Infrared Polarization Filters. TECHART J. Arts Imaging Sci. 2016, 3, 50–54. [Google Scholar] [CrossRef]
  10. Li, J.; Ye, D.H.; Chung, T.; Kolsch, M.; Wachs, J.; Bouman, C. Multi-target detection and tracking from a single camera in Unmanned Aerial Vehicles (UAVs). In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Daejeon, South Korea, 9–14 October 2016; pp. 4992–4997. [Google Scholar]
  11. Rozantsev, A.; Lepetit, V.; Fua, P. Flying objects detection from a single moving camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR2015), Boston, MA, USA, 7–12 June 2015; pp. 4128–4136. [Google Scholar]
  12. Rozantsev, A.; Lepetit, V.; Fua, P. Detecting Flying Objects using a Single Moving Camera. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 879–892. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, Z.; Cao, Y.; Ding, M.; Zhuang, L.; Wang, Z. Spatial and Temporal Context Information Fusion Based Flying Objects Detection for Autonomous Sense and Avoid. In Proceedings of the 2018 International Conference on Unmanned Aircraft Systems (ICUAS), Dallas, TX, USA, 12–15 June 2018; pp. 569–578. [Google Scholar]
  14. Du, L.; Gao, C.; Feng, Q.; Wang, C.; Liu, J. Small UAV Detection in Videos from a Single Moving Camera. In Proceedings of the CCF Chinese Conference on Computer Vision, Tianjin, China, 11–14 October 2017; Springer: Singapore, 2017; pp. 187–197. [Google Scholar]
  15. Wu, P.F.; Xiao, F.; Sha, C.; Huang, H.P.; Wang, R.C.; Xiong, N.X. Node Scheduling Strategies for Achieving Full-View Area Coverage in Camera Sensor Networks. Sensors 2017, 17, 1303. [Google Scholar] [CrossRef]
  16. Han, R.; Yang, W.; Zhang, L. Achieving Crossed Strong Barrier Coverage in Wireless Sensor Network. Sensors 2018, 18, 534. [Google Scholar] [CrossRef] [PubMed]
  17. Rossi, P.S.; Ciuonzo, D.; Kansanen, K.; Ekman, T. On energy detection for MIMO decision fusion in wireless sensor networks over NLOS fading. IEEE Commun. Lett. 2015, 19, 303–306. [Google Scholar] [CrossRef]
  18. Rossi, P.S.; Ciuonzo, D.; Kansanen, K.; Ekman, T. Performance analysis of energy detection for MIMO decision fusion in wireless sensor networks over arbitrary fading channels. IEEE Trans. Wirel. Commun. 2016, 15, 7794–7806. [Google Scholar] [CrossRef]
  19. Rossi, P.S.; Ciuonzo, D.; Ekman, T. HMM-based decision fusion in wireless sensor networks with noncoherent multiple access. IEEE Commun. Lett. 2015, 19, 871–874. [Google Scholar] [CrossRef]
  20. Stein, D.W. Detection of random signals in Gaussian mixture noise. IEEE Trans. Inf. Theory 1995, 41, 1788–1801. [Google Scholar] [CrossRef]
  21. Zhuang, X.; Huang, J.; Potamianos, G.; Hasegawa-Johnson, M. Acoustic fall detection using Gaussian mixture models and GMM supervectors. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 69–72. [Google Scholar]
  22. Wang, T.; Qiao, M.; Chen, J.; Wang, C.; Zhang, W.; Snoussi, H. Abnormal global and local event detection in compressive sensing domain. AIP Adv. 2018, 8, 055224. [Google Scholar] [CrossRef]
  23. Stauffer, C.; Grimson, W.E.L. Adaptive Background Mixture Models for Real-Time Tracking. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR1999), Fort Collins, CO, USA, 23–25 June 1999; Volume 2, pp. 246–252. [Google Scholar]
  24. Kozick, R.J.; Sadler, B.M. Maximum-likelihood array processing in non-Gaussian noise with Gaussian mixtures. IEEE Trans. Signal Process. 2000, 48, 3520–3535. [Google Scholar]
  25. Cevher, V.; Sankaranarayanan, A.; Duarte, M.F.; Reddy, D.; Baraniuk, R.G.; Chellappa, R. Compressive Sensing for Background Subtraction; Springer: Berlin/Heidelberg, Germany, 2008; pp. 155–168. [Google Scholar]
  26. Fornasier, M.; Rauhut, H. Compressive sensing. In Handbook of Mathematical Methods in Imaging; Springer: New York, NY, USA, 2011; pp. 187–228. [Google Scholar]
  27. Liu, Y.; Pados, D.A. Compressed-Sensed-Domain L1-PCA Video Surveillance. IEEE Trans. Multimedia 2016, 18, 351–363. [Google Scholar] [CrossRef]
  28. Lu, C.; Shi, J.; Jia, J. Online Robust Dictionary Learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR2013), Portland, OR, USA, 23–28 June 2013; pp. 415–422. [Google Scholar]
  29. Baraniuk, R.G. A lecture on compressive sensing. IEEE Signal Process. Mag. 2007, 24, 118–121. [Google Scholar] [CrossRef]
  30. Candès, E.J. The restricted isometry property and its implications for compressed sensing. C. R. Math. 2008, 346, 589–592. [Google Scholar] [CrossRef]
  31. Candès, E.J.; Tao, T. Decoding by linear programming. IEEE Trans. Inf. Theory 2005, 51, 4203–4215. [Google Scholar] [CrossRef] [Green Version]
  32. Li, L.; Li, H.; Li, T.; Gao, F. Infrared small target detection in compressive domain. Electron. Lett. 2014, 50, 510–512. [Google Scholar] [CrossRef]
  33. Wang, C.; Qin, S. Background Modeling of Infrared Image in Dynamic Scene with Gaussian Mixture Model in Compressed Sensing Domain. Acta Autom. Sin. 2018, 44, 1212–1226. [Google Scholar]
  34. Wang, T.; Qiao, M.; Lin, Z.; Li, C.; Snoussi, H.; Liu, Z.; Choi, C. Generative Neural Networks for Anomaly Detection in Crowded Scenes. IEEE Trans. Inf. Forensics Secur. 2019, 14, 1390–1399. [Google Scholar] [CrossRef]
  35. Gan, L. Block Compressed Sensing of Natural Images. In Proceedings of the International Conference on Digital Signal Processing, Cardiff, UK, 1–4 July 2007; pp. 403–406. [Google Scholar]
  36. Cho, T.S.; Butman, M.; Avidan, S.; Freeman, W.T. The patch transform and its applications to image editing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR2008), Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  37. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 1–39. [Google Scholar] [CrossRef]
  38. Chen, C.; Li, S.; Qin, H.; Hao, A. Robust salient motion detection in non-stationary videos via novel integrated strategies of spatio-temporal coherency clues and low-rank analysis. Pattern Recognit. 2016, 52, 410–432. [Google Scholar] [CrossRef]
  39. Wright, J.; Ganesh, A.; Rao, S.; Ma, Y. Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Matrices via Convex Optimization. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 7–10 December 2009; pp. 2080–2088. [Google Scholar]
  40. Wang, C.; Qin, S. Approach for moving small target detection in infrared image sequence based on reinforcement learning. J. Electron. Imaging 2016, 25, 053032. [Google Scholar] [CrossRef]
  41. Waters, A.E.; Sankaranarayanan, A.C.; Baraniuk, R. SpaRCS: Recovering low-rank and sparse matrices from compressive measurements. In Proceedings of the Advances in Neural Information Processing Systems, Granada, Spain, 12–14 December 2011; pp. 1089–1097. [Google Scholar]
  42. Lee, K.; Bresler, Y. ADMiRA: Atomic decomposition for minimum rank approximation. IEEE Trans. Inf. Theory 2010, 56, 4402–4416. [Google Scholar] [CrossRef]
  43. Needell, D.; Tropp, J.A. CoSaMP: Iterative signal recovery from incomplete and inaccurate samples. Appl. Comput. Harmon. Anal. 2009, 26, 301–321. [Google Scholar] [CrossRef] [Green Version]
  44. Berg, A.; Ahlberg, J.; Felsberg, M. A thermal Object Tracking benchmark. In Proceedings of the IEEE International Conference on Advanced Video and Signal Based Surveillance, Karlsruhe, Germany, 25–28 August 2015; pp. 1–6. [Google Scholar]
  45. Kaewtrakulpong, P.; Bowden, R. An Improved Adaptive Background Mixture Model for Real-time Tracking with Shadow Detection. In Video Based Surveillance Systems; Remagnino, P., Jones, G.A., Paragios, N., Regazzoni, C.S., Eds.; Springer: Boston, MA, USA, 2002; pp. 135–144. [Google Scholar] [Green Version]
  46. He, J.; Balzano, L.; Szlam, A. Incremental gradient on the Grassmannian for online foreground and background separation in subsampled video. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR2012), Providence, RI, USA, 16–21 June 2012; pp. 1568–1575. [Google Scholar]
  47. Zhou, X.; Yang, C.; Yu, W. Moving Object Detection by Detecting Contiguous Outliers in the Low-Rank Representation. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 597–610. [Google Scholar] [CrossRef] [Green Version]
  48. Sobral, A. BGSLibrary: An opencv c++ background subtraction library. In Proceedings of the IX Workshop De Visão Computacional, Botafogo, Rio de Janeiro, ‎Brazil, 3–5 June 2013; pp. 1–6. [Google Scholar]
Figure 1. A demonstration of candidate patches identification based on GMM_CS.
Figure 1. A demonstration of candidate patches identification based on GMM_CS.
Sensors 19 02168 g001
Figure 2. Local foreground-background images separation via RPCA.
Figure 2. Local foreground-background images separation via RPCA.
Sensors 19 02168 g002
Figure 3. Comprehensive scheme for flying small target detection.
Figure 3. Comprehensive scheme for flying small target detection.
Sensors 19 02168 g003
Figure 4. Some experiment results of flying small target detection on visible image sequence obtained from different algorithms (a) groundtruth, (b) GMMv1, (c) GRASTA, (d) DECOLOR and (e) the proposed method.
Figure 4. Some experiment results of flying small target detection on visible image sequence obtained from different algorithms (a) groundtruth, (b) GMMv1, (c) GRASTA, (d) DECOLOR and (e) the proposed method.
Sensors 19 02168 g004
Figure 5. Curves of recall and precision with different threshold of overlap scores for flying small target detection on visible image sequence (a) recall and (b) precision.
Figure 5. Curves of recall and precision with different threshold of overlap scores for flying small target detection on visible image sequence (a) recall and (b) precision.
Sensors 19 02168 g005
Figure 6. Some experiment results of flying small target detection on infrared image sequence obtained from different algorithms (a) groundtruth, (b) GMMv1, (c) GRASTA, (d) DECOLOR and (e) the proposed method.
Figure 6. Some experiment results of flying small target detection on infrared image sequence obtained from different algorithms (a) groundtruth, (b) GMMv1, (c) GRASTA, (d) DECOLOR and (e) the proposed method.
Sensors 19 02168 g006
Figure 7. Curves of recall and precision with different threshold of overlap scores for flying small target detection on infrared image sequence (a) recall and (b) precision.
Figure 7. Curves of recall and precision with different threshold of overlap scores for flying small target detection on infrared image sequence (a) recall and (b) precision.
Sensors 19 02168 g007
Table 1. A brief summary of the notations used throughout the rest of this paper.
Table 1. A brief summary of the notations used throughout the rest of this paper.
NotationsDescriptions
sthe number of Gaussian distribution
w i the weight of the i t h Gaussian distribution
μ i the mean of the i t h Gaussian distribution
σ i 2 the variance of the i t h Gaussian distribution
x a vector stacked by an image with n pixels
y a compressive sensing measurement vector with m dimensions
Ψ a sparse basis matrix composed by q n-dim basis vectors
Φ a measurement matrix composed by m n-dim Gaussian distribution
Θ the weighting coefficients vector of x in the Ψ domain
ξ a constant for Gaussian distribution identification
α the learning rate of weight of the Gaussian distribution
β the learning rate of mean and variance of the Gaussian distribution
τ the value of the Gaussian distribution matched or not
J k the value of the k t h linear measurement matched or not
δ a threshold for identifying image patch maintain target or not
D a matrix composed by r rows and c columns block matrices
Ga matrix composed by a group of vectorized consecutive patches
Lthe low-rank matrix decomposed from matrix G via RPCA
Sthe sparse matrix decomposed from matrix G via RPCA
Rthe rank of the low-rank matrix L
Kthe sparse degree of the sparse matrix S
λ a weight parameter for tradeoff the matrices L and S
u i the i t h column vector of matrix U after SVD
v i the i t h column vector of matrix V after SVD
Ω L a subset of the entire set of singular vectors O
Ω S a subset of the entire set of position support terms Q

Share and Cite

MDPI and ACS Style

Wang, C.; Wang, T.; Wang, E.; Sun, E.; Luo, Z. Flying Small Target Detection for Anti-UAV Based on a Gaussian Mixture Model in a Compressive Sensing Domain. Sensors 2019, 19, 2168. https://doi.org/10.3390/s19092168

AMA Style

Wang C, Wang T, Wang E, Sun E, Luo Z. Flying Small Target Detection for Anti-UAV Based on a Gaussian Mixture Model in a Compressive Sensing Domain. Sensors. 2019; 19(9):2168. https://doi.org/10.3390/s19092168

Chicago/Turabian Style

Wang, Chuanyun, Tian Wang, Ershen Wang, Enyan Sun, and Zhen Luo. 2019. "Flying Small Target Detection for Anti-UAV Based on a Gaussian Mixture Model in a Compressive Sensing Domain" Sensors 19, no. 9: 2168. https://doi.org/10.3390/s19092168

APA Style

Wang, C., Wang, T., Wang, E., Sun, E., & Luo, Z. (2019). Flying Small Target Detection for Anti-UAV Based on a Gaussian Mixture Model in a Compressive Sensing Domain. Sensors, 19(9), 2168. https://doi.org/10.3390/s19092168

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop