Next Article in Journal
Horizontal Visibility in Forests
Next Article in Special Issue
A Residual Attention and Local Context-Aware Network for Road Extraction from High-Resolution Remote Sensing Imagery
Previous Article in Journal
Field-Based High-Quality Emissivity Spectra Measurement Using a Fourier Transform Thermal Infrared Hyperspectral Imager
Previous Article in Special Issue
Building Extraction from Remote Sensing Images with Sparse Token Transformers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hyperspectral Target Detection with an Auxiliary Generative Adversarial Network

School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710129, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2021, 13(21), 4454; https://doi.org/10.3390/rs13214454
Submission received: 10 October 2021 / Revised: 29 October 2021 / Accepted: 2 November 2021 / Published: 5 November 2021

Abstract

:
In recent years, the deep neural network has shown a strong presence in classification tasks and its effectiveness has been well proved. However, the framework of DNN usually requires a large number of samples. Compared to the training sets in classification tasks, the training sets for the target detection of hyperspectral images may only include a few target spectra which are quite limited and precious. The insufficient labeled samples make the DNN-based hyperspectral target detection task a challenging problem. To address this problem, we propose a hyperspectral target detection approach with an auxiliary generative adversarial network. Specifically, the training set is first expanded by generating simulated target spectra and background spectra using the generative adversarial network. Then, a classifier which is highly associated with the discriminator of the generative adversarial network is trained based on the real and the generated spectra. Finally, in order to further suppress the background, guided filters are utilized to improve the smoothness and robustness of the detection results. Experiments conducted on real hyperspectral images show the proposed approach is able to perform more efficiently and accurately compared to other target detection approaches.

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) obtained by remote sensing systems usually contain hundreds of continuous narrow bands over a wide range of an electromagnetic spectrum. Each pixel of HSIs forms a continuous spectrum, which is able to provide abundant useful spectral information that is beneficial to distinguish different objects [1]. By exploiting the different electromagnetic spectra of materials, targets of interest in the hyperspectral scene can be detected and identified through effective processing algorithms.
As one of the most important research areas, target detection (TD) is basically a binary classifier which aims at separating target pixels from the background with known target spectra. Furthermore, just as fields of synthetic aperture radar (SAR) target recognition [2], electromagnetic sources recognition [3] and single-image super-resolution [4], which are very likely to be confronted with the problem of insufficient training samples, TD tasks also face the challenge of a small size of labeled training samples. Because of the small fraction of pixels being labeled as targets, the method of statistical hypothesis testing is usually exploited in classical target detection algorithms. Representative algorithms include the linear spectral matched filter (SMF), matched subspace detector (MSD) and adaptive coherence/cosine estimator (ACE). These algorithms define the target spectral characteristics by either a single target spectrum or a target subspace, while modeling the background statistically by a Gaussian distribution or a subspace representing either the whole or the local background statistics. Through the generalized likelihood ratio test (GLRT), the output for a certain pixel is obtained [1]. As kernel methods in machine learning are introduced to target detection, the approaches mentioned above are further extended to nonlinear versions, such as kernel SMF and kernel MSD [5], which can solve nonlinear problems to some extent. However, how to design a suitable kernel remains a problem. Constrained energy minimization (CEM), as a linear filter, is another target detection approach which constrains a desired target signature while being able to minimize the total energy of the output of other known signatures [6].
Meanwhile, target detection algorithms based on sparse representation are also extensively explored by researchers. Central to sparse representation is the idea that a signal can be expressed as a linear combination of very few atoms of an over-complete dictionary consisting of a set of training samples from all classes. Then, the class information can be revealed by the sparse coefficients under the assumption that signals from different classes lie in different spaces [7]. Typical approaches include the original sparse representation-based target detector (STD) and its variants: sparse representation-based binary hypothesis detector (SRBBH) [8] and sparse and dense hybrid representation-based target detector (SDRD) [9]. On the basis of STD, SRBBH perceives the detection problem as a competition between the background-only subspace and the target-and-background subspace under two different hypotheses. However, different from SRBBH, SDRD emphasizes that the background and the target sub-dictionaries collaboratively compete with each other, which facilitates the learning of sparse and dense hybrid representations for the test pixels.
Although progress has been achieved and many problems in detection tasks have been solved, challenges are still there within the realm of hyperspectral target detection. First, due to sensor noise and other factors, the target spectrum of the same material may vary dramatically, which makes it hard to represent the target spectrum using only a single spectrum or a subspace. Further, since spatial resolutions of HSIs are usually lower than that of multi-spectral and chromatic images, mixing pixels covering several materials still exist, which makes detection tasks much more complicated. Moreover, the problem of target detection is mostly nonlinear, but the approaches mentioned earlier are more likely to make decisions under linear assumptions, which does not do justice to the complex reality of HSI processing. Furthermore, it is noteworthy that many of these approaches just fail to work well when nonlinearity increases.
Recently, the deep neural network (DNN), which has achieved great success in the field of computer vision, has attracted the attention of the HSI processing community. However, due to the small training set with very few target spectra in the task of hyperspectral detection, DNNs designed for large datasets can hardly be utilized directly. In order to utilize the great representation power of a DNN, the training strategy, as well as the network structure and function must be carefully designed.
Furthermore, based on the whole new avenue opened up by DNNs, approaches of [10,11,12,13,14] have been proposed accordingly, in which the neural network has been exploited as an unsupervised feature extractor. Specifically, in [10], a distance-constrained stacked sparse autoencoder is utilized for feature extraction, which is able to maximize the distinction between the target pixels and the other background pixels in the feature space, and the process is followed by background suppression via a simple detector based on a radial basis function kernel. The authors of Ref. [11] introduce two adversarial losses in the autoencoder to reconstruct enhanced hyperspectral images and utilize the RX detector for anomaly detection. Meanwhile, the autoencoder also proves effective to be exploited to extract spectral features [12] which are further processed with band selection and spatial post-processing techniques for background suppression. Similar processing techniques can be found in deep latent spectral representation learning proposed in [13]. Another approach that is comparable is 3-D macro–micro residual autoencoder (3-D-MMRAE) in which a convolutional autoencoder is built to extract the spectral and spatial features from both the macro and the micro branches, and the extracted features are further fed to a hierarchical radial basis function (hRBF) detector for target preservation and background suppression [14]. The results have demonstrated the effectiveness of these approaches in HSI processing, but the underlying mechanism of the transformation process using an autoencoder remains to be uncovered.
Besides unsupervised approaches, supervised approaches have also garnered much attention. Furthermore, among such approaches, the subtraction pixel pair feature (SPPF), which tries to enlarge the training set by using pairs between target spectra and background spectra [15], eclipses many others as a very effective approach. To be specific, the approach acquires a sufficiently large number of samples using SPPF and ensures the excellent performance of the multi-layer convolutional neural network. Perhaps more effective is the approach of [16], which combines the representation-based method together with the linear mixing model to generate adequate background and target samples to train the neural network. Whilst being compared to [15], the subtraction process in [16] is performed in the final layer rather than the input layer, which improves the detection performance.
Apart from the approaches mentioned above, efforts performed from other angles which prove fruitful are [17,18,19]. The authors of Ref. [17] utilize a semi-supervised learning method through a two-step training process, which unfolds with the unsupervised pre-training of a generative adversarial network, followed by the fine-tuning of a discriminator through limited labeled samples. The authors of Ref. [18] realize few shot learning based on domain adaptive approaches. To be specific, a network including both feature fusion and channel attention is built for feature extraction and, then, the learned features from different sensors are correlated through a discriminator whereby the weights learned in the source domain are able to be transferred to the target domain for the following few shot learning. The authors of Ref. [19] approach the detection problem based on multiple instance learning, in which a sparsity constraint is introduced for the estimated attention factor of the positive training bags and ensures the approach yields superior performance.
In this paper, we propose a novel hyperspectral target detection approach with an auxiliary generative adversarial network. It is a neural network which can be trained with very limited target spectra directly without using any other preprocessing approaches. Similar to the general generative adversarial network [20], the proposed network mainly contains two parts: a generator and a discriminator. The generator generates simulated target spectra and background spectra, which ease the problem of insufficient training samples. Then, the initial detection results can be acquired by a classifier which is highly correlated with the discriminator. Since the neural network only deals with the spectral information of pixels, post-processing techniques are added to make full use of the spatial information. Furthermore, in order to preserve the edges of the targets, several guided filters [21] are utilized which can further suppress the background and improve the smoothness as well as the robustness of the detection results.
Extensive experiments have been conducted on real HSIs and the results show the proposed approach is more effective compared to the conventional target detection approaches. The contributions of the proposed approach are summarized as follows:
(1)
A novel hyperspectral target detection framework based on the generative adversarial network is proposed, which can exhibit good performance while being trained with limited target spectra.
(2)
During the training process, a generator capable of generating simulated target spectra is obtained. Furthermore, the generated targets as well as the background spectra allow us to better understand the real distributions of the target and the background.
(3)
A combination of guided filters is adopted to further suppress the background, which ensures the smoothness and robustness of the results and makes the processing of various sizes of targets possible.
The remainder of this article is organized as follows: Section 2 introduces the relevant theories and the proposed approach based on the generative adversarial network in detail. Section 3 shows the analysis process and the experimental results. Section 4 puts forward a brief discussion on similar approaches and sheds light upon the future work. Section 5 draws the conclusion.

2. Materials and Methods

Before introducing the proposed approach, we first explicate the theories related to generative adversarial network and guided filter, which were used for spectral target detection and spatial smoothing in the proposed approach, respectively. In the following section of this paper, we examined the circumstantial process.

2.1. Generative Adversarial Network

Generative adversarial network (GAN) was first proposed and used for estimating generative models in [20]. To achieve good performance, GAN utilizes two models: a generator and a discriminator, both of which can be realized by neural networks.
The basic idea of GAN is that the generator tries to generate fake samples and fool the discriminator, while the discriminator tries to distinguish between real samples coming from the data and fake samples coming from the generator. This adversarial competition drives both the generator and the discriminator to improve their performance until the generated or fake samples cannot be distinguished from the real ones. This process can be formulated in a more mathematical manner.
Suppose p d a t a ( x ) , p z ( z ) denote data distribution and a random noise distribution, respectively. The generator builds a mapping p z ( z ) : G ( z ; θ g ) p g ( x ) , where p g ( x ) denotes the distribution of generated samples. As for the discriminator, let D ( x ; θ d ) denote the probability of x coming from p d a t a ( x ) rather than p g ( x ) . Then, the objective function of the generator and the discriminator, denoted as L g and L d , can be formulated as Equations (1) and (2), respectively.
min G L g = E z p z ( z ) [ log ( 1 D ( G ( z ) ) ) ]
max D L d = E x p d a t a ( x ) [ log D ( x ) ] + E z p z ( z ) [ log ( 1 D ( G ( z ) ) ) ]
Thus, during alternate iteration of the training process, the final value function V ( D , G ) can be formulated as Equation (3).
min G max D V ( D , G ) = E x p d a t a ( x ) [ log D ( x ) ] + E z p z ( z ) [ log ( 1 D ( G ( z ) ) ) ]
The value function in Equation (3) can be further interpreted as a distance metric, i.e., Jensen–Shannon divergence. In practice, however, vanilla GAN proposed in [20] is usually hard to train in deep networks. Researchers holding different viewpoints have been working hard to propose various variants which can be optimized more easily, including the activation function, objective function and network architectures, etc.
For example, in [22], several strategies for training stable deep convolutional GANs are suggested, including adopting batchnorm, using ReLU activation in the generator other than the output layer and using LeakyReLU activation in the discriminator. On the other hand, through the variational divergence estimation network, the training objective of GAN is generalized to arbitrary f-divergences in [23]. Moreover, in order to accommodate different application scenarios, modified network architectures are introduced, such as information maximizing GAN (InfoGAN) [24] and auxiliary classifier GAN (ACGAN) [25].
Generally speaking, generative adversarial network is mainly utilized either to model the distribution of samples or to generate samples. Although there exist several criteria for evaluating the quality of generated samples, open issues still remain in the particular fields.
In the proposed approach, we exploit the core idea of GAN and try to detect targets through discrimination. To be more concrete, a detection network is built with an auxiliary GAN which provides supplementary samples and shares several layers with the detector. Furthermore, to fix the problem of the detection network in being only able to deal with spectral information, a spatial post-processing technique was employed in the training process.

2.2. Guided Filter

Guided image filtering is one kind of self-adaptive image filters. It computes the filtering process according to the content of a guided image which can be the input image itself or another image. To put it more specifically, the guided filter is assumed to be a local linear model between the guided image I and the filtering output q. Suppose q is a linear transformation of I in a local area w k centered at the pixel k [21]:
q i = a k I i + b k , i w k
where a k and b k denote linear coefficients which are assumed to be constant in w k . On the other hand, the output q is modeled as the input p which eliminates a few other parts n, defined as:
q i = p i n i
When Equations (4) and (5) are combined together, the previous tough issue is transformed into a minimization problem with a quadric formula:
E ( a k , b k ) = i w k [ ( a k I i + b k ) p i ] 2 + ϵ a k 2
where ϵ is a regularization parameter which penalizes large a k .
As an edge-preserving smooth operator, guided filter is faster and works better near edges when compared to the popular bilateral filters. Therefore, we adopted guided filter as a post-processing technique in the detection algorithm to further suppress the background information and the tiny isolated points in the previous detection maps.

2.3. Proposed Approach

In this section, we explain the proposed detection approach in detail. Firstly, a general introduction about the detection framework with circumstantial explanations was given. After that, the architectures of the network along with the training objective were introduced. Finally, a bank of guided filters used for spatial smoothing was illustrated briefly.

2.3.1. General Framework

Given the fact that labeled target spectra are quite limited, unlabeled spectra including targets and backgrounds should, therefore, be utilized much more efficiently. To this end, our approach utilized the generative adversarial network to model the distributions of both the target spectra and the background spectra. Furthermore, to our delight, we found when the generative adversarial network converged, the generator would well generate both the simulated target spectra and the background spectra, and the discriminator was also able to extract useful information to be used for target detection. Hence, a target detector could be merged with the discriminator of a GAN. The general framework of the proposed approach is shown in Figure 1.
As demonstrated in Figure 1, the whole process of the proposed approach was divided into two phases: network training and detecting. In the training process, the target spectra together with the unlabeled spectra to be detected were regarded as the targets and the backgrounds, respectively, and were fed into the general adversarial network for training. Although it seemed venturesome to deem all of the unlabeled spectra as backgrounds, the generative adversarial network proved to be able to counteract the wrong apportionment very effectively with simulated samples. When the training process converged, the discriminator of GAN was selected as the detector in the detecting phase. In this process, we obtained an initial detection map through the discriminator based on the spectral information. Furthermore, to further suppress the background and the tiny outliers, guided filters were utilized to improve the smoothness and robustness of the detection results.

2.3.2. Network Architecture and Objective Function

With respect to the specific network architectures, it should be noted that the neural network during training was divided into a generator network and a discriminator network. Furthermore, we followed different criteria while building these two networks for target detection.
As shown in Figure 1, a residual structure was adopted in the generator, which was very expressive and could be used to well model the complicated distributions of the target and the background spectra. Meanwhile, in order to stabilize the training process, spectral normalization (SN) layers which were proposed in [26] were added between every two fully connected layers. As for the discriminator, we designed a dual branch structure. One branch output the truth value indicating whether the sample was real or false, and the other output the class label indicating whether the sample was target or background. The dual branches shared as many layers as possible, linking the generative adversarial network with a detector. Parameters of our network are listed in Table 1.
As the discriminator in the proposed network was able to provide two kinds of probabilities through the dual branches, viz. one for the truth value and the other for the class label, the loss of the discriminator denoted as L d could, accordingly, be represented by two parts, i.e., L a d v for the truth loss and L c l c for the class loss, which were similar to that of ACGAN as proposed in [25] and could be defined as Equation (7):
L d = L a d v + L c l c = [ i max ( 0 , 1 s r i ) + j max ( 0 , 1 + s g j ) ] i , j [ c log c ^ + ( 1 c ) log ( 1 c ^ ) ]
where s r i and s g j denote the truth values for real and generated samples separately and c and c ^ denote the real and the predicted class labels for all the samples, respectively. As shown in Equation (7), the first two terms represented the hinge version of the adversarial loss as proposed in [26,27], while the third term represented the cross-entropy between the real and the predicted class labels. Through integrated loss, as the network converged, the discriminator could not only distinguish between real and generated samples, but also provide a class label for each sample.
As for the objective of the generator (denoted as L g ), the case became quite simple. As demonstrated in Equation (8):
L g = L g e n + α × L r e g = j s g j + α × p | X ^ p + 1 + X ^ p 1 2 X ^ p |
where X ^ p denotes a certain value on position p of the generated spectrum X ^ , the first term denotes the ordinary generative loss and is denoted as L g e n , and the second term, represented as L r e g , refers to the second-order difference and denotes a relatively weak but useful penalty which could help to smoothen the generated spectra and stabilize the training process. Worth noting, α is a penalty constant which was empirically set as 1.
As for the training process of the proposed network, it was very similar to that of conventional GAN. Specifically, the technique of back-propagation was utilized in the training process to minimize L d and L g alternately. When the network converged, it came to the detection phase. In this phase, the generator was removed and the discriminator was utilized as a detector to deal with the real spectra. Since we could utilize the discriminator at different times and with different weights, several initial detection maps could be acquired. To further suppress the background and the tiny outliers, necessary spatial processing techniques were adopted.

2.3.3. Bank of Guided Filters

Furthermore, since spatial correlation was useful in detecting certain objects in HSIs, a bank of guided filters could be utilized to smoothen the initial detection results. The framework of spatial filtering with guided filters is illustrated in Figure 2.
As shown in Figure 2, the initial detection maps obtained at different times were firstly filtered by the bank of filters individually. In this case, all the filters would adopt the initial input image as the guided image and they differed from each other only in radius r and the regularization coefficient ϵ . Then, different degrees and scales of smoothening procedures were conducted on the initial detection maps. After that, max pooling was performed on the filtered images to gather the most valuable information. At last, an overall average value regarding the images acquired at different times was calculated so as to obtain a relatively stable detection result.

3. Results

To evaluate the robustness and effectiveness of the proposed approach, experiments were conducted for the target detection of HSIs. Firstly, the assessment criteria of target detection were provided along with the hyperspectral datasets that were utilized in this paper. Then, the preferences of the proposed network were analyzed through the experimental results. Finally, comparisons were performed between our approach and some other detection approaches such as ACE, CEM, STD, etc. The results showed the proposed approach could yield superior detection performance in terms of both efficiency and accuracy.

3.1. Assessment Criteria

In order to assess the detection performance reliably, a 2D receiver operating characteristic (ROC) curve was adopted together with the intuitional detection maps. As an effective metric for the ROC curve, the area under the curve (AUC) of the false positive rate ( P f ) versus the true positive rate ( P t ) was calculated. P f , P t could be defined as Equation (9):
P f = FP FP + TN , P t = TP TP + FN
where FP and TP denote the false positive samples (backgrounds which were detected as targets) and the true positive samples (targets which were detected as targets), respectively, and FN and TN denote the false negative samples (targets which were detected as backgrounds) and the true negative samples (backgrounds which were detected as backgrounds), respectively. Therefore, P f was also known as the false alarm rate and P t the recall rate in the field of TD.

3.2. Datasets

Three HSI datasets were utilized in our experiments which were described as follows.

3.2.1. Airport–Beach–Urban Dataset

This dataset is available online and consists of scenes from airports, beaches and urban areas. Three images from each scene were selected for the following experiments, and were denoted as D1, D2 and D3, respectively. All of them were acquired by the airborne visible/infrared imaging spectrometer (AVIRIS) sensor and contained 100 × 100 pixels in the spatial domain. After removing the water-absorption and noisy bands, each group of HSIs included 205, 193, 207 bands, respectively. Their color composites and the corresponding reference maps are shown in Figure 3, and five target signatures were utilized as the training samples for each image.

3.2.2. San Diego Airport Dataset

The second dataset was acquired by the AVIRIS sensor over the San Diego airport in CA, USA. The original HSIs contained 400 × 400 pixels in the spatial domain and 189 bands after removing the water-absorption and noisy bands. Two small parts were utilized for the experiments. They all contained 100 × 100 pixels and were denoted as D4 and D5, respectively. The color composites and the corresponding reference maps are illustrated in Figure 4a,b. Each of them contained three planes which were considered as the targets to be detected and five target signatures were selected as the training samples for each image.

3.2.3. Urban Dataset

The third dataset was collected by the hyperspectral digital imagery collection experiment (HYDICE) sensor and its spatial and spectral resolutions were 2 m and 10 nm, respectively. A small part denoted as D6 with 80 × 100 pixels and 162 bands was utilized in the experiments. The color composites and the corresponding reference maps are shown in Figure 4c. There were 21 target pixels to be detected, where 3 target signatures were utilized as the training samples.

3.3. Preferences

Before further discussion, explanations were first determined here to parameter preferences, including the ratio between the target signature and the unlabeled spectra in one batch during training, the learning rate along with the number of classifiers and the guided filters during detection.
Earlier in this paper, we mentioned that the training set was composed of the target spectra and the unlabeled spectra. Since the former contained far fewer samples than the latter, only a small fraction of unlabeled spectra was selected as the backgrounds in each batch for training in the experiments. In this way, the imbalance between the two kinds of samples was greatly alleviated. In addition, the amount of target signature was doubled by adding Gaussian noises. Furthermore, the detection performance evaluated by the AUC ( P f , P t ) under different ratios between the target spectra (including noisy samples) and the unlabeled spectra in one batch is shown in Figure 5a.
From the results, we could see the ratio between the target spectra and the unlabeled spectra had an obvious influence on the detection performance, and the optimum ratio varied from different datasets. Specifically, for D1 and D5, as the proportion of unlabeled spectra in one batch increased, the AUC increased until it converged. For D3, D4 and D6, when the proportion of unlabeled spectra increased, the AUC decreased slightly. Furthermore, for D2, with the increasing of the proportion of unlabeled spectra, the AUC first increased slightly and then decreased dramatically. Judging from the different situations, we selected 1:10 as a proper ratio between the target spectra and the unlabeled spectra for the following experiments.
With respect to the learning rate, we adopted an imbalanced learning rate as in [28] with l d = [ 10 4 , 10 3 , 10 2 , 10 1 ] , l g = l d / 2 for the discriminator and the generator, respectively, and employed RMSprop as the optimizer. Since we used spectral normalization and the smooth penalty to stabilize the training process, the network could converge in less than two thousand iterations. The detection performance evaluated by the AUC ( P f , P t ) under different learning rates is shown in Figure 5b.
According to the results of Figure 5b, the learning rate had a significant influence on the detection performance, especially for D1 and D2. With a low degree of network convergence, a small learning rate such as 10 4 led to a decrease in the performance while a larger learning rate such as 10 1 could lead to a divergent network, resulting in a dramatic performance collapse. Meanwhile, the AUC changed slightly and exhibited a relatively robust performance when the learning rate fell into a range from 10 3 to 10 2 . Given this fact, the learning rate of 10 3 was adopted in the following experiments.
After the network converged, it came to the detection phase. Discriminators at 15 different moments were utilized as the detectors, and 9 different guided filter preferences were selected to smoothen the initial detection maps. The pairs between the radius r and regularization coefficient ϵ of guided filters were set to be r = [ 1 , 2 , 3 ] and ϵ = [ 0 . 4 2 , 0 . 6 2 , 0 . 8 2 ] . The detection performance before and after filtering with different filters is listed in Figure 5c. The proposed approach adopted three different settings for filters, viz. ( 1 , 0 . 8 2 ) , ( 2 , 0 . 6 2 ) and ( 3 , 0 . 4 2 ) , so as to form a filter bank as mentioned in the previous section.
As indicated in Figure 5c, the AUC ( P f , P t ) varied as the guided filter setting changed. The use of guided filtering improved the detection performance in almost every dataset, and the case was especially true for D1 and D3. However, the influence of filter settings on different datasets was quite different. For D1, ( 3 , 0 . 8 2 ) was the best setting for ( μ , ϵ ) of guided filters, while for D3, ( 1 , 0 . 4 2 ) was the optimal setting. Meanwhile, the AUC was almost the same for both D4 and D5. As the proposed approach adopted a bank of guided filters, it could always perform well in various datasets.
As neural networks on which the proposed approach was based on are often time-consuming with high computational complexity, the relevant processing complexity was explicitly addressed here to further demonstrate the effectiveness of the proposed approach. Specifically, the processing cost was evaluated in terms of the processing time (in seconds) and the amounts of learnable parameters, with results illustrated in Table 2. Meanwhile, the programs in the proposed approach were ran through Tensorflow in Python on a computer with a Quad-Core core processor of 3.40 GHz (i5-7500), GTX 1070 and 16G memory. As indicated in Table 2, a range of 0.15 to 0.17 million parameters (varying with the number of bands) was employed in the proposed network, the number of which was relatively small and was able to yield efficient detection results. Meanwhile, as the proposed approach was pixel-based, the processing time would change as the size of datasets changed. For this reason, D6 (with a smaller size of 80 × 100) showed a better detection efficiency than the other datasets.

3.4. Comparisons with Other Approaches

In this paper, we compared the proposed approach with two traditional approaches (i.e., ACE [1] and CEM [6]), two sparse-based approaches (i.e., sparse representation for target detection (STD) [7], and combined sparse and collaborative representation for target detection (CSCR) [29]) and a neural network-based approach (i.e., semi-supervised classification based on GAN (SSC) [17]). A global background estimation strategy was utilized in ACE, while an adaptive local dual window approach was adopted for the background dictionary estimation as in [29]. Furthermore, in consideration of the varied sizes of different targets, dual windows with various sizes were adopted for different datasets. For D3, the dual window size was [ 3 , 17 ] , while for D6, the dual window size was set to be [ 5 , 9 ] . Moreover, the dual window sizes for the remaining datasets were set as [ 5 , 13 ] . The detection performance is shown in Table 3, Figure 6 and Figure 7.
Table 3 shows the detection performance evaluated by AUC ( P f , P t ). From this table, we could see the proposed approach performed well and was able to achieve the highest AUC ( P f , P t ) in almost every dataset except for D3 (the result of which was only 0.0017 lower than the maximum value). The other approaches performed well in certain datasets, but +are likely to fall behind the proposed approach when dealing with other datasets. Furthermore, it is noteworthy that our approach was able to outperform the approach of SSC (which bears strong resemblance to our approach) in almost every dataset, especially in D2, which further validated the effectiveness of the unified training process with a composite loss function. Based on this fact as well as the other facts that were elaborated earlier, we could draw the conclusion that the proposed approach was more robust and could deal with various sorts of HSIs without resetting the hyperparameters.
Figure 6 shows the comparison results between different algorithms in a more visual way. The performance of a specific approach was displayed in one column. From this figure, we could observe that the detection results were quite in accord with what was shown in Table 3, which also supported the conclusion that the proposed approach was more robust when dealing with various HSIs. Furthermore, with the use of guided filters, the contour edges were preserved to a considerable extent.
Besides what was mentioned above, additional points that could be determined from the ROC curve of Figure 7 were: for D1 and D4, the proposed approach outperformed the other approaches under various false positive rates (also known as false alarm rates) and was able to allow P t to reach its peak value at a very small false alarm rate. Meanwhile, for D5 and D6, the gap between the proposed approach and the other approaches was very small when the false alarm rate was bigger than 10 2 . Furthermore, the former displayed a more competitive performance when the false alarm rate dropped to a value below 10 3 . With regard to D2 and D3, the situation became more complex. When the false alarm rate dropped to a low value, P t in certain approaches was higher. However, when the false alarm rate was higher than a certain value (about 2 × 10 3 for D2, 10 2 for D3), P t of the proposed approach outperformed the other approaches and quickly reached its maximum point. Such property of the proposed approach would make a much higher AUC as shown in Table 3.

4. Discussion

It was found that similarities exist between and among approaches of [17,18] and our approach as in these approaches, the technique of adversarial training process was exploited to fix the problem of limited labeled samples. However, it should be noted that in [17,18], the training process was separated into different phases, which was likely to result in a weakened correlation between networks. To avoid the gap between the different training phases, a neural network with a composite loss function was built in our approach, which was able to crack the problem to a comparatively satisfactory extent.
Another point that may need to be addressed more clearly here is the proposed approach utilized generated samples in the training process, the technique of which usually falls under data augmentation. However, visible differences were present between the proposed approach and other augmentation approaches such as [30,31]. Specifically, the two processes, i.e., data augmentation and the training of the target detector, were realized simultaneously as a whole in the proposed approach, but were realized sequentially in the others. Furthermore, more importantly, as a data-driven approach, no prior assumptions were needed for GAN in the proposed approach, whereas for other approaches, hypotheses (e.g., linear mixing model or subspace) are often needed for data augmentation.
Alongside the efforts performed in the proposed approach, further research can be conducted on how to develop an approach which is entirely based on neural networks for target detection using a single-target signature.

5. Conclusions

In sum, we proposed a novel hyperspectral target detection framework with an auxiliary generative adversarial network in this paper. Simulated samples of target and background spectra were generated by the generative neural network to facilitate the target detection process. Extensive experiments conducted on the HSI datasets showed the proposed target detection approach equipped with the post-processing technique of guided filtering was more competitive when compared to other approaches.

Author Contributions

Y.G. and Y.F. conceived and designed the study. Y.G. constructed the neural network, implemented the experiments and drafted the manuscript. X.Y. contributed to the improvement of the model and the manuscript. Y.F. provided the overall guidance to the project, reviewed and edited the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by Shaanxi Key Industry Innovation Chain Project (Grant Number: 2019ZDLGY14-02-02).

Data Availability Statement

Publicly available datasets were analyzed in this study. The datasets can be found here: http://levir.buaa.edu.cn/code_cn.htm, https://www.erdc.usace.army.mil/Media/Fact-Sheets/Fact-Sheet-Article-View/Article/610433/hypercube/ and http://xudongkang.weebly.com/data-sets.html, accessed on 3 November 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Nasrabadi, N.M. Hyperspectral target Detection: An overview of current and future challenges. IEEE Signal Process. Mag. 2014, 31, 34–44. [Google Scholar]
  2. Chen, H.; Zhang, F.; Tang, B.; Yin, Q.; Sun, X. Slim and Efficient Neural Network Design for Resource-Constrained SAR Target Recognition. Remote Sens. 2018, 10, 1618. [Google Scholar] [CrossRef] [Green Version]
  3. Matuszewski, J.; Pietrow, D. Recognition of electromagnetic sources with the use of deep neural networks. In Proceedings of the XII Conference on Reconnaissance and Electronic Warfare Systems, Ołtarzew, Poland, 19–21 November 2018; SPIE: Bellingham, DC, USA, 2019; Volume 11055, p. 110550D. [Google Scholar]
  4. Xu, W.; Chen, R.; Huang, B.; Zhang, X.; Liu, C. Single Image Super-Resolution Based on Global Dense Feature Fusion Convolutional Network. Sensors 2019, 19, 316. [Google Scholar] [CrossRef] [Green Version]
  5. Kwon, H.; Nasrabadi, N.M. A Comparative Analysis of Kernel Subspace Target Detectors for Hyperspectral Imagery. Eurasip J. Adv. Signal Process. 2006, 2007, 1–13. [Google Scholar] [CrossRef] [Green Version]
  6. Geng, X.; Ji, L.; Sun, K.; Zhao, Y. CEM: More Bands, Better Performance. Geosci. Remote Sens. Lett. 2014, 11, 1876–1880. [Google Scholar] [CrossRef]
  7. Chen, Y.; Nasrabadi, N.M.; Tran, T.D. Sparse Representation for Target Detection in Hyperspectral Imagery. IEEE J. Sel. Top. Signal Process. 2011, 5, 629–640. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Du, B.; Zhang, L. A Sparse Representation-Based Binary Hypothesis Model for Target Detection in Hyperspectral Images. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1346–1354. [Google Scholar] [CrossRef]
  9. Guo, T.; Luo, F.; Zhang, L.; Tan, X.; Liu, J.; Zhou, X. Target Detection in Hyperspectral Imagery via Sparse and Dense Hybrid Representation. IEEE Geosci. Remote Sens. Lett. 2020, 17, 716–720. [Google Scholar] [CrossRef]
  10. Shi, Y.; Lei, J.; Yin, Y.; Cao, K.; Li, Y.; Chang, C.-I. Discriminative Feature Learning With Distance Constrained Stacked Sparse Autoencoder for Hyperspectral Target Detection. Geosci. Remote Sens. Lett. IEEE 2019, 16, 1462–1466. [Google Scholar] [CrossRef]
  11. Jiang, T.; Xie, W.; Li, Y.; Du, Q. Discriminative Semi-Supervised Generative Adversarial Network for Hyperspectral Anomaly Detection. In Proceedings of the IGARSS 2020-2020 IEEE International Geoscience and Remote Sensing Symposium, Waikoloa, HI, USA, 26 September–2 October 2020; pp. 2420–2423. [Google Scholar]
  12. Xie, W.; Yang, J.; Lei, J.; Li, Y.; Du, Q.; He, G. SRUN: Spectral Regularized Unsupervised Networks for Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1463–1474. [Google Scholar] [CrossRef]
  13. Xie, W.; Lei, J.; Yang, J.; Li, Y.; Du, Q.; Li, Z. Deep Latent Spectral Representation Learning-Based Hyperspectral Band Selection for Target Detection. IEEE Trans. Geosci. Remote Sens. 2020, 58, 2015–2026. [Google Scholar]
  14. Shi, Y.; Li, J.; Yin, Y.; Xi, B.; Li, Y. Hyperspectral Target Detection With Macro-Micro Feature Extracted by 3-D Residual Autoencoder. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 4907–4919. [Google Scholar] [CrossRef]
  15. Du, J.; Li, Z. A Hyperspectral Target Detection Framework With Subtraction Pixel Pair Features. IEEE Access 2018, 6, 45562–45577. [Google Scholar] [CrossRef]
  16. Zhu, D.; Du, B.; Zhang, L. Two-Stream Convolutional Networks for Hyperspectral Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6907–6921. [Google Scholar] [CrossRef]
  17. Zhan, Y.; Wu, K.; Liu, W.; Qin, J.; Yang, Z.; Medjadba, Y.; Wang, G.; Yu, X. Semi-Supervised Classification of Hyperspectral Data Based on Generative Adversarial Networks and Neighborhood Majority Voting. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 5756–5759. [Google Scholar]
  18. Shi, Y.; Li, J.; Li, Y.; Du, Q. Sensor-Independent Hyperspectral Target Detection With Semisupervised Domain Adaptive Few-Shot Learning. IEEE Trans. Geosci. Remote Sens. 2021, 59, 6894–6906. [Google Scholar] [CrossRef]
  19. Jiao, C.; Chen, C.; Gou, S.; Wang, X.; Yang, B.; Chen, X.; Jiao, L. L1 Sparsity-Regularized Attention Multiple-Instance Network for Hyperspectral Target Detection. IEEE Trans. Cybern. 2021, 1–14. [Google Scholar] [CrossRef]
  20. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 27 (Nips 2014) 2014, 27, 2672–2680. [Google Scholar]
  21. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  22. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  23. Nowozin, S.; Cseke, B.; Tomioka, R. f-GAN: Training generative neural samplers using variational divergence minimization. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 271–279. [Google Scholar]
  24. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2180–2188. [Google Scholar]
  25. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  26. Yoshida, Y.; Miyato, T. Spectral Norm Regularization for Improving the Generalizability of Deep Learning. arXiv 2017, arXiv:1705.10941. [Google Scholar]
  27. Zhang, H.; Goodfellow, I.; Metaxas, D.; Odena, A. Self-attention generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 7354–7363. [Google Scholar]
  28. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. GANs trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6629–6640. [Google Scholar]
  29. Li, W.; Du, Q.; Zhang, B. Combined sparse and collaborative representation for hyperspectral target detection. Pattern Recognit. 2015, 48, 3904–3916. [Google Scholar] [CrossRef]
  30. Rochac, J.F.R.; Zhang, N.; Thompson, L.; Oladunni, T. A data augmentation-assisted deep learning model for high dimensional and highly imbalanced hyperspectral imaging data. In Proceedings of the 2019 9th International Conference on Information Science and Technology (ICIST), Hulunbuir, China, 2–5 August 2019; pp. 362–367. [Google Scholar]
  31. Yang, X.; Dong, M.; Wang, Z.; Gao, L.; Zhang, L.; Xue, J.H. Data-augmented matched subspace detector for hyperspectral subpixel target detection. Pattern Recognit. 2020, 106, 107464. [Google Scholar] [CrossRef]
Figure 1. General network of our approach.
Figure 1. General network of our approach.
Remotesensing 13 04454 g001
Figure 2. Spatial filtering with guided filters.
Figure 2. Spatial filtering with guided filters.
Remotesensing 13 04454 g002
Figure 3. Color composites and reference maps: (a) D1, (b) D2 and (c) D3.
Figure 3. Color composites and reference maps: (a) D1, (b) D2 and (c) D3.
Remotesensing 13 04454 g003
Figure 4. Color composites and reference maps: (a) D4, (b) D5 and (c) D6.
Figure 4. Color composites and reference maps: (a) D4, (b) D5 and (c) D6.
Remotesensing 13 04454 g004
Figure 5. AUC ( P f , P t ) under different (a) ratios between samples in one batch, (b) learning rates of the discriminator, (c) settings of guided filters.
Figure 5. AUC ( P f , P t ) under different (a) ratios between samples in one batch, (b) learning rates of the discriminator, (c) settings of guided filters.
Remotesensing 13 04454 g005
Figure 6. Detection results of different approaches: (a) color composites, (b) reference maps, (c) ACE, (d) CEM, (e) STD, (f) CSCR, (g) SSC and (h) the proposed.
Figure 6. Detection results of different approaches: (a) color composites, (b) reference maps, (c) ACE, (d) CEM, (e) STD, (f) CSCR, (g) SSC and (h) the proposed.
Remotesensing 13 04454 g006
Figure 7. ROC curve ( P f , P t ) of different datasets: (a) D1, (b) D2, (c) D3, (d) D4, (e) D5 and (f) D6.
Figure 7. ROC curve ( P f , P t ) of different datasets: (a) D1, (b) D2, (c) D3, (d) D4, (e) D5 and (f) D6.
Remotesensing 13 04454 g007
Table 1. Parameter of the proposed network.
Table 1. Parameter of the proposed network.
NetworkLayerTypeShapeActivation
GeneratorInputFully Connected 16 × 128 None
SN ReLU
Res. 1–3Fully Connected 128 × 64 None
SN ReLU
Fully Connected 64 × 128 None
OutputFully Connected 128 × N b a n d s None
SN Tanh
DiscriminatorInputFully Connected N b a n d s × 256 Leaky ReLU
Fully Connected 256 × 128 Leaky ReLU
Branch 1Fully Connected 128 × 32 Leaky ReLU
Fully Connected 32 × 1 Sigmoid
Branch 2Fully Connected 128 × 32 Leaky ReLU
Fully Connected 32 × 2 Sigmoid
Table 2. The processing cost of the proposed approach in various datasets.
Table 2. The processing cost of the proposed approach in various datasets.
DatasetsD1D2D3D4D5D6
Time (sec.)Train11.8312.2512.0612.0311.899.57
Test3.053.123.062.993.082.90
Learnable Parameters 1.5 × 10 5 1.7 × 10 5
Table 3. Detection results evaluated by AUC ( P f , P t ) of different approaches.
Table 3. Detection results evaluated by AUC ( P f , P t ) of different approaches.
DatasetAUCACECEMSTDCSCRSSCProposed
D1 ( P f , P t ) 0.91550.91990.88540.91630.98910.9962
D2 ( P f , P t ) 0.87860.88400.91300.95810.94200.9922
D3 ( P f , P t ) 0.85300.81600.94210.97170.99410.9924
D4 ( P f , P t ) 0.92480.91630.88830.95380.98840.9968
D5 ( P f , P t ) 0.99530.99650.99660.99790.99290.9987
D6 ( P f , P t ) 0.97890.98950.99620.99680.99740.9976
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, Y.; Feng, Y.; Yu, X. Hyperspectral Target Detection with an Auxiliary Generative Adversarial Network. Remote Sens. 2021, 13, 4454. https://doi.org/10.3390/rs13214454

AMA Style

Gao Y, Feng Y, Yu X. Hyperspectral Target Detection with an Auxiliary Generative Adversarial Network. Remote Sensing. 2021; 13(21):4454. https://doi.org/10.3390/rs13214454

Chicago/Turabian Style

Gao, Yanlong, Yan Feng, and Xumin Yu. 2021. "Hyperspectral Target Detection with an Auxiliary Generative Adversarial Network" Remote Sensing 13, no. 21: 4454. https://doi.org/10.3390/rs13214454

APA Style

Gao, Y., Feng, Y., & Yu, X. (2021). Hyperspectral Target Detection with an Auxiliary Generative Adversarial Network. Remote Sensing, 13(21), 4454. https://doi.org/10.3390/rs13214454

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop