Next Article in Journal
Electrochemical Determination of Nitrite by Au Nanoparticle/Graphene-Chitosan Modified Electrode
Next Article in Special Issue
Rapid Airplane Detection in Remote Sensing Images Based on Multilayer Feature Fusion in Fully Convolutional Neural Networks
Previous Article in Journal
Road Anomalies Detection System Evaluation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Effective Image Denoising Method for UAV Images via Improved Generative Adversarial Networks

1
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
2
Department of Mathematics, University of California, Irvine, CA 92697, USA
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2018, 18(7), 1985; https://doi.org/10.3390/s18071985
Submission received: 25 April 2018 / Revised: 17 June 2018 / Accepted: 20 June 2018 / Published: 21 June 2018
(This article belongs to the Special Issue High-Performance Computing in Geoscience and Remote Sensing)

Abstract

:
Unmanned aerial vehicles (UAVs) are an inexpensive platform for collecting remote sensing images, but UAV images suffer from a content loss problem caused by noise. In order to solve the noise problem of UAV images, we propose a new methods to denoise UAV images. This paper introduces a novel deep neural network method based on generative adversarial learning to trace the mapping relationship between noisy and clean images. In our approach, perceptual reconstruction loss is used to establish a loss equation that continuously optimizes a min-max game theoretic model to obtain better UAV image denoising results. The generated denoised images by the proposed method enjoy clearer ground objects edges and more detailed textures of ground objects. In addition to the traditional comparison method, denoised UAV images and corresponding original clean UAV images were employed to perform image matching based on local features. At the same time, the classification experiment on the denoised images was also conducted to compare the denoising results of UAV images with others. The proposed method had achieved better results in these comparison experiments.

1. Introduction

As a rapid evolution technology, the increased availability of unmanned aerial vehicles (UAVs) has drawn attention for their capability to generate ultra-high spatial resolution images. UAV however is a low-altitude remote sensing platform, which is affected by lightning, ground electromagnetic waves, illumination change, and mechanical noise from the UAV itself. These factors are sources of noise in UAV images, therefore, it is especially important to study how to remove noise from UAV images.
In recent years, there is a growing body of research on noise removal in remote sensing images. Liu et al. [1] used an auxiliary noise-free image as a prior, proposing a denoising method for remote sensing images based on partial differential equations. Rajapriyadharshini et al. [2] split the noisy images into several disjoint local regions and clustered the noisy images into several disjoint local regions to denoise SAR images. Bhosale et al. [3] designed wavelet filter to restore the remote sensing images and explored the effects of noise. Wang et al. [4] proposed high-order balanced multi-band multiwavelet packet transforms to denoise remote sensing images and claimed that an appropriate band number can improve the denoising performance. Xu et al. [5] provided a method based on blockwise nonlocal means algorithm to denoise repetitive image patches in remote sensing images. Penna et al. [6] utilized non local means with the stochastic distances to denoise SAR images. Kim et al. [7] adopted background registration processing and robust principle component analysis, proposing a method of noise filtering of LWIR/MWIR Imaging sensors.
Some researchers studied the denoising of remote sensing images based on sparse expression. Chang et al. [8] combined unidirectional total variation and sparse representation to learn a dictionary trained to fit the input data to remove random noises from remote sensing images. Cerra et al. [9] reduced the weight of sparse unmixing problem proposing a denoising method based on sparse reconstruction of simulated EnMAP data. Xu et al. [10] used the nonlocal sparse model and the iterative regularization technique to denoise SAR images.
Nowadays, deep learning is breeding new ideas and convolutional neural networks (CNNs) have become recognized as an efficient method that automatically learns deep-level feature representations from images. Denoising algorithms for natural images, based on deep learning are an emerging trend. Jain et al. [11] synthesized training samples from specific noise models and used convolutional networks as unsupervised learning procedure and image processing architecture. Xie et al. [12] put forward a method of image denoising that combined sparse representation and deep neural networks pre-trained by denoising auto-encoders. Burger et al. [13] introduced a plain multi-layer perceptron of mapping from a noisy image to a noise-free image. Wu et al. [14] used rectified linear function instead of sigmoid function as the hidden layer activation function of deep neural networks to achieve image denoising. Xu et al. [15] provided a method of using deep convolutional neural networks as reliable a support for robust deconvolution against artifacts for image restoration. Li et al. [16] combined sparse coding and auto-encoder to achieve image denoising. Mao et al. [17] put forward networks with multiple layers of convolution and deconvolution operators to learned end-to-end mappings from a denoised image to a clean image.
Many practical remote sensing applications require clear textures for ground objects in remote sensing images. Because of the wide range of remote sensing images, remote sensing image denoising must consider the overall spatial distribution of ground objects, which is different from the natural images. It is easier to extract deep texture features in remote sensing images from the deeper neural network structure. Nevertheless, the deeper networks are often more complicated to train because of the internal covariate shift. Batch normalization can address this problem due to the fact that it back propagates the gradients through the normalization parameters and preserves the representation ability of the networks [18,19,20]. Residual learning framework solves the gradient vanishing of deeper neural networks because it explicitly lets each few stacked layers fit a residual mapping instead of expecting these layers directly fit a desired underlying mapping [21,22,23,24]. Visual Geometry Group networks (VGGs) are beneficial for the deep-level representation of images [25,26,27], which is of great help for modeling mappings of highly complex ground features. Johnson et al. [27] utilized pre-trained VGG nets to extract high-level features by optimizing the perceptual loss function. Kiasari et al. [28] used perceptual loss to alleviate the problem of blurry image generation. Generative Adversarial Networks (GAN) with constantly improving optimization strategies have been applied widely to image generation and classification problems [20,22,29,30]. Dosovitskiy et al. [31] combined Euclidean distances with GAN to calculate the distances between image features extracted.
On the basis of distinctive demand for texture features from UAV images and combining with the continuous development of deep learning methodology, we propose a method based on generative adversarial networks to obtain clearer feature textures for denoised UAV images. In this paper, pretrained VGG networks are employed to extract deep-level complex features of UAV images and a perceptual reconstruction loss is combined with the pixel-based Euclidean loss for constant optimizations of the game theoretic min-max model to generate fake clean UAV images using powerful generation capability of GAN. In order to make the networks structure deeper and sturdier, the batch-normalization is used to regularize training data and multiple residual blocks are utilized to build the networks.

2. Methods

2.1. Deep Architecture

GAN can effectively learn the distributions of clean and noisy UAV images, using clean UAV images to denoise UAV images. The proposed neural network based on GAN directly learns the mapping between clean UAV images and noisy UAV images, based on the work of Goodfellow et al. [29], who generated realistic data by effectively learning the distributions of a training data set. They adopted a min-max adversarial game theoretic optimization framework to train a generative model G and discriminative model D simultaneously. GAN continuously train a model G so that the probability distributions of the images generated by G are indistinguishable from real images, thus fooling model D. According to this idea, we consider the UAV image image-denoising problem as a generative adversarial problem, aiming to learn a mapping directly from clean UAV images to noisy UAV images by constructing a GAN-based deep networks called Denoise UAV image Generative Adversarial Network (DUGAN). In order to learn a generator that can fool a discriminative model (D) so it can distinguish clean UAV images from generated denoised UAV images. Thus, a generative model G and a discriminative model D were designed.

2.1.1. Generative Model

The primary aim of generative model G is to generate clean denoised UAV images from real clean UAV images. As the original GAN model is unstable, artifacts in the output images are synthesized by generative model G, which do not meet ground-object texture requirements of UAV image applications. Therefore, the critical is to design a deeper structure to generate denoised UAV images. We constructed a generative model G containing 14 residual blocks (Figure 1) to train the deeper networks in the G model, efficiently.
The gradient can easily vanish in the process of back propagation while training the deeper networks, which may result in the losses of image details and textures. The residual networks make a reference to each layer’s input and learn a residual function rather than learn some functions that do not have references. This residual function is easier to optimize to solve the gradient vanishing problem structurally, at the level of the deeper networks, but greatly increases the number of network layers In the G model. Each residual block of generative model G comprises two convolutional layers and two batch normalization layers, then a skip connection is established in each residual block. Skip connection back-propagates the gradient to deeper layers and can pass gradients through multiple residual blocks smoothly, which helps recover the details in UAV images and assists the CNN to effectively denoise UAV images.
The distributions of data will have influence on the training of deep networks. When the depth of a network increases, the overall distribution of the activation input gradually approaches the upper and lower limits of the value interval of the non-linear function of the value interval, resulting in a slow network convergence. Batch normalization forces the distributions of the input values of any neuron in each layer of the neural networks back to the standard normal distribution, so that the active input values fall into the sensitive area of the nonlinear function. Small changes can lead to large changes in the loss function, which makes the training faster and easier. In our generative model G, the activation functions of two batch normalization layers in the residual blocks are relu. In our generative model G, the activation functions of two batch normalization layers in the residual blocks are relu. In addition to these residual blocks, we add two convolutional layers to generate simulated clean UAV images using the activation function, tanh. The specific settings of each layer of the generative model are as follows:
C(r,64)→C(64)B(r)C(64)B(r)SC→…12…→C(64)B(r)C(64)B(r)SC→C(64)→C(t,3)
Here, C(r,64) denotes a set of convolutional layers with 64 feature maps and activation function relu; C(64)B(r)C(64)B(r)SC represents a residual block; B(r) is a batch normalizational layer with activation function relu, and SC denotes a skip connection. There are in total, 14 residual blocks. C(t,3) represents a convolutional layer with three feature maps and activation function tanh.

2.1.2. Discriminative Model

GAN continuously trains the generative model G until the generated images cannot be differentiated by the discriminative model D from these actual, real image samples. Generative model G continuously generates better images so that the distributions of the images are undifferentiated from the distributions of real images. The discriminative model D is employed to distinguish simulated clean UAV images synthesized by the generative model from the corresponding actual clean UAV images. Model D can also be regarded as a judge and guidance for the generative model G. We constructed a discriminative model D that alternately updates G and D to solve the min-max adversarial game-theoretic optimization problem (Equation (1)):
min G   max D   E G C I p g [ log ( 1 D ( G C I ) ) ] + E I C p r [ log D ( I C ) ]
Here, IC represents the actual real clean image and G C I denotes the generated clean image, p r is the sample distribution of the real clean image, p g is the sample distribution of the generated clean image. The proposed discriminative model is shown at the bottom part of Figure 1. The specific settings of each layer of the discriminative model are as follows:
C(lr,64)→C(128)BN(lr)→C(256)BN(lr)→C(512)BN(lr)→C(1024)BN(lr)→C(2048k)BN(lr)→C(128)BN(lr)→C(128)BN(lr)→C(512)BN(lr)→SC→D
Here, lr denotes the activation function leakyrelu; C(128)BN(lr) denotes a set of convolutional layers with 128 feature maps followed by batch-normalization with activation function leakyrelu, and D is the dense layer for outputting generated clean UAV images that are indistinguishable from actual real clean UAV images. The feature maps were increased from 256 to 2048, and suitable for pre-trained VGG networks [32].

2.2. Loss Function

The definition of loss function is critical for the performance of the proposed method. Some losses of image restoration are optimized at pixel-level [31,33,34] so that images are typically, overly smooth and thus lack high frequency content and have poor perceptual quality. Some researchers argued that reconstructed results would be better to optimize a perceptual loss function by minimizing perceptual differences between reconstructed images and the true ground images [27,35,36]. The quality of the image features can be improved by the perceptual reconstruction loss so that it can meet the requirements of UAV images for features acquired from ground objects textures. GAN have powerful ability of image generation by alternatively updating generative networks G and discriminative networks D. Dosovitskiy et al. [31] combined Euclidean distances with generative adversarial training to establish a loss function. The solutions of pixel loss optimization problems often result in perceptually unsatisfying solutions with overly smooth textures. Combining the perceptual reconstruction loss function with VGG networks the networks will encourage networks to enjoy feature representations of noisy images similar to those of actual clean images. Therefore, we propose a new advanced loss equation for better denoising results of UAV images. Perceptual reconstruction loss, generative adversarial loss and Euclidean loss are combined together to formulate the proposed loss function, which is as follows:
L D U G A N = x L p e + y L g a + L p i
Here, L p e is perceptual reconstruction loss, an appropriate measure for features extracted from a pretrained VGG networs instead of low-level pixel-wise error measures; L g a is adversarial loss; and L p i is pixel loss between noisy pixel of the noisy UAV images and pixel of the clean UAV images. x and y are respectively the weights of L p e and L g a . The perceptual reconstruction loss based on the relu activation layers of the pretrained 19 layer VGG networks is defined as in [26,30]. The aim is to minimize the distances between high-level features, and L p e defined as Equation (3):
L p e = 1 C i W i H i c = 1 C i w = 1 W i h = 1 H i | | V ( U I C I c , w , h ) V ( U I G C I c , w , h ) | | 2 2
Here, C i , W i , H i represent the channels, width and height of the images respectively, and V represents a non-linear CNN transformation pretrained by VGG19. U I G C I denotes generated clean UAV image and U I C I denotes the corresponding actual clean UAV image.
The generative adversarial loss encourages our networks to obtain better solutions laying in the manifold of reconstructed images by trying to fool the discriminative model, which is defined based on the probability that the discriminant model considers the generated denoised UAV images to be actual, clean UAV images, as shown in Equation (4):
L g a = n = 1 N log D ( U I G C I )
Here, D ( U I G C I ) is the probability that the generated images are real clean UAV images. We minimize L g a continuously for better denoising results of UAV images. The per-pixel Euclidean loss is defined as Equation (5):
L p i = 1 C W H c = 1 C w = 1 W h = 1 H | | U I C I c , w , h U I G C I c , w , h | | 2 2
Here, U I G C I is generated clean UAV image and U I C I represents the corresponding real clean UAV image.

3. Experiments

3.1. Experimental Setting

3.1.1. Experimental Data

Because of the lack of UAV image data sets for denoising training and assessment, a new UAV image set for training and testing the proposed networks was built for our experiments. The UAV image training set was obtained by a CW-30 UAV in Guiyang city, Guizhou Province. The camera was a H5D-50 with the focal length of 50 mm; the size of entire image was 8176 × 6132 pixels and the flying height was 600–800 m. For the convenience of training, we cut entire images into small images using Photoshop; 400 images of 360 × 360 pixels were used as the clean training data and different levels of noise were added to these clean images to make them noisy images. The testing set includes two parts; one part consists of 40 pieces of UAV images and the other consists of 100 pieces of smaller UAV images, comprised of images of cars and trucks. The testing set was obtained from other mapping areas. One part of the testing set was collected with a CW-30 UAV in Yangjiang city, Guangdong Province. The camera was a SWDC5 with a focal length of 50 mm; the size of entire image is 8206 × 6078 pixels and the flying height is 600–800 m. The other test dataset was obtained using a CW-10 UAV in Wuhan city, Hubei Province. The camera was a ILCE-7R with a focal length of 28 mm; the size of entire image is 7360 × 4916 pixels and the flying height is 400–600 m.

3.1.2. Prameters Setting and Model Details

The entire networks are trained on a Nvidia GRID M60-8Q (8G) GPU using the tensorflow framework, the number of training iterations is 160 k. Due to the limited memory of the computer, the batch size of our experiment was 1. We used Aadm as an optimization algorithm and set the learning rate at 0.9. In the training process, we set x = 0.5 × 10−3, y = 2 × 10−6 (in Equation (2)) by experimental experiences. All strides were 1 in the generative model G, while all other convolutions were composed of 3 × 3 sized kernels, except for the last convolution with a 1 × 1 sized kernel. In the discriminative model D, the first six layers were composed of 4 × 4 sized kernels with a stride of 2; the next layer was composed of 1 × 1 sized kernels with a stride 1 and the last two layers were composed of 3 × 3 sized kernels with a stride 1. In generative model G and discriminative model D, all padding modes padding edges of kernels.

3.2. Comparison and Qualitative Evaluation

We added synthesized noise to the testing images with three noise levels: 20, 35 and 55, and compare the proposed method with several state-of-the-art methods used for denoising of remote sensing images. These comparative results are presented in the following subsections.

3.2.1. Comparion in the Traditional Way

Figure 2 shows results when the noise level was 35. Images in columns a, b, c, d, and e are five randomly picked testing images showing different ground objects. The labels a1-e1 denotes the actual clean UAV images, the a2-e2 denotes the synthetic noise image, and the a3–e3 denotes the UAV image denoised by method [8]. The label a4–e4 denotes the denoised UAV image by method [5], the a5–e5 identifies the UAV image denoised by method [17], and the a6–e6 denotes the UAV image denoised by the proposed method. In each of the images, small rectangles in white and blue identify areas enlarged for examination in the larger white and blue rectangles in each series of images with different levels of noise.
In Figure 2 and Figure 3, the experimental results of five randomly picked testing images of different ground objects have been provided. In the testing experiments with different noise levels, it can be observed that the proposed method preserves more distinct ground edges and more clear ground objects textures. Meanwhile, the denoised UAV images by DUGAN are better in overall visual effect and closer to the true ground objects in terms of UAV image structures. Due to the fact that we have relatively deeper networks and the loss function can better preserve the overall styles of the UAV images, deeper ground feature information can be extracted, which is of great help for the later application of UAV images.
As can be seen in the images shown in Figure 2, at noise level 35, the proposed method outperforms the other tested methods. For example, in the e series of images, the full size and enlarged portion of the image processed using proposed method is clearer, with sharper edges than the results from the other tested methods. Figure 3 is similar to Figure 2, with the same labeling scheme, order, and enlarged details, but shows results at noise level 55. It can be seen in Figure 2 and Figure 3, a qualitative visual comparison of the proposed method and the other tested methods at different noise levels shows that the proposed method preserves distinct ground edges and clear ground object textures. The denoised UAV images produced by the proposed method are better in overall visual effect and closer to the true ground objects in terms of UAV image structures. Because the proposed model have relatively deeper networks and the loss function thus preserves the structure of terrain shown in the UAV images, so that deeper ground feature information can be extracted, which is of great help in later applications using these UAV images.
Table 1 and Table 2 quantitatively compare denoised images obtained using the tested denoising methods using the Peak Signal to Noise Ratio (PSNR) and Structural Similarity Index (SSIM). Table 1 presents comparative image denoising results using PNSR for the tested methods at different noise levels. The columns indicate the tested method, while rows 3–7 represent the PSNR values of denoised images obtained using the test denoising methods when the noise level is 20. The rows 10–14 represent the PSNR values of denoised images obtained with several denoising methods when the noise level is 35, rows 17–21 represent the PSNR values of denoised images obtained with several denoising methods when the noise level is 55. Table 2 is similar to Table 1, but shows results using SSIM for the tested methods at different noise levels.
As can be seen in Table 1 and Table 2, the proposed method has the highest PSNR and SSIM [37] in different noise levels can be observed, this is consistent with the visual effects of the denoised UAV images.

3.2.2. Compare Denoising Results Using Image Matching

To verify our method further, the denoised UAV images gathered by several methods are matched with the real clean UAV images so that the results of the matching experiments were employed to compare the outcomes of denoising. Scale Invariant Feature Transform (SIFT) [38] is a well-known matching algorithm in image matching based on matching methods of image local features. The quality of the matching results can often reflect the similarities of the local characteristics of the two images, and the number of matching point pairs is the standard for judging the quality of matching results. In terms of the professional characteristics, the textures of ground features are important expressions of the UAV images of local features. Therefore, the matching results of denoising UAV images by SIFT can indicate the denoising results.
In order to ensure the objectivity of the experiment, the denoising results of the above five pieces of UAV images are used to match with the corresponding real clean UAV images by SIFT algorithm. Figure 4 and Figure 5 shows the matching results by SIFT. In Figure 4, when the noise level is 35, the labels b1–e1 indicates the SIFT matching result of the original clean image and the denoised image obtained by method [8], the labels b2–e2 indicates the SIFT matching result of the original clean image and the denoised image obtained by method [5], the labels b3–e3 indicates the SIFT matching result of the original clean image and the denoised image obtained by method [17], the labels b4–e4 indicates the SIFT matching result of the original clean image and the denoised image obtained by the proposed method. Figure 5 is similar to Figure 4, but shows results at noise level 55.
In Table 3, Column 1 denotes five randomly picked testing images, Column 2 denotes correct matching pairs of the original clean image and the denoised image obtained by method [8], Column 3 denotes correct matching pairs of the original clean image and the denoised image obtained by method [5], Column 4 denotes correct matching pairs of the original clean image and the denoised image obtained by method [17], Column 5 denotes correct matching pairs of the original clean image and the denoised image obtained by proposed method.
As is shown in Figure 3, Figure 4 and Figure 5, it can be observed that the denoised images generated by proposed method in two noise levels (noise level: 35, 55) obtain more correct matching pairs than other methods in number.

3.2.3. Compare Denoising Results Using Image Classification

In this experiment, we used 100 denoised UAV images (60 cars and 40 trucks) to conduct images classification experiment to compare denoising results of several methods (Figure 6). The classification networks consist of five layers: one input layer, three hidden convolutional neural networks (convolution kernel size is 3), and a softmax output layer. We used 1 and 0 to represent clean images of cars and trucks, respectively. Then, we employed 80 clean UAV images to train classification networks and 20 clean UAV images to test the trained classification networks. After completing the classification networks training, the classification correct rate of the testing images reaches 85%. The definition of classification correct rate is shown in Equation (6):
Correct   rate = N C N T
where, N C is the number of correctly classified UAV images and N T is the number of total UAV images. We add different noise levels (noise level: 35, 55) to those 100 clean images, and obtain denoised images by several denoising methods. The trained classification networks are used to classify the denoised UAV images. From Table 4, it can be seen that the classification results of denoised images obtained by our method enjoy a higher accuracy rate, which can reflect indirectly that denoised images obatined by our method are more similar to clean images.
Through these three comparative experiments with different noise levels, it can be observed that the denoised images obtained by the proposed method can obtain better denoising results. In the SIFT matching algorithm, the denoised images of our method gets more correct matching pairs, which shows that our method can better restore the local characteristics of UAV images and preserve the textures, as observed via classification experiments with denoised images that denoised images generated by our method are more similar to real clean images.

4. Conclusions

In this paper, we use perceptual reconstruction loss function and pixel-based loss function and propose a method of denoising UAV images based on generative adverserial networks. According to the special requirements of the UAV images for the textures of ground objects, multiple residual blocks are used to build a deep learning framework, which makes the denoised images obtain more details of the texture features. The denoising results of this method yield better results in the traditional evaluation methods. Meanwhile, in the experiments based on local feature matching and image classification, good results are achieved, which is helpful for subsequent applications of UAV images. In the deep networks, each layer of the networks can be viewed as a filter. The deeper networks have more filters of UAV images denoising, which make the entire networks become more nonlinear. The nonlinearity of the networks is the critical factor of the proposed method, which makes it superior to other denoising methods. In the future work, we will further explore how to simulate real UAV noise so as to obtain more ideal denosing effect of UAV images based on adversative learning.

Author Contributions

R.W. executed all the analyses and wrote most of the paper; X.X. and R.W. reviewed the content and offered substantial improvement for this paper; B.G. and Q.Q. development and coordination of the project and helped in the interpretation of the results; R.C. shaped significant ideas for the analyses of this paper.

Funding

This work was supported by the National Natural Science Foundation of China (No. 91638203).

Acknowledgments

We are thankful for Stephen C. McClure for providing us English editing of the manuscript freely.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, L.; Huang, F.; Li, G.; Liu, Z. Remote-Sensing Image Denoising Using Partial Differential Equations and Auxiliary Images as Priors. IEEE Geosci. Remote Sens. Lett. 2012, 9, 358–362. [Google Scholar] [CrossRef]
  2. Rajapriyadharshini, R.; Benadict, R.J. SAR Image Denoising Via Clustering Based Linear Discriminant Analysis. In Proceedings of the International Conference on Innovations in Information Embedded and Communication Systems 2015, Coimbatore, India, 19–20 March 2015; pp. 1–6. [Google Scholar]
  3. Bhosale, N.P.; Manza, R.R.; Kale, K.V. Design and Development of Wavelet Filter to Remove Noise from ASTER Remote Sensing Images. Int. J. Appl. Eng. Res. 2015, 10, 29–34. [Google Scholar]
  4. Wang, H.; Wang, J.; Jin, H.; Li, H.; Jia, D.; Ma, Y.; Liu, Y.; Yao, F.; Yang, Q. High-order balanced M-band multiwavelet packet transform-based remote sensing image denoising. EURASIP J. Adv. Signal Process. 2016, 2016, 10. [Google Scholar] [CrossRef]
  5. Xu, S.; Zhou, Y.; Xiang, H.; Li, S. Remote Sensing Image Denoising Using Patch Grouping-Based Nonlocal Means Algorithm. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2275–2279. [Google Scholar] [CrossRef]
  6. Penna, P.A.A.; Mascarenhas, N.D.A. (Non-)homomorphic approaches to denoise intensity SAR images with non-local means and stochastic distances. Comput. Geosci. 2017, 18, 127–128. [Google Scholar] [CrossRef]
  7. Kim, B.H.; Kim, M.Y.; Chae, Y.S. Background Registration-Based Adaptive Noise Filtering of LWIR/MWIR Imaging Sensors for UAV Applications. Sensors 2018, 18, 60. [Google Scholar] [CrossRef] [PubMed]
  8. Chang, Y.; Yan, L.; Fang, H.; Liu, H. Simultaneous Destriping and Denoising for Remote Sensing Images With Unidirectional Total Variation and Sparse Representation. IEEE Geosci. Remote Sens. Lett. 2014, 11, 1051–1055. [Google Scholar] [CrossRef]
  9. Cerra, D.; Bieniarz, J.; Storch, T.; Müller, R.; Reinartz, P. Restoration of EnMAP Data through Sparse Reconstruction. In Proceedings of the 7th Workshop on Hyperspectral Image and Signal Processing: Evolution in Remote Sensing—WHISPERS 2015, Tokyo, Japan, 2–5 June 2015. [Google Scholar]
  10. Xu, B.; Cui, Y.; Li, Z.; Yang, J. An Iterative SAR Image Filtering Method Using Nonlocal Sparse Model. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1635–1639. [Google Scholar] [CrossRef]
  11. Jain, V.; Seung, H.S. Natural Image Denoising with Convolutional Networks. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–11 December 2008; pp. 769–776. [Google Scholar]
  12. Xie, J.; Xu, L.; Chen, E. Image Denoising and Inpainting with Deep Neural Networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 341–349. [Google Scholar]
  13. Harmeling, S. Image denoising: Can plain Neural Networks compete with BM3D? In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 2392–2399. [Google Scholar]
  14. Wu, Y.; Zhao, H.; Zhang, L. Image Denoising with Rectified Linear Units. In Proceedings of the International Conference on Neural Information Processing 2014, Kuching, Malaysia, 3–6 November 2014; pp. 142–149. [Google Scholar]
  15. Xu, L.; Ren, J.S.J.; Liu, C.; Jia, J. Deep convolutional neural network for image deconvolution. In Proceedings of the 27th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; pp. 1790–1798. [Google Scholar]
  16. Li, H.M. Deep Learning for Image Denoising. Int. J. Signal Process. Image Process. Pattern Recognit. 2014, 7, 171–180. [Google Scholar] [CrossRef]
  17. Mao, X.J.; Shen, C.; Yang, Y.B. Image Restoration Using Very Deep Convolutional Encoder-Decoder Networks with Symmetric Skip Connections. In Proceedings of the Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  18. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Ioffe, S.; Szegedy, S. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  20. Huang, R.; Zhang, S.; Li, T.; He, R. Beyond Face Rotation: Global and Local Perception GAN for Photorealistic and Identity Preserving Frontal View Synthesis. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2458–2467. [Google Scholar]
  21. Caballero, J.; Ledig, C.; Aitken, A.; Acosta, A.; Totz, J.; Wang, Z.; Shi, W. Real-Time Video Super-Resolution with Spatio-Temporal Networks and Motion Compensation. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  22. Huang, X.; Li, Y.; Poursaeed, O.; Hopcroft, J.; Belongie, S. Stacked Generative Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1866–1875. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, R.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. He, K.; Zhang, X.; Ren, R.; Sun, J. Identity Mappings in Deep Residual Networks. In Proceedings of the European Conference on Computer Vision (ECCV2016), Amsterdam, The Netherlands, 22–29 October 2016; pp. 630–645. [Google Scholar]
  25. Ren, S.; Girshick, R.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  26. Dai, J.; Li, Y.; He, K.; Sun, J. R-FCN: Object Detection via Region-based Fully Convolutional Networks. In Proceedings of the Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  27. Johnson, J.; Alahi, A.; Li, F.F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the European Conference on Computer Vision (ECCV2016), Amsterdam, The Netherlands, 22–29 October 2016; pp. 694–711. [Google Scholar]
  28. Kiasari, M.A.; Moirangthem, D.S.; Lee, M. Generative Moment Matching Autoencoder with Perceptual Loss. In Proceedings of the ICONIP 2017, Guangzhou, China, 14–18 November 2017; pp. 226–234. [Google Scholar]
  29. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D. Generative Adversarial Networks. Adv. Neural Inf. Process. Syst. 2014, 3, 2672–2680. [Google Scholar]
  30. Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv, 2015; arXiv:1511.06434. [Google Scholar]
  31. Dosovitskiy, A.; Brox, T. Generating images with perceptual similarity metrics based on deep networks. In Proceedings of the 2016 Advances in Neural Information Processing Systems (NIPS), Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  32. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
  33. Mathieu, M.; Couprie, C.; Lecun, Y. Deep multi-scale video prediction beyond mean square error. arXiv, 2016; arXiv:1511.05440. [Google Scholar]
  34. Yu, X.; Porikli, F. Ultra-resolving face images by discriminative generative networks. In Proceedings of the European Conference on Computer Vision (ECCV2016), Amsterdam, The Netherlands, 22–29 October 2016; pp. 318–333. [Google Scholar]
  35. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar]
  36. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5967–5976. [Google Scholar]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  38. Lowe, D. Distinctive image features from scale-invariant key points. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
Figure 1. An overview of the proposed DUGAN method. The networks include generative model G and discriminative model D.
Figure 1. An overview of the proposed DUGAN method. The networks include generative model G and discriminative model D.
Sensors 18 01985 g001
Figure 2. Testing results of several denoising methods for UAV images with several ground objects. In each group of images, the 1st image (a1,b1,c1,d1,e1) is the ground truth; the 2nd image (a2,b2,c2,d2,e2) is a noise image with noise level 35; and the 3rd image (a3,b3,c3,d3,e3) presents the denoising results of method [8]. The 4th image (a4,b4,c4,d4,e4) presents the denoising results of method [5]; the 5th image (a5,b5,c5,d5,e5) presents the denoising results of method [17]; and the 6th image (a6,b6,c6,d6,e6) presents the denoising results of the proposed method.
Figure 2. Testing results of several denoising methods for UAV images with several ground objects. In each group of images, the 1st image (a1,b1,c1,d1,e1) is the ground truth; the 2nd image (a2,b2,c2,d2,e2) is a noise image with noise level 35; and the 3rd image (a3,b3,c3,d3,e3) presents the denoising results of method [8]. The 4th image (a4,b4,c4,d4,e4) presents the denoising results of method [5]; the 5th image (a5,b5,c5,d5,e5) presents the denoising results of method [17]; and the 6th image (a6,b6,c6,d6,e6) presents the denoising results of the proposed method.
Sensors 18 01985 g002aSensors 18 01985 g002bSensors 18 01985 g002cSensors 18 01985 g002d
Figure 3. Testing results of several denoising methods for UAV images with several ground objects. In each group of images, the 1st image (a1,b1,c1,d1,e1)is the ground truth; the 2nd image (a2,b2,c2,d2,e2) is a noisy image with noise level 55; and the 3rd image (a3,b3,c3,d3,e3) presents the denoising results of method [8]. The 4th image (a4,b4,c4,d4,e4) presents the denoising results of method [5]; the 5th image (a5,b5,c5,d5,e5) presents the denoising results of method [17]; and the 6th image (a6,b6,c6,d6,e6) presents the denoising results of the proposed method.
Figure 3. Testing results of several denoising methods for UAV images with several ground objects. In each group of images, the 1st image (a1,b1,c1,d1,e1)is the ground truth; the 2nd image (a2,b2,c2,d2,e2) is a noisy image with noise level 55; and the 3rd image (a3,b3,c3,d3,e3) presents the denoising results of method [8]. The 4th image (a4,b4,c4,d4,e4) presents the denoising results of method [5]; the 5th image (a5,b5,c5,d5,e5) presents the denoising results of method [17]; and the 6th image (a6,b6,c6,d6,e6) presents the denoising results of the proposed method.
Sensors 18 01985 g003aSensors 18 01985 g003bSensors 18 01985 g003cSensors 18 01985 g003d
Figure 4. Due to the length concerns, we only show the matching results of 3 images (b,c,e). In each matching image, the left image is the original real clean UAV image, and the right image is the denoised UAV image (noise level 35). The 1st image (b1,c1,e1) presents matching results from SIFT and the denoised UAV images using method [8] and original real clean UAV images. The 2nd image (b2,c2,e2) presents matching results from SIFT and the denoised UAV images using method [5] and the original real clean UAV images. The 3rd image (b3,c3,e3) presents matching results from SIFT and the denoised UAV images created by method [17] and original real clean UAV images; and the 4th image (b4,c4,e4) presents matching results from SIFT and the denoised UAV images using proposed method and the original real clean UAV images.
Figure 4. Due to the length concerns, we only show the matching results of 3 images (b,c,e). In each matching image, the left image is the original real clean UAV image, and the right image is the denoised UAV image (noise level 35). The 1st image (b1,c1,e1) presents matching results from SIFT and the denoised UAV images using method [8] and original real clean UAV images. The 2nd image (b2,c2,e2) presents matching results from SIFT and the denoised UAV images using method [5] and the original real clean UAV images. The 3rd image (b3,c3,e3) presents matching results from SIFT and the denoised UAV images created by method [17] and original real clean UAV images; and the 4th image (b4,c4,e4) presents matching results from SIFT and the denoised UAV images using proposed method and the original real clean UAV images.
Sensors 18 01985 g004aSensors 18 01985 g004b
Figure 5. In each matching image, the left image is the original real clean UAV image, and the right image is the denoised UAV image (noise level 55). The 1st image (b1,c1,e1) presents matching results from SIFT and the denoised UAV images using method [8] and original real clean UAV images. The 2nd image (b2,c2,e2) presents matching results from SIFT and the denoised UAV images using method [5] and the original real clean UAV images. The 3rd image (b3,c3,e3) presents matching results from SIFT and the denoised UAV images created by method [17] and original real clean UAV images; and the 4th image (b4,c4,e4) presents matching results from SIFT and the denoised UAV images using proposed method and the original real clean UAV images.
Figure 5. In each matching image, the left image is the original real clean UAV image, and the right image is the denoised UAV image (noise level 55). The 1st image (b1,c1,e1) presents matching results from SIFT and the denoised UAV images using method [8] and original real clean UAV images. The 2nd image (b2,c2,e2) presents matching results from SIFT and the denoised UAV images using method [5] and the original real clean UAV images. The 3rd image (b3,c3,e3) presents matching results from SIFT and the denoised UAV images created by method [17] and original real clean UAV images; and the 4th image (b4,c4,e4) presents matching results from SIFT and the denoised UAV images using proposed method and the original real clean UAV images.
Sensors 18 01985 g005aSensors 18 01985 g005bSensors 18 01985 g005c
Figure 6. (a) Represent the noisy images of cars and trucks with different noise levels; (b) Represent the denoising results of method [8]; (c) Represent the denoising results of method [5]; (d) Represent the denoising results of method [17]; (e) Represent the denoising results of proposed method.
Figure 6. (a) Represent the noisy images of cars and trucks with different noise levels; (b) Represent the denoising results of method [8]; (c) Represent the denoising results of method [5]; (d) Represent the denoising results of method [17]; (e) Represent the denoising results of proposed method.
Sensors 18 01985 g006
Table 1. Quantitative measurement results using PSNR (dB) on UAV testing images with different noise levels.
Table 1. Quantitative measurement results using PSNR (dB) on UAV testing images with different noise levels.
ImagesMethod [8]Method [5]Method [17]Ours
Noise Level: 20
a30.754330.958131.017431.2741
b30.396530.558330.696730.9458
c31.690731.967831.853432.1033
d29.635829.814529.954330.2104
e30.223530.416930.557630.7562
Average results 130.594230.765730.985631.1123
ImagesNoise Level: 35
a28.816528.696728.887129.0564
b28.308428.575828.431228.9687
c29.487429.732129.727629.9576
d27.517227.695427.895828.1354
e28.291328.478228.618428.7858
Average results28.435828.612728.796828.9854
ImagesNoise Level: 55
a26.186926.311226.498426.7156
b25.927626.116426.365826.5942
c27.196227.333527.479527.6614
d25.495625.644225.784125.9612
e26.051726.265426.312726.5154
Average results26.001426.242426.491226.6576
1 The average results of 40 UAV test images.
Table 2. Quantitative measurement results using SSIM on UAV testing images with different noise levels.
Table 2. Quantitative measurement results using SSIM on UAV testing images with different noise levels.
ImagesMethod [8]Method [5]Method [17]Ours
Noise Level: 20
a0.87990.88180.88410.8857
b0.87760.87970.88040.8823
c0.88570.88970.88950.8912
d0.87040.87230.87410.8765
e0.87310.87630.87790.8806
Average results0.87560.87810.87960.8821
ImagesNoise Level: 35
a0.83450.83620.83790.8406
b0.83090.83390.83520.8371
c0.83980.84410.84330.8468
d0.82310.82490.82710.8289
e0.82760.83030.83190.8343
Average results0.82890.83120.83330.8369
ImagesNoise Level: 55
a0.76820.77010.77190.7745
b0.76390.76610.76790.7701
c0.77390.77920.77980.7819
d0.72450.72710.72890.7315
e0.75890.76180.76410.7658
Average results0.76010.76470.76620.7689
Table 3. The comparison of the correct matching pairs between denoised images and the corresponding clean UAV images obtained by SIFT.
Table 3. The comparison of the correct matching pairs between denoised images and the corresponding clean UAV images obtained by SIFT.
ImagesMethod [8]Method [5]Method [17]Ours
Noise Level 35/55Noise Level 35/55Noise Level 35/55Noise Level 35/55
a110/53114/74127/78150/88
b160/101205/141240/145274/154
c96/66104/70118/76135/78
d136/115138/129157/144192/154
e177/154192/162211/179260/215
Table 4. The comparison of the correct rate of denoised images classification.
Table 4. The comparison of the correct rate of denoised images classification.
Noise LevelCorrect Rate of Method [8]Correct Rate of Method [5]Correct Rate of Method [17]Correct Rate of Ours Method
3567%71%73%78%
5562%65%70%72%

Share and Cite

MDPI and ACS Style

Wang, R.; Xiao, X.; Guo, B.; Qin, Q.; Chen, R. An Effective Image Denoising Method for UAV Images via Improved Generative Adversarial Networks. Sensors 2018, 18, 1985. https://doi.org/10.3390/s18071985

AMA Style

Wang R, Xiao X, Guo B, Qin Q, Chen R. An Effective Image Denoising Method for UAV Images via Improved Generative Adversarial Networks. Sensors. 2018; 18(7):1985. https://doi.org/10.3390/s18071985

Chicago/Turabian Style

Wang, Ruihua, Xiongwu Xiao, Bingxuan Guo, Qianqing Qin, and Ruizhi Chen. 2018. "An Effective Image Denoising Method for UAV Images via Improved Generative Adversarial Networks" Sensors 18, no. 7: 1985. https://doi.org/10.3390/s18071985

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop