Next Article in Journal
Physical Extraction and Feature Fusion for Multi-Mode Signals in a Measurement System for Patients in Rehabilitation Exoskeleton
Previous Article in Journal
A Fiber Bragg Grating (FBG)-Enabled Smart Washer for Bolt Pre-Load Measurement: Design, Analysis, Calibration, and Experimental Validation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Image Super Resolution by Combining Compressive Sensing and Deep Learning

1
University of Chinese Academy of Sciences, Beijing 101408, China
2
Key Laboratory of Space Active Opto-Electronics Technology, Shanghai Institute of Technical Physics of the Chinese Academy of Sciences, Shanghai 200083, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(8), 2587; https://doi.org/10.3390/s18082587
Submission received: 12 June 2018 / Revised: 26 July 2018 / Accepted: 3 August 2018 / Published: 7 August 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
Super resolution methods alleviate the high cost and high difficulty in applying high resolution infrared image sensors. In this paper we present a novel single image super resolution method for infrared images by combining compressive sensing theory and deep learning. Low resolution images can be regarded as the compressed sampling results of the high resolution ones in compressive sensing. With sparsity in this theory, higher resolution images can be reconstructed. However, because of diverse level of sparsity for different images, the output contains noise and loss of high frequency information. Deep convolutional neural network provides a solution to relieve the noise and supplement some missing high frequency information. By concatenating two methods, we manage to produce better results in super resolution tasks for infrared images than SRCNN and ScSR. PSNR and SSIM values are used to quantify the performance. Applying our method to open datasets and actual infrared imaging experiments, we also find better visual results are preserved.

1. Introduction

Nowadays high resolution (HR) images, possessing richer scene information and better visual quality than low resolution (LR) ones, are more desirable in many circumstances. However, the instrumentation limits make the HR images expensive and hard to achieve [1]. This problem is much more severe for infrared (IR) image sensors than visible (VIS) ones. Due to long wave-length, low resolution IR images always suffer from missing details including texture, contexture, edge information, etc. [2]. Less difficulty in optics and sensors manufacturing, super-resolution (SR) method is the most common task that widely used in many areas such as medical imaging [3], remote sensing [4], face recognition [5] and microscopy [6].
SR solutions are grouped into two categories: multi-frame SR (MFSR) and single-image SR (SISR) [7]. For MFSR, a sequence of LR images are captured to compose a HR image using the relative geometric and/or photometric displacements from the target HR image [8]. However, the necessary highly related sequences of images are not often available. In this paper, we focus on single image super resolution (SISR). As it is an inherently ill-posed problem, we have to rely on strong prior information to accomplish the task [9]. Sparsity based methods and learning based methods represent two typical ways of utilizing prior information [10].
Images, as is 2-D signal, exhibit sparsity in some domain, which enables Compressive Sensing (CS) theory to reconstruct the original HR images with LR ones with less sample rate. CS theory has already been proved to be effective and powerful in SR tasks [11]. Many SISR problems have been analyzed under different sparse bases, such as Wavelet [12], Discrete Cosine Transformation (DCT) [13] and Discrete Fourier Transform (DFT) [14]. Recently more practical applications reveal that a signal is more sparse with respect to an over-complete dictionary than a basis [15]. Besides, in order to accurately reconstruct the coefficients of the original signal in sparse domain, optimal reconstruction methods are needed. Different methods possess different performance, while we choose the iteratively reweighted least squares (IRLS) as the optimal reconstruction algorithm for its high reconstruction performance through experiments [16]. Its mechanism will be discussed later in this paper.
Apart from sparsity-based methods, learning-based ones also benefit from prior information. As deep-learning has recently prospered, many learning-based algorithms have been used in SISR, such as VGG [17], ResNet [18] and GAN [19]. Initially SRCNN [20] was the first 3-layered Convolutional Neural Network (CNN) utilized for SR tasks. Lately, for better performance the network structure goes much deeper. Besides, residual learning networks, when used in SR tasks, have been proved to possess better visual performance and Peak Signal to Noise Ratio (PSNR) performance.
In recent several years, many researchers have tried to combine CS and deep learning to produce better SR task solutions. Duan et al. [21] used deep learning to capture the image features and apply them to reconstruct HR images with the help of the sparsity in CS. Bora et al. [22] used generative models to replace the sparsity bases in CS and achieve satisfying results.
In this paper, we provide a novel combination architecture. We take advantage of sparsity in CS to recover the high frequency information in HR images. Then we build a deep-layer CNN to promote the performance of IRLS in CS. Residual learning [23] ensures that with our algorithm it is easier to optimize the results by denoising and reconstructing the output image of CS. By concatenating the two methods we achieve better performance than SRCNN [20] and ScSR [24] that utilize sparsity and a neural network alone. In simulations and actual infrared imaging experiments, we apply our method to IR images and we verify its performance both visually and quantitatively.

2. Super-Resolution Framework

2.1. Super-Resolution with Compressive Sensing Theory

CS theory combines sampling and compression into non-adaptive linear measurement process [25] at a rate significantly below the Nyquist [26]. The classical CS acquisition process can be depicted as:
y = Φ x = Φ Ψ s = θ s .
Here y M is the vector of stacking measurements. x N ( M < N ) is the original compressible signal. Φ is the M × N measurement matrix and θ Φ Ψ where Ψ is the N × N basis matrix. Vector s is the coefficients of x in the Ψ domain. Usually a Gaussian random matrix will be used as Φ . In SISR tasks, y will be regarded as the low projection of the HR image x , and Φ is corresponding to a downsample matrix in SISR [13]. Referring to the binning process of image sensors [27], we believe that one pixel in a LR image equals to the average of corresponding k × k neighbor pixels in HR one. Therefore Φ with M × N dimension, where N / M = k 2 , should function as this downsampling process [28].
x in the spatial domain can be represented by vector s in the Ψ domain, which is K -sparse ( K < N coefficients in s are non-zero). With sufficient sampling rate, s will be correctly recovered from Equation (1) by solving such an l p -norm optimization problem:
min s 1 2 s p p ,   s . t . y = Φ x = Φ Ψ s = θ s .
Ψ , the sparsity basis, has been widely proved validity using wavelet basis [29]. In our algorithm, we utilize DCT basis instead, because of its better performance under numerous experimental conditions. In this paper, we will use Peak Signal to Noise Ratio (PSNR) and structural similarity index (SSIM) to quantify the performance of the SR method. After testing on widely used 400 images [30] of size 180 × 180 , we find that on average the DCT basis outperforms the wavelet basis by 14% higher in PSNRs and 26% higher in SSIMs.
Corresponding to the basis, the reconstruction algorithm is also important for our SR tasks. In order to solve this underdetermined equation finding accurate x , many optimization methods have been developed these years, such as Orthogonal Matching Pursuit (OMP) [31], Subspace Pursuit [32], Relevance Vector Machine (RVM) [33] and Iteratively reweighted least squares (IRLS) [16]. Iteratively reweighted least squares (IRLS) is selected for better visual and quantitative results, where p = 1 .
The IRLS method we use is based on solving (2) with modified objective function that at each iteration the function approaches k = 1 N | s | p [27]. Simply, we substitute the 𝓁 p objective function in (2) with a weighted 𝓁 2 norm:
min s i = 1 N w i s i 2 ,   s . t . y = Φ x = Φ Ψ s = θ s .
where w i = | s i ( n 1 ) | p 2 is the first-order approximation to the 𝓁 p objective function. w i changes at each iteration until w i s i 2 is sufficiently close to s p p in (4) after convergence. Then the solution of (3) is:
s ( n 1 ) = Q n θ T ( θ Q n θ T ) 1 y ,
where Q n is the diagonal matrix with entries:
1 / w i = | s i ( n 1 ) | 2 p .
The convergence criterion for each iteration stage can be depicted as:
s n s n 1 1 + s n 1 < μ 100 .
After (6) is attained, μ is reduced by a factor of 10, and the iterative procedure is repeated until μ < 10 8 [34].
In conclusion, the HR image x can be depicted by sparse vector s in Ψ domain. The input of the algorithm, the original LR image, is regarded as the compressed measurements. Finally, x can be resolved with reconstruction algorithm. The detailed parameters in this algorithm are demonstrated in Algorithm 1.
Algorithm 1. IRLS Method for Super-Resolution
Parameters: p = 1, use DCT basis as Ψ , down-sampling matrix Φ , N/M = 2 or 3, μ = 1 .
Step 1: Initialize the size of output image and the formation of sparsity basis.
Step 2: Do the inner loop:
2.1 Initialize n 1 , s 0 = ( 0 , 0 , 0 ) and Q ( 0 ) = O.
2.2 Update Q ( n ) using (5).
2.3 Compute s n using (4).
2.4 If (6) is satisfied, go to step 3; otherwise, let n = n + 1 and go to step 2.2.
Step 3: Update the regularization parameter, μ = μ / 10 .
Step 4: If μ < 10 8 , finish; else, go to Step 2.

2.2. Image Denoising and Reconstruction with Deep Learning

Practically, it is hard to find an absolutely correct y , which represents the HR image in SR tasks. Mostly the algorithms will come to a local optimal solution that makes the output images contain fixed pattern noise, which is illustrated in Figure 1. By comparing the output of CS, bicubic method and the original HR image, visually we find that CS preserves more texture information and less blur effect, but contains some fixed pattern noise. After using our CNN, the noise is visually alleviated. Although the PSNR of CS output is 0.64 dB higher than bicubic, the SSIM of CS is 0.031 lower. As SSIM calculates the covariance value of the images representing the structural information of the objects in images [35], studies show that it is more vulnerable to fixed-pattern noise than pixel difference-based measurement, PSNR [36]. Therefore a method that protects the high spatial frequency information while wiping out the fixed pattern noise is necessary. After using our CNN, the structure of which will be discussed in the following paragraph, the PSNR is increased to 34.18 dB and the SSIM is increased to 0.9719. These values proves that our CNN is effective in denoising and reconstruction.
From the results, we believe that the CNN not only deals with the fixed pattern noise, but also helps supplement more high frequency information. As the images change, the level of sparsity changes as well. Some HR images may contain more high frequency information that won’t be recovered by a certain sparsity basis, causing the limits of the CS method, which means using CS alone won’t recover all the high frequency information. In that case, we also need more efforts to supplement the missing information during the SR process. Deep learning with powerful image processing ability has been applied to many tasks like image denoising, demosaicing [37] and reconstruction [38]. Zhang et al. [29] designed a deep convolutional neural network (CNN) for image Gaussian denoising, which is called DnCNN. Residual learning and batch normalization greatly benefit its performance. Inspired by DnCNN, we modified its network architecture to accomplish the denoising and high frequency information supplementation in our SR tasks.
The most essential part of our CNN model is the residual learning. Although the output of CS contains fixed pattern noise, we are not able to describe its formation with a designed rule in order to eliminate it. However deep learning provides us with trainable convolutional filter, in which case the noise of each HR image can be detected and eliminated after training the CNN model. Residual learning enables us to train each layer of CNN to fit the residual mapping instead of the original image. Formally, we denote the HR output of CS as H ( j ) , and the original HR image, which is the ground truth, as G ( j ) . Here j denotes the index of each image. The residual image R ( j ) = G ( j ) H ( j ) , represents the fixed pattern noise of each image. Researches have revealed that residual image is easier to be optimized by CNN [23]. Figure 2 shows the proposed SR architecture when training.
The target of our CNN is to estimate the residual image of every CS output for promoting the performance. The averaged mean square error between estimated residual image and the true residual one:
l ( Θ ) = 1 2 N i = 1 N R ˜ ( Θ , i ) R ( Θ , i ) 2 ,
denotes the loss function to learn the trainable parameters Θ in CNN. Corresponding to i th training image, R ˜ ( Θ , i ) represents the estimated residual image produced by our CNN, while R ( Θ , i ) represents the true residual image used for training.
Researches reveal that the depth of network is of great importance for better results [23]. Therefore, we challenge to modify the CNN into a deeper network with 30 layers. Inspired by DnCNN, our network consists of three types of layer, which is shown in Figure 3.
In the first layer, we utilize 64 filters of 3 × 3 size as the convolution kernels to generate 64 feature maps. And rectified linear units (ReLU, m a x ( 0 , · ) ) are utilized as the nonlinear activation function for speeding up the optimization. The 28 hidden layers are of the same formation. 64 filters of size 3 × 3 × 64 are connected with batch-normalization (BN) [39] in the hidden layers for accelerating training speed. For the last layer, a 3 × 3 × 64 convolution is used for reconstructing the residual image.
After simulation experiments, we find that Adaptive Moment Estimation (Adam) optimization [40] algorithm outperforms Stochastic Gradient Descent (SGD) [41]. Therefore, we choose Adam as the optimization method for our CNN. Adam is a first-order gradient-based optimization algorithm, which is based on adaptive estimates of lower-order moments of the gradients. The pseudo-code is shown in Algorithm 2.
Algorithm 2. Adam Method for Optimization
Parameters: α is the stepsize; β 1 , β 2 [ 0 , 1 ) , λ [ 0 , 1 ) are the exponential decay rates for the moment estimates; l ( Θ ) is the loss function with parameter Θ .
Step 1: Initialize the parameters as β 1 = 0.9 , β 2 = 0.999 , λ = 1 10 8 , α = 0.001 .
Step 2: Initialize the vectors.
m 0 0 is the initial first moment vector.
v 0 0 is the initial second moment vector.
t 0 is the initial timestep.
Step 3: Do the inner loop:
3.1 t t + 1 . Update the timestep.
3.2 β 1 , t β 1   λ t 1 . Decay the first moment running average coefficient.
3.3 g t θ f t ( θ t 1 ) . Get gradients corresponding to loss function at timestep t.
3.4 m t β 1 , t · v t 1 + ( 1 β 1 , t ) · g t . Update biased first moment estimate.
3.5 v t β 2 · v t 1 + ( 1 β 2 ) · g t g t . Update biased second raw moment estimate.
3.6 m t ^ m t / ( 1 β 1 , t ) . Compute bias-corrected first moment estimate.
3.7 v t ^ v t / ( 1 β 2 ) . Compute bias-corrected second raw moment estimate.
3.8 Θ t Θ t 1 α · m t ^ / ( v t ^ + ϵ ) . Update parameters, where ϵ is for preventing the denominator to be zero.
3.9 if Θ t is converged, go to step 4; otherwise go to step 3.1.
Step 4: Return Θ t .
Most parameters of Adam are set the same as the ones in [40], as the mini-batch size is 128 and the learning rate decays exponentially from 1 × 10−1 to 1 × 10−4 during 50 epochs of training.
We use the MatConvNet package in Matlab 2017a to train our CNN. A Intel® CoreTM i5-4670k CPU operating at 3.4 GHz and an Nvidia 1080Ti GPU are used. Experiments show that the deeper the network goes, the better PSNR performance becomes, as is shown in Figure 4. However, for a deep network of 30 layers and 128 mini-batch size, a great burden has been placed on the GPU memory. 30 layers with 128 mini-batch size is up to the limit of the GPU memory.

2.3. The Whole Super-Resolution Algorithm Architecture

In Figure 5 we show the whole architecture when using the proposed method to accomplish the SR target. After training process, the CNN is used to eliminate the fixed pattern noise in the output of CS SR method and supplement some high spatial frequency information to it.

3. Simulation Results

Before applying our method to real scenes captured by infrared sensors, we test it with some open datasets by comparing it with SRCNN [20] and ScSR [24] that utilize sparsity and neural network alone. Considering that there are not enough open image data sets for training at infrared wavelengths, we choose widely used 400 VIS images [29] of size 180 × 180 as the training dataset. The experimental results show that the model trained by VIS dataset functions well when dealing with IR images. A larger training dataset is more preferable, but leads to more training time pressure. After testing we find that 400 images are enough to get high performance, and the training time is acceptable. About 10 h for training is needed for our CNN. This trained model in VIS is used for super resolution tasks in VIS images and IR images.
We apply our method to six infrared images collected from the OSU thermal pedestrian database, OSU Color and Thermal Database and Terravic Motion Infrared Database of the OTCBVS dataset collection [42]. Besides, we also apply our method to six widely used VIS images to prove the robustness. Figure 6 shows the overview of the 12 total images regarded as the test set. It is worth highlighting that the training set should not share the same images with the test set in order to avoid a logical paradox. Therefore the 12 IR and VIS images are not included in the 400 images for training.
In this paper the upscaling factors are set 2 and 3. We down-sample the HR image into two LR one by merging 2 × 2 or 3 × 3 neighbor pixels on average in order to simulate two kinds of LR images. We compare the SR images with the original HR ones by quantifying the performance in PSNR and SSIM, the results of which is shown in the Table 1 and Table 2. Besides the execution time is also provided in the tables for considering the complexity of our algorithm.
Before discussing the SR reconstruction performance, the execution time of three methods also attracts great interest. SRCNN exhibits the least time consumption, while ScSR and our algorithm need far more execution time. Most time of our algorithm is spent on solving the optimization problem for compressive sensing architecture in (2). This is because the time complexity of IRLS is high, despite its better accuracy in reconstruction. Another fact that draws great attention is that our algorithm needs far less time for SR of upscaling factor of 3 than of 2, unlike ScSR and SRCNN. The reason is that LR images produced by merging 3 × 3 neighbor pixels from HR ones contain lower spatial resolution and less amount of information than those produced by merging 2 × 2 pixels, which means fewer constraint conditions in (3) and fewer dimensions of vector s in the Ψ domain in (3). After fewer iterations, IRLS will comes to the nearly accurate answers to get HR estimation. Therefore, we may predict that for even larger upscaling factors, our algorithm may perform much better in execution time.
We find that the proposed method has great advantages in PSNR values, while performing a little better than SRCNN and ScSR in SSIM values. We choose image 6, the infrared surveillance, in the test set as an example to show the performance visually. Figure 7 illustrates the visual comparison of three methods. The original HR image is of 360 × 240 pixels. After down-sampling, two kinds of LR image images are produced, which are of 180 × 120 pixels and of 120 × 80 pixels. We produce the SR images with SRCNN, ScSR and our method. The zoomed HR images are placed on the right.
The whole images comparison provides us the overall perception of different methods, where the texture feature in our method appears to be clearer. Moreover, in our results, the surroundings near the objects are of less distraction and less noise. From the zoomed images, we find that the edges in the output of our method are more distinct. In details, the contours of the zebra crossing in our method possesses higher fidelity and higher contrast compared to the one in SRCNN and ScSR. We believe this advantage may help a lot in further image recognition tasks.

4. Imaging Experiments

In this section, we apply our method to an infrared image sensor to testify the portability and generality. As demonstrated in Figure 8, we use MARS-VLW-RM4 from the Sofradir Company (Palaiseau, France) as the infrared image sensor, whose original resolution is 320 × 256 . Its sensitivity to infrared radiation in the Very Long-Wave band (8–12 μm) make ensure its applicability for military and civilian surveillance purposes. However due to the high cost of manufacturing, it is difficult to increase the resolution. Using CS theory and deep learning, we are able to produce higher resolution infrared images without changing the original sensor.
The parameters and trained models are the same as the ones in the simulation section. As lack of ground truth for HR infrared images, the performance will be judged visually in this section. With upscaling factor of 2 and 3, we will produce HR images of 640 × 512 and 960 × 768 resolution. The results are shown in Figure 9.
Visual comparison between the LR and HR images and the 3 different methods demonstrate the advantage of our method. In HR ones, image details, like textures and contours, are more sufficient and mosaic effects caused by LR image sensor are relieved. Therefore, higher resolution infrared images which surpass the original image sensor’s resolution are available by using our method. Moreover, compared to zoomed images of ScSR, our results contain less blur and sharper features. As to SRCNN, its reconstruction noise of the windowsill in the zoomed images shows its inferiority to our method.

5. Conclusions

In this paper we present a novel super resolution method that is the combination of compressive sensing theory and deep learning. Our method consists of two parts. The first one utilizes the spatial sparsity of CS theory to reconstruct a HR image which contains higher frequency information. The second part uses the trained network to remove the fixed pattern noise that was introduced in the first part and supplement some additional high frequency information which is learnt from the training set. Its high performance helps us to acquire higher resolution infrared images without suffering from the high cost and difficulty in applying large infrared sensors. The performance has been demonstrated visually and quantitatively in the simulation tasks. Our method possesses better performance with higher PSNR and SSIM values than SRCNN and ScSR in both visible and infrared datasets. We apply our method to a Very-Long-Wave band infrared sensor to testify its portability and generality. With low resolution infrared sensor, we are able to produce higher resolution images.
As our work only analyzes the monochrome images, we expect more studies will focus on spectral images’ super-resolution problems.

Author Contributions

Conceptualization, X.Z. and J.W.; Methodology, X.Z.; Software, X.Z.; Validation, X.Z., C.L. and Q.M.; Formal Analysis, C.L.; Investigation, Q.M. and Y.Z.; Resources, Y.Z.; Data Curation, S.L.; Writing-Original Draft Preparation, X.Z.; Writing-Review & Editing, X.Z.; Visualization, X.Z.; Supervision, J.W.; Project Administration, J.W.; Funding Acquisition, J.W.

Funding

This work is supported by the Chinese Academy of Sciences Innovation Fund (grant CXJJ-16S054).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shi, W.; Caballero, J.; Huszar, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar] [CrossRef]
  2. Liu, F.; Han, P.; Wang, Y.; Li, X.; Bai, L.; Shao, X. Super resolution reconstruction of infrared images based on classified dictionary learning. Infrared Phys. Technol. 2018, 90, 146–155. [Google Scholar] [CrossRef]
  3. Greenspan, H. Super-Resolution in Medical Imaging. Comput. J. 2009, 52, 43–63. [Google Scholar] [CrossRef]
  4. Shen, H.; Ng, M.K.; Li, P.; Zhang, L. Super-resolution reconstruction algorithm to modis remote sensing images. Comput. J. 2008, 52, 90–100. [Google Scholar] [CrossRef]
  5. Gunturk, B.K.; Batur, A.U.; Altunbasak, Y.; Rd, H.M.; Mersereau, R.M. Eigenface-domain super-resolution for face recognition. IEEE Trans. Image Process. 2003, 12, 597. [Google Scholar] [CrossRef] [PubMed]
  6. Quan, T.; Li, P.; Long, F.; Zeng, S.; Luo, Q.; Hedde, P.N.; Nienhaus, G.U.; Huang, Z. Ultra-fast, high-precision image analysis for localization-based super resolution microscopy. Opt. Express 2010, 18, 11867–11876. [Google Scholar] [CrossRef] [PubMed]
  7. Nasrollahi, K.; Moeslund, T.B. Super-resolution: A comprehensive survey. Mach. Vis. Appl. 2014, 25, 1423–1468. [Google Scholar] [CrossRef]
  8. Li, X.; Hu, Y.; Gao, X.; Tao, D.; Ning, B. A multi-frame image super-resolution method. Signal Process. 2010, 90, 405–414. [Google Scholar] [CrossRef]
  9. Kim, K.I.; Kwon, Y. Single-image super-resolution using sparse regression and natural image prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1127–1133. [Google Scholar] [CrossRef] [PubMed]
  10. Sreeja, S.J.; Wilscy, M. Single image super-resolution based on compressive sensing and TV minimization sparse recovery for remote sensing images. In Proceedings of the 2013 IEEE Recent Advances in Intelligent Computational Systems, Trivandrum, India, 19–21 December 2013; pp. 215–220. [Google Scholar] [CrossRef]
  11. Kulkarni, N.; Nagesh, P.; Gowda, R.; Li, B. Understanding compressive sensing and sparse representation-based super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 778–789. [Google Scholar] [CrossRef]
  12. Sen, P.; Darabi, S. Compressive image super-resolution. In Proceedings of the 2009 Conference Record of the Forty-Third Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 1–4 November 2009; pp. 1235–1242. [Google Scholar] [CrossRef]
  13. Yang, S.; Wang, M.; Sun, Y.; Sun, F.; Jiao, L. Compressive sampling based single-image super-resolution reconstruction by dual-sparsity and non-local similarity regularizer. Pattern Recognit. Lett. 2012, 33, 1049–1059. [Google Scholar] [CrossRef]
  14. Bertocco, M.; Frigo, G.; Narduzzi, C.; Tramarin, F. Resolution enhancement in harmonic analysis by compressive sensing. In Proceedings of the IEEE International Workshop on Applied Measurements for Power Systems, Aachen, Germany, 25–27 September 2013; pp. 40–45. [Google Scholar] [CrossRef]
  15. Baraniuk, R.; Foucart, S.; Needell, D.; Plan, Y.; Wootters, M. One-bit compressive sensing of dictionary-sparse signals. Inf. Inference A J. IMA 2018, 7, 83–104. [Google Scholar] [CrossRef]
  16. Chartrand, R.; Yin, W. Iteratively reweighted algorithms for compressive sensing. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March–4 April 2008; pp. 3869–3872. [Google Scholar] [CrossRef]
  17. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar] [CrossRef]
  18. Li, D.; Wang, Z. Video superresolution via motion compensation and deep residual learning. IEEE Trans. Comput. Imaging 2017, 3, 749–762. [Google Scholar] [CrossRef]
  19. Ledig, C.; Theis, L.; Huszar, F.; Caballero, J.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; Shi, W. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar] [CrossRef]
  20. Dong, C.; Chen, C.L.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the EVVC 2014—European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  21. Duan, G.; Hu, W.; Wang, J. Research on the natural image super-resolution reconstruction algorithm based on compressive perception theory and deep learning model. Neurocomputing 2016, 208, 117–126. [Google Scholar] [CrossRef]
  22. Bora, A.; Jalal, A.; Price, E.; Dimakis, A.G. Compressed sensing using generative models. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 537–546. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  24. Li, X.; Lu, X.; Yuan, H.; Yan, P.; Yuan, Y. Geometry constrained sparse coding for single image super-resolution. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 1648–1655. [Google Scholar] [CrossRef]
  25. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef] [Green Version]
  26. Boufounos, D.; Liu, D.; Boufounos, P.T. A lecture on compressive sensing. IEEE Signal Process. Mag. 2007, 24, 1–9. [Google Scholar]
  27. Nasibov, H.; Kholmatov, A.; Akselli, B.; Nasibov, A.; Baytaroglu, S. Performance analysis of the CCD pixel binning option in particle-image velocimetry measurements. IEEE/ASME Trans. Mechatron. 2010, 15, 527–540. [Google Scholar] [CrossRef]
  28. Lu, F.; Au, O.C. Novel 2-D MMSE subpixel-based image down-sampling for matrix displays. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing, Dallas, TX, USA, 14–19 March 2010; pp. 986–989. [Google Scholar] [CrossRef]
  29. Fan, N. Super-resolution using regularized orthogonal matching pursuit based on compressed sensing theory in the wavelet domain. In Proceedings of the International Conference on Computer Graphics, Tianjin, China, 11–14 August 2009; pp. 349–354. [Google Scholar] [CrossRef]
  30. Zhang, K.; Chen, Y.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed]
  31. Tropp, J.A.; Gilbert, A.C. Signal recovery from random measurements via orthogonal matching pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef]
  32. Dai, W.; Milenkovic, O. Subspace Pursuit for Compressive Sensing Signal Reconstruction. IEEE Trans. Inf. Theory 2009, 55, 2230–2249. [Google Scholar] [CrossRef] [Green Version]
  33. He, L.; Carin, L. Exploiting structure in wavelet-based Bayesian compressive sensing. IEEE Trans. Signal Process. 2009, 57, 3488–3497. [Google Scholar] [CrossRef]
  34. Miosso, C.J.; Borries, R.V.; Argaez, M.; Velazquez, L.; Quintero, C.; Potes, C.M. Compressive sensing reconstruction with prior information by iteratively reweighted least-squares. IEEE Trans. Signal Process. 2009, 57, 2424–2431. [Google Scholar] [CrossRef]
  35. Al-Najjar, Y. Comparison of image quality assessment: PSNR, HVS, SSIM, UIQI. Int. J. Sci. Eng. Res. 2012, 3, 1. [Google Scholar]
  36. Megha, G.; Yashpal, L.; Vivek, L. Analytical relation & comparison of PSNR and SSIM on babbon image and human eye perception using matlab. Int. J. Adv. Res. Eng. Appl. Sci. 2015, 4, 108–119. [Google Scholar]
  37. Satya, K.; JayaChandra, T. Deep learning approach for image denoising and image demosaicing. Int. J. Comput. Appl. 2017, 168, 18–26. [Google Scholar] [CrossRef]
  38. Rivenson, Y.; Zhang, Y.; Günaydın, H.; Teng, D.; Ozcan, A. Phase recovery and holographic image reconstruction using deep learning in neural networks. Light Sci. Appl. 2018, 7, 17141. [Google Scholar] [CrossRef]
  39. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  40. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  41. Bottou, L. Large-scale machine learning with stochastic gradient descent. In Proceedings of the 19th International Conference on Computational Statistics, Paris, France, 22–27 August 2010; pp. 177–186. [Google Scholar] [CrossRef]
  42. OTCBVS Benchmark Dataset Collection. Available online: http://vcipl-okstate.org/pbvs/bench/ (accessed on 12 June 2018).
Figure 1. Illustration of fixed pattern noise in CS method when upscaling factor is set 2 for demonstration. Subfigure (a) is the output of bicubic SR method; subfigure (b) is the output of CS; subfigure (c) is the reconstructed output of CNN; subfigure (d) is the original HR image. The corresponding zoomed pictures are placed on the right. The PSNR and SSIM of CS are 28.86 dB and 0.8950. The PSNR and SSIM of bicubic are 28.22 dB and 0.9260. The PSNR and SSIM of CNN are 34.18 dB and 0.9719.
Figure 1. Illustration of fixed pattern noise in CS method when upscaling factor is set 2 for demonstration. Subfigure (a) is the output of bicubic SR method; subfigure (b) is the output of CS; subfigure (c) is the reconstructed output of CNN; subfigure (d) is the original HR image. The corresponding zoomed pictures are placed on the right. The PSNR and SSIM of CS are 28.86 dB and 0.8950. The PSNR and SSIM of bicubic are 28.22 dB and 0.9260. The PSNR and SSIM of CNN are 34.18 dB and 0.9719.
Sensors 18 02587 g001
Figure 2. The residual learning progress of our method.
Figure 2. The residual learning progress of our method.
Sensors 18 02587 g002
Figure 3. The architecture of the CNN.
Figure 3. The architecture of the CNN.
Sensors 18 02587 g003
Figure 4. The average PSNR & SSIM upscaling factor 2 & 3 versus different layers CNNs under 50-epoch training when applied to the 12-image test set.
Figure 4. The average PSNR & SSIM upscaling factor 2 & 3 versus different layers CNNs under 50-epoch training when applied to the 12-image test set.
Sensors 18 02587 g004
Figure 5. The whole architecture of the proposed SR method.
Figure 5. The whole architecture of the proposed SR method.
Sensors 18 02587 g005
Figure 6. The 12 IR and VIS images used for performance evaluation.
Figure 6. The 12 IR and VIS images used for performance evaluation.
Sensors 18 02587 g006
Figure 7. Visual comparison of SRCNN, ScSR and the proposed method with upscaling factors of 2 and 3.
Figure 7. Visual comparison of SRCNN, ScSR and the proposed method with upscaling factors of 2 and 3.
Sensors 18 02587 g007aSensors 18 02587 g007b
Figure 8. The imaging structure of our infrared system. MARS-VLW-RM4 from Sofradir Company is the infrared image sensor of 320 × 256 resolution. The lens is suitable for Very Long-Wave band of f = 60 mm and F = 2.0 from Lenstech Company in Beijing, China.
Figure 8. The imaging structure of our infrared system. MARS-VLW-RM4 from Sofradir Company is the infrared image sensor of 320 × 256 resolution. The lens is suitable for Very Long-Wave band of f = 60 mm and F = 2.0 from Lenstech Company in Beijing, China.
Sensors 18 02587 g008
Figure 9. Imaging results comparison of the SR output.
Figure 9. Imaging results comparison of the SR output.
Sensors 18 02587 g009aSensors 18 02587 g009b
Table 1. SR results with upscaling factor of 2.
Table 1. SR results with upscaling factor of 2.
ImageSRCNNScSRProposed Method
PSNRSSIMTime/sPSNRSSIMTime/sPSNRSSIMTime/s
133.030.95292.533.290.966217.634.780.96299.6
236.780.96331.636.340.968517.738.080.968912.1
334.420.97001.734.260.971821.134.650.970215.2
440.590.97691.641.360.979320.941.530.978615.4
535.340.96521.735.930.969120.936.100.967815.2
630.630.81181.830.340.814124.431.080.815414.1
728.200.90051.527.560.894017.729.480.911114.5
832.660.93981.431.750.933317.333.720.946453.0
932.510.96181.530.840.952016.734.180.971957.6
1028.550.91801.528.410.916917.929.590.925457.2
1136.190.93819.835.840.935370.136.740.938056.8
1232.980.920110.232.440.914769.333.550.926554.0
Table 2. SR results with upscaling factor of 3.
Table 2. SR results with upscaling factor of 3.
ImageSRCNNScSRProposed Method
PSNRSSIMTime/sPSNRSSIMTime/sPSNRSSIMTime/s
128.220.86661.827.420.885147.628.730.90015.0
232.060.92221.431.500.919451.332.420.92594.2
329.470.90451.528.360.907052.229.270.91174.2
436.590.94521.435.610.954253.537.300.95284.0
530.930.90111.530.890.909652.530.570.90454.8
628.480.71341.627.970.712760.528.720.72165.1
726.530.84271.326.110.834245.227.240.85965.1
830.440.91171.228.690.897743.031.180.926918.7
929.040.91051.326.940.883544.830.270.933422.3
1026.120.86931.325.750.864046.027.080.883822.4
1133.400.90979.232.660.9041183.333.480.912422.7
1230.790.86369.830.140.8543186.731.170.873122.4

Share and Cite

MDPI and ACS Style

Zhang, X.; Li, C.; Meng, Q.; Liu, S.; Zhang, Y.; Wang, J. Infrared Image Super Resolution by Combining Compressive Sensing and Deep Learning. Sensors 2018, 18, 2587. https://doi.org/10.3390/s18082587

AMA Style

Zhang X, Li C, Meng Q, Liu S, Zhang Y, Wang J. Infrared Image Super Resolution by Combining Compressive Sensing and Deep Learning. Sensors. 2018; 18(8):2587. https://doi.org/10.3390/s18082587

Chicago/Turabian Style

Zhang, Xudong, Chunlai Li, Qingpeng Meng, Shijie Liu, Yue Zhang, and Jianyu Wang. 2018. "Infrared Image Super Resolution by Combining Compressive Sensing and Deep Learning" Sensors 18, no. 8: 2587. https://doi.org/10.3390/s18082587

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop