Next Article in Journal
A Survey of Blockchain Enabled Cyber-Physical Systems
Previous Article in Journal
Impact of SCHC Compression and Fragmentation in LPWAN: A Case Study with LoRaWAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RGB-IR Cross Input and Sub-Pixel Upsampling Network for Infrared Image Super-Resolution

1
School of Physics and Optoelectronic Engineering, Xidian University, No. 2 South Taibai Road, Xi’an 710071, China
2
Research and Development of Infrared Detection Technology, China Aerospace Science and Technology Corporation, Shanghai 201109, China
3
Shanghai Aerospace Control Technology Institute, Shanghai 201109, China
4
National Institute of Informatics, Tokyo 101-8430, Japan
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(1), 281; https://doi.org/10.3390/s20010281
Submission received: 3 December 2019 / Revised: 22 December 2019 / Accepted: 28 December 2019 / Published: 3 January 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Deep learning-based image super-resolution has shown significantly good performance in improving image quality. In this paper, the RGB-IR cross input and sub-pixel upsampling network is proposed to increase the spatial resolution of an Infrared (IR) image by combining it with a color image of higher spatial resolution obtained with a different imaging modality. Specifically, this is accomplished by fusion of the features map of two RGB-IR inputs in the reconstruction of an infrared image. To improve the accuracy of feature extraction, deconvolution is replaced by sub-pixel convolution to upsample image in the network. Then, the guided filter layer is introduced for image denoising of IR images, and it can preserve the image detail. In addition, the experimental dataset, which is collected by us, contains large numbers of RGB images and corresponding IR images with the same scene. Experimental results on our dataset and other datasets demonstrate that the method is superior to existing methods in accuracy and visual improvement.

1. Introduction

Image super-resolution (SR) has become a hot research topic to improve the image resolution by means of software algorithm. The key is to obtain an estimation of a High Resolution (HR) image from the Low Resolution (LR) input. Usually, Infrared (IR) detectors can be used for night video monitoring, biomedicine, forest fire fighting and safe driving. However, IR images captured by infrared imaging devices always suffer from low resolution, low contrast and blur details. Therefore, it is more difficult to extract information of interesting objects from IR images than that of color images.
Traditional research on image super-resolution mainly focuses on two aspects. One is trying to find the mapping relationship between HR and LR images with respect to pixels or local patches, which always involves Neighbor Embedding (NE) and Anchored Neighborhood Regression (ANR). Hong Chang introduced Neighbor Embedding [1], which can be reconstructed by its neighbors in the feature space and not highly depend on sample. Howerver, the number of fixed neighborhoods would cause overfitting or under-fitting problems. Radu Timofte, et al. proposed Anchored Neighborhood Regression [2,3] which takes the dictionary atom as the neighborhood to reduce the operational complexity and running time, but it loses flexibility. The other kind of method is sparse coding, which aims to learn the similarity between LR and HR patches via large datasets. Jianchao Yang presented the sparse coding algorithm  [4,5] via patch-based sparse coding and two dictionaries to achieve the image super resolution, which has great improvement in SR (super-resolution) quality. However, the dictionary is not complete and the image edge quality is not high. In summary, the traditional algorithm has limited SR effect on image super-resolution, and its speed is not fast.
Recently, end-to-end learning methods were successfully applied into the field of SR, object detection, obstacle recognition [6]. Dong et al. firstly combined the Convolutional Neural Network (CNN) method with SR, and proposed the Super Resolution Convolution Neural Network (SRCNN [7]) algorithm and Fast SRCNN (FSRCNN [8]), with better performance than traditional algorithms. J. Kim et. al proposed the Very Deep Convolutional Networks (VDSR [9]), which has 20 convolution layers to further exploit context information. To avoid the overfitting of deep networks, J. Kim et al. introduced the Deeply-Recursive Convolutional Network (DRCN [10]) with residual structure and skip-connections. In the last two years, many modules were dated to solve low resolution problems. Jiayi Ma [11] presented a dense discriminative network that is composed of several aggregation modules (AM) and aggregate features progressively in an efficient way. Still, the above algorithm deep learning network can only process a single data source, and can only use the characteristics of a single sensor. There are certain limitations to enhancing details. There are also some limitations can only handle a single sensor image.
Meanwhile, various approaches via multi-source data (infrared image and color image) were proposed to increase the image high-frequency information as much as possible [12]. To solve the problems of lighting, T. Y. Han [13] proposed a method which fuses multiple LR images taken at different camera positions to synthesize a HR image. C. W. Tseng [14] proposed a network in which infrared images and color images are concatenated as input to construct a high-resolution IR image. However, as single-input networks, these algorithms ignore the high-frequency information of visible images captured by visible detectors. In addition, most networks have not got trained network models that take paired infrared and visible images. To solve this problem, we propose a network to mine information of visible image for high-resolution images reconstruction. The network has multi-modal input of infrared and visible light images to make full use of the feature information of multiple sensors.
Infrared detectors can clearly display target images at night or partially occluded. In the limitation to the hardware system of Infrared detectors, the infrared images have a few problems: low resolution, blurring, and random noise. Apart from training convolutional networks with multi-sensors images as input, the key to achieving infrared image super-resolution is the removal of infrared noise. In this paper, we applied guided filter [15] to suppress the noise in IR images.
The structure of our RGB-IR cross input and sub-pixel up-sampling network is shown in Section 3. The main contributions are listed as follows: (1) Paired IR and RGB images are used as inputs, unlike most existing works which only take single-sensor images. (2) Sub-pixel convolution is applied to extract multi-channel features from RGB images, which is used to enhance the IR image. (3) Guided filter layer is adopted to reduce the influence of noise in IR images, which makes the IR image reconstruction effective. (4) Dataset of IR-COLOR2000 is captured by ourselves for training ideal network.
The specific operation is as follows. First, a guided filter layer is introduced to suppress the noise in IR images. Then, the RGB image is convoluted by sub-pixel convolution filter to obtain feature images, and the HR image is represented by the sum of the RGB image feature and IR upsampling image feature. In addition, we take our dataset, called IR-COLOR2000, as the dataset for training the ideal network. Finally, experimental results show that the proposed algorithm is superior to several comparison algorithms.
The remainder is organized as follows. Section 2 mainly contains the convolutional neural network for image SR, the existing datasets for super SR domain, and infrared image denosing algorithm. Section 3 presents the RGB-IR cross input and sub-pixel upsampling model, the importance of guided filter layer and sub-pixel convolution model. Section 4 follows the algorithm simulation for IR image and RGB image, respectively. Finally, conclusions are drawn in Section 5.

2. Related Works

2.1. CNN-Based SR

CNN-based works for SR algorithms basically focus on the single modal image (color image or infrared image). The Super-Resolution Convolution Neural Network (SRCNN) demonstrates that the three-layer network extracted the convolutional feature to reconstruct image, which runs at high speed but is very sensitive to noise. To extract deeper image features and avoid gradient attenuation, Kim proposed a Very Deep Convolutional Networks (VDSR) method, which learned only the high-frequency residuals between the input and the output. In addition, it reduced the training time of learning a large number of low-frequency similar parts. Moreover, the Laplacian Pyramid Super-Resolution Network (LapSRN [16]) can get the intermediate feature image and predict the next HR image by adding feature images. For the reason that the reconstruction HR image is gradually enlarged in the scale of 2, the algorithm running speed is relatively fast. However, these algorithms ignore the features of infrared images and have not yet got trained models from multimodal features of infrared and color  images.
The structure of the SRCNN, FSRCNN, VDSR, DRCN, and LapSRN networks are shown in Table 1. CNN-based super-resolution methods have many forms of network structure in Table 1, from the simplest 3-layer network to a 24-layer residual network. Among them, they are divided into direct reconstruction and progressive reconstruction according to the number of upsampled images. Methods with direct reconstruction perform pre-upsampling from the LR to HR patches, and progressive reconstruction is multiple post upsampling steps. The input of the network is LR or an interpolated image (LR + bicubic). Depth is the number of convolutional layers passed from input to 4x output. In addition, the loss function uses L2 loss function or other loss functions.
Compared with other network structures, LapSRN has the following advantages: progressive image reconstruction to reduce network parameters, residual structure to deepen the network, and use of the Charbonnier loss function suitable for the pyramid structure.

2.2. Laplacian Pyramid Network

The mainstream framework of image SR is to study the end-to-end network over one type of image. In this section, a deep Laplacian pyramid network (LapSRN) [16] is introduced for SR. The LapSRN contains two sub-networks of feature extraction and image reconstruction as Figure 1. We extracted feature images in feature extraction branch, then sent the feature into upsampling images to get residual result in image reconstruction branch for high resolution. Furthermore, the network output HR image is obtained from RGB image features and upscaling images at each level. In other words, the network uses multiple pyramid layers, and each layer shares the extracted features with the upsampling image in progressive reconstruction at scale 2. After training on three pyramid layers network, the output we obtained is 8 times the input.
Loss function. The loss function is the same Charbonnier penalty function as LapSRN. To represent the difference between HR image and the SR image, we assumed that the residual image of the s-level pyramid is r s , the image after the up-sampling is x s , and the corresponding HR image is y s = r s + x s . The corresponding pyramid level Y s is generated from downsampling the HR and the bicubic interpolation and loss function is
L o s s ( Y , y ; θ ) = 1 N i = 1 i = N S = 1 S = L δ Y s y s
L o s s ( Y , y ; θ ) = 1 N i = 1 i = N S = 1 S = L δ ( Y x s ) r s
where δ ( x ) = X 2 + ε 2 is explained δ ( ) in Equation (2); X represents ( Y x s ) r s ; ε is penalty, and the value is very small. L is the number of pyramid layer (L = 1, 2, 3); i denotes the pixels in image, N is number of pixels in image.

2.3. SR Datasets

Image super-resolution datasets are common datasets in the field of image processing. The most widely used datasets are ImageNet [17], Urban 100 [18], Greneral-100 [8], Set5, Set14, Manage109 [19], BSD300 and BSD500 [20]. The test data that is often used in image processing are Set5, Set14, Manage109, and so on. Databases of infrared and visible images [21,22,23] were applied in different computer vision tasks such as image enhancement, pedestrian detection, and region segmentation. However, most datasets do not consider IR images of high resolution. Therefore, we devote ourselves to capture pairs of IR and color images to train the deep network efficiently.
Dataset of IR-COLOR2000. The datasets contain both day and night images, which were generally captured in the outdoor (gardens, playgrounds, and buildings). The human in the night is salient in IR images, indicating that IR is sensitive to thermal radiation. Information about outdoor images is more sufficient than that of indoor images.
To effectively train the deep network, the database should contain large numbers or different types of images. In this paper, datasets that include IR and RGB images are obtained via two independent cameras, an uncooled long-wave detector IRT102, and Canon 60D. Besides, each pair of RGB and IR image faces the same environment. The data contains images of day and night, usually captured outdoors (gardens, playgrounds, and buildings). For the same scene image, the target size and the resolution of infrared images use the local self-similar-based image registration algorithm [24]. The database (named IR-COLOR2000) captured here consists of thousands of RGB and IR images with different illumination or scene. In particular, several pairs of images are shown in Figure 2.

2.4. IR Image Denoising

In the process of IR image SR, the image has low contrast and fuzzy edge. In particular, there are types of denoising algorithms, including variational, Total Variational (TV [25]), Partial Differential Equation (PDE [26,27]), Block-Matching, and the 3D filtering (BM3D [28,29]) method. TV is used to reduce the degradation of flat areas of the image, but it has faults of complex calculations and slow convergence. The purpose of PDE is to deal with fuzzy edges and edge position movement. PDE has a good effect on eliminating noise close to the Gaussian distribution, but it is not ideal for impulsive noise. BM3D is a noise reduction method that improves the sparse representation of images in the transform domain but the BM3D algorithm to perform random noise is not good.
Although the quality of IR images is improved, there is also much noise that needs to be reduced via images priors, such as smoothness and self-similarity. So the appropriate method for IR image denoising needs to be explored in the IR image super resolution procession without priors, such as [30]. Subsequently, the curvature filter [31] and guided filter methods [15,32], which can obtain well edge-preserving smoothing and less noise infrared images, are adopted for denoising. With a color image used as the guidance, the IR image can be well reconstructed, and the depth image edge can be preserved completely [14]. At the same time, we found that the network-based guided filtering [33] method can be trained to achieve an end-to-end network for image enhancement.

3. The Proposed Algorithm

In this section, based on LapSRN, this paper improves the LapSRN in three aspects for the characteristics of infrared images. we describe the proposed RGB-IR cross input and sub-pixel upsampling network. In the proposed network (Figure 3), pairs of IR and RGB images are used as input together, sub-pixel convolution is applied to optimize feature extraction network and a guided filter layer is adopted to reduce the influence of noise in IR images.
Several defects in IR image reconstruction, such as background noise, low contrast, and blur, result in bad performance. Therefore, an improved LapSRN based on IR-RGB images is presented to deal with IR image SR. RGB-IR cross input and sub-pixel upsampling network is also made up of RGB image feature extraction and IR image reconstruction. Sub-pixel convolution excels deconvolution in image upsampling to optimize feature extraction network especially. Moreover, the guided filter layer plays an important role in infrared image denoising and edges preserving.

3.1. RGB-IR Cross Input and Sub-Pixel Upsampling Network

Not to increase the depth of the network, the multi-modal image is used as input in our network for more details. That is, features of the visible image are add in infrared image reconstruction.
It is superior to other networks in that the proposed model includes sub-pixel layer for upsampling and a guided filtering layer for infrared image denoising. As Figure 3 shows, the features from RGB image is extracted by feature extraction group which concatenates the feature from “convolution” → “Relu” → “convolution” → “Relu” →⋯→ “sub-pixel convolution” → “convolution”. Then, the features from RGB images are added to the reconstruction of infrared images that go through “guided filtering” and “Deconvolution” to up-sampling processing. In this way, each image feature and infrared feature combination will enlarge the image scale by two times. Multi-scales 2 N image super-resolution needs to pass the feature to the next level for N feature extraction, and the infrared feature upsampling process, finally completes the reconstruction convolution. The detail information such as the kernel functions, depths and I/O dimensions of all “convs” in “Feature Extraction” and “Image Reconstruction” branches are shown in Table 2.
Network-intermediate layer images include feature image ×2, feature image ×4, and denoising results. Two sets of images are listed in Figure 4 which are the images of the middle layer in the proposed network at ×4 scale.
Residuals output of different scales and corresponding scale reconstruction results are obtained through cascade learning. The branch of image feature extraction takes RGB image as input to extract detailed features for the reconstruction of infrared image, and gradually samples and adds feature images of the same scale in the process of image reconstruction.
The feature extraction network extracts the visible image features, and upsamples the feature image through the sub-pixel convolution layer, and upsamples infrared image via the transposed convolution in the infrared image reconstruction, and the output image size is twice than the input image. When the super-resolution reconstruction scale is ×2, ×4, and ×8, the sub-pixel upsampling network feature extraction and the infrared image reconstruction upsampling operation are repeated m times (m = 1, 2, 3 ). The upsampling layer extracts image feature via a sub-pixel convolution layer. In each sub-pixel convolution part, the upsampled feature image is twice the size of its former feature image. It is worth noting that the input in the upsampling layer is IR images, and the RGB image is another input in the part of the feature extraction, which gives more detail information compared to handle SR.

3.2. Sub-Pixel Convolution

Taking into account that direct interpolation of deconvolution upsampling causes image blur, we have to use progressive upsampling for network upsampling and use subpixel convolution instead of deconvolution to reduce the loss of detail in the upsampling process. Sub-pixel convolution combines pixel values of multiple channels into one feature map, thus adding features in the network and changing the feature image size.
The sub-pixel layer plays a role as upsampling achieved by sub-pixel convolution in the extraction network. In contrast with upsampling, sub-pixel convolution combines feature images from multiple channels into one image. Pixel values in feature images are multiple channels value at the same position. The sub-pixel convolution layer can upscale the final LR feature maps into the HR output in the paper [34]. The feature map I F is the input of sub-pixel upsampling layer; f L ( I F ) is the output of sub-pixel upsampling layer; W L and b L is sub-pixel upsampling layer parameters. The implementation of sub-pixel convolution can be expressed as where ϕ ( ) is sub-pixel convolution operation, called shuffle. That operation changes the feature shape from W × H × r 2 tensor to r W × r H . The effects of sub-pixel convolution operation are shown in Figure 5. Sub-pixel convolution combines a single pixel on a multi-channel feature into a single pixel on a feature. Figure 5 shows the operation of sub-pixel convolution intuitively. To upsample the feature map of r times size, we need to generate r 2 feature maps of the same size. The operation of sub-pixel convolution is to assemble r 2 same size feature maps into a larger r times map.
f L ( I F ) = ϕ ( W L f L 1 ( I F ) + b L )
For the sake of improving and optimizing the network, sub-pixel convolution is used instead of deconvolution for image upsampling. Sub-pixel convolution avoids the danger of large numbers of zeros in general deconvolution. Besides, As can be seen from the comparison training in Figure 6, the sub-pixel convolution module has always had an absolute advantage over the simple upsampling in training 100 epochs on the dataset of IR-COLOR2000 for ×4 SR.
To prove the benefits of sub-pixel convolution, comparative experiments were conducted between networks with transposed convolutions and networks with sub-pixel convolution to improve network performance and rich feature details. As can be seen from the comparison training in Figure 6, the sub-pixel convolution module has always had an absolute advantage over the simple upsampling in training 100 epochs. The sub-pixel upsampling network in the feature extraction model makes full use of sub-pixel convolution. The experiment result shows that sub-pixel convolution is used for image feature extraction, which greatly improves the signal-to-noise ratio (PSNR) of the image.

3.3. Guided Filter Layer

Apart from training convolutional networks for more detail infrared image, the key to achieve infrared image super-resolution is the removal of infrared noise. It is necessary to enhance the infrared image and remove noise for uncooled long-wave infrared images containing noise. A denoising method is needed to eliminate noise and enhance the image in convolutional networks. Guided filtering has edge-preserving and denoising. It is through the network to train the guided filter parameters for image denoising enhancement.
The influence of noise can be reduced by the guided filter method. Furthermore, an RGB image is used as the guidance to complete IR image SR. The guided filter layer of this article is identical to the previous works in principle. The guided images we use are different. Generally, the input image is used as a guided image to maintain the edge. In this paper, the corresponding RGB image is used as a guided image to reduce the noise of the infrared image.
The guided filter process is linear translation-variant, and the output q is obtained by the guided filtering method of the input and the guidance I.
q i = a k I i + b k
where, a k and b k are the linear coefficients with restriction of the window w k ; k is the center point of w k window. In the filtering calculation, i represents the pixels of the window w k .
To seek a solution of minimizing the difference between output image q and input image p, it can also maintain the linear model Equation (4). The Equation (5) is the cost function in the window w k . ϵ  is a penalizing parameter.
E ( a k , b k ) = i w k ( ( a k I i + b k p i ) 2 + ϵ a k 2 )
a k = 1 w i w k I i p i μ k p k ¯ σ k 2 + ϵ
b k = p k ¯ a k μ k
Here, μ k and σ k 2 are respectively represented as the mean and variance of I in w k , w is the number of pixels in w k and 1 w i w k p i is the mean of p.
The flow chart and step of the presented algorithm are shown in Algorithm 1 and Figure 3, respectively.
Algorithm 1 Guided Filtering Layer for image processing
 Input: low-resolution IR image I l
           low-resolution RGB image G l
           Radius r and Regularization term ϵ
 Output: Guided filtered image as output O h
1. G l mean matrix calculation: G l ¯ = f μ ( G l , r )
2. I l mean matrix calculation: I l ¯ = f μ ( I l , r )
3. G l Correlation calculation: G l 2 ¯ = f μ ( G l G l , r )
4. G l , I l Correlation calculation: G l I l ¯ = f μ ( G l I l , r )
5. G l Variance calculation: G l = G l 2 ¯ G l ¯ G l ¯
6. G l , I l Covariance calculation: G l I l = G l I l ¯ G l ¯ G l ¯
Calculate guided filtering parameters a k and b k according to Equations. (6-7):
7. a k = G l I l G l + ϵ
8. b k = I l ¯ A l G l ¯
9. O h = a k G l + b k
/* f μ ( ) is a mean filter
To eliminate image noise the guided filter layer was trained in the RGB-IR cross input and sub-pixel upsampling model, then compare the SR effect on the synthetic data, such as images added salt & pepper noise. The noise image as input is illustrated in Figure 7, to perform image super resolution simulation and check whether the model is sensitive to noise. In Figure 7, the denoising ability of algorithms was analyzed and judged by the difference image between the noise image and the SR result. The less information contained in the difference image, the better the image SR result and the denoising effect. In Figure 7, it can be clearly seen that the proposed method is better than the comparison algorithm. It can be seen from Figure 7 that the proposed method has a good effect of removing noise. However, it can be seen from the value of PSNR that the salt & pepper noise has a bad influence on the image SNR.
It is observed that four noise images referring to the A+ method and the SRCNN method still contain much noise and lose edge or texture information. However, both the VDSR method and the proposed method do well to reduce image noise. For further comparison, the proposed method is superior to the VDSR algorithm in a subjective visual sense. In summary, the guided filter can improve image quality and reduce IR noise effectively.

3.4. The Loss Function of Multi-scale

The network uses 2×, 4×, and 8× samples to train multi-scale super-resolution models. It is necessary to construct network structures at different scales (2, 4, 8) and reduce the differences between SR images and HR images at different scales.
s c a l e = 2 , 4 , 8 L s c a l e ( Y , y , θ ) = L o s s ( Y , y , θ )
We note that the L o s s ( Y , y , θ ) is Charbonnier loss, and the scale in loss is upsampling scales for SR.

4. Experiment Results

4.1. Training & Testing

In this section, all experiments are implemented in Matlab 2016a on a PC with GPU Titan V, 12 GB RAM in ubuntu 16.04 system.
The training datasets (IR-COLOR2000), which contain 2000 pairs of images, are captured by ourselves. The images in the datasets main contain the infrared images and corresponding RGB images. Especially, we crop each input image into patches with a size of 128 × 128. The samples were generated by the operators of flipping horizontally and flipping vertically or rotating 90 , 180 , 270 . After the training is completed, the pre-trained model will be able to make 2×, 4× SR respectively. When testing, just select the model and enter the image to be tested.
We simultaneously conduct extensive evaluations of existing super-resolution models as baselines on our IR-COLOR2000 datasets. Then, the performance of several deep models, including the proposed RGB-IR cross input and sub-pixel upsampling network are evaluated. Three sets of experiments are performed: (1) Three pairs of IR images and RGB images as input respectively, we compare the effects of SR with single image input and two image input; (2) SR performance comparisons with state-of-the-arts algorithm by the upsampling scale of 2 and 4. we introduce the datasets for training and testing as follows.
Qualitative and quantitative performance of our model in comparison with state-of-the-art ones: A+, SCSR, SCN [35], SRCNN, VDSR, LapSRN, are presented as follows to evaluate the results. Peak Signal Noise Rate (PSNR) and Self-Similarity (SSIM) are used as the quality metric to evaluate the performance of all methods. Besides, visual comparisons on IR-COLOR2000 dataset are shown in Figure 8, Figure 9 and Figure 10 with the scale factor of 2. SR quantitative results of scale ×2 and ×4 are given in Table 3. In Figure 8, regions highlighted by green rectangles are magnified, and the difference between SR image and ground truth is clear for easy visual inspection. The result of RGB-IR cross input and sub-pixel upsampling network performance well in image reconstruction and simulation run at high speed.
We use PSNR and SSIM to evaluate the performance of our datasets. SSIM metric indicated the pixel similarity and local structure similarity between reconstructed HR image and ground truth. In Table 3, the performance of the proposed method is higher than A+, SCSR, SCN, SRCNN, VDSR, DRCN, LapSRN, especially on a scale of 4. Figure 8, Figure 9 and Figure 10 show SR experiment results by different algorithms. Reconstructed contextual information of the proposed method including edges and textures more clear than the others such as ailing in the Building, the waistcoat of Laborer, and the head of the man in Figure 8, Figure 9 and Figure 10.

4.2. Comparison to the State-of-the-Art

Since the previous image super-resolution reconstruction is pure infrared image super-resolution or visible image super-resolution. Furthermore, the existing input multi-modal data is all for image fusion, no dual input super-resolution reconstruction algorithm for images. For the sake of fairness, this paper makes more comparisons with single-input super-resolution reconstruction algorithms. Different from image fusion, this paper uses infrared and visible images for image super-resolution. Pairs of IR images and RGB images as input respectively, we compare the effects of SR with single image input and two images. To compensate for the contrast algorithm without double input, I will increase more single input super-resolution contrast experiments. Three pairs of images from the KAIST [36] and OutdoorUrban [37] dataset are used to test the performance of the proposed algorithm under low illumination.
To further highlight the performance of our SR method, a set of images in low light were selected for simulation experiments in Figure 11. OutdoorUrban dataset is images fusion datasets made by Nigel J. W. Morri in the Statistics of Infrared Images [37]. The contrast of the 4× scales experiment is performed by a set of dimly light images using the algorithms Bicubic, DRCN [10], RDN [38], SRGAN [39], CGAN [40] and WGAN [41] and proposed method which embedded the visible features into the infrared image in Figure 12. The PSNR and SSIM of Figure 11 and Figure 12 are shown in Table 4.
The experiment on night images from the KAIST dataset is shown in Figure 12. Clear RGB images is helpfuly for infrared images recovery, and improved SR results can be seen in Figure 12h. Also, the reconstruction image makes the inconspicuous features of infrared image clearer, which is consistent with human visual features. However, when the RGB Image is not clear enough, our method hasn’t any advantages of the image SR. It is obvious that the partial brightness of the lamp in the figure is different when the super-resolution scale is four times and two times, and the effect of the proposed method is indeed better than the others especially the evaluation values in Figure 12.

4.3. Analysis& Discussion

Six pairs of images from KAIST [36] and OutdoorUrban [37] dataset respectively as Figure 13 are chosen for SR experiments on A+, SCSR, SRCNN, DRCN, SRGAN, WGAN, CGAN, and our method. Infrared images and visible images are from two datasets of OutdoorUrban and KAIST. Figure 13a,b are in slightly weak light; Figure 13c,d are in low light; Figure 13e,f are in normal light. The PSNR and SSIM of Figure 13 are in Table 4.
Further verifying the stability of the algorithm, the data in KAIST and OutdoorUrban are chosen for quality comparison in Table 4.
The CGAN, WGAN, SRGAN, and RDN in Table 4 are not models for infrared images, so the experimental results are not optimal. The data in Table 4 shows that the algorithm in this paper is more stable than RDN and GAN-based methods under weak light and normal light conditions. The Evaluation indexes of Table 4 marked in blue are the best results of image SR. When visible light is low, the algorithm in this paper loses some advantages. In general, it can be seen from the mean signal-to-noise ratio of the six groups of experiments that the proposed algorithm is more stable than other algorithms.
To prove that infrared image noise reduction is effective for super-resolution networks, we conducted an experiment on three pairs of images from the OutdoorUrban [37] dataset. We use the proposed algorithm to reconstruct the above three pairs of images as shown in Figure 14 and calculate the image quality evaluation results. The results without denoising process are shown in Figure 14a and the results with denoising are shown in Figure 14b.
As can be seen in Figure 14, the results of group (b) are better than that of group (a). For the evaluation values of PSNR and SSIM, it can be noticed that the PSNR of (b) is nearly 1dB higher than (a). In addition, the image similarity parameters SSIM has not decreased significantly. This indicates that our denoising scheme can improve performance.
Since RGB-IR cross input and sub-pixel upsampling network make full use of the advantage of multiple model images, the network obtained a larger receptive field and less noise infrared image. Experimental results show that the proposed algorithm is very successful and effectively to improve image resolution.

4.4. Running Time

In this work, we can design and train different pre-trained models. Also, we need not train the model with odd scales, because the proposed network could handle the upsampling rate is odd. When dealing with an odd scale such as 3×, the input image is upsampled to the nearest scale by 4×, then the super-resolution result is downsampled to the target scale. Moreover, our method is less time consumption in comparison to the methods A+, SCN, SRCNN, VDSR, DRCN, RDN [38], SRGAN [39], CGAN [40] and WGAN [41] as illustrated in Table 5.
The running time is related to the depth and the input size of the network. The structure of SCN, SRCNN, and VDSR are relatively simple, but the input image size is larger than our proposed network because the input of that three networks are directly up-sampled to the ideal size, but our algorithm is gradually up-sampled to the ideal size. Thus the inference time of our proposed is shorter than these networks. The structure of DRCN, RDN, SRGAN, CGAN, and WGAN algorithms are deeper than our algorithm, which corresponds to longer inference time.

5. Conclusions

In this paper, we presented two input networks that employ the guided upsampling Laplacian Pyramid Network for super-resolution. The deep network can be optimized to be faster via guided filtering and sub-pixel convolution upsampling step by step. A significant improvement can be shown visually by using an RGB image to guide the IR input image and combining the RGB feature image with the IR feature image. The proposed RGB-IR cross input and sub-pixel upsampling network reduced the IR image noise problem and improved the IR super-resolution image quality by adding RGB details. Relative evaluations on datasets demonstrate that the proposed algorithm performs satisfactorily compared to the other SR methods in terms of visual quality.

Author Contributions

J.D. has made substantial contributions to the algorithm design. K.Q., W.T. and J.D. made substantial contributions to the algorithm implementation and manuscript preparation. Z.Z., W.T. and J.D. have made substantial contributions to the experimental data collection and the result analysis. H.Z., Y.Y. and L.G. revised the final manuscript. All authors have read and agreed to the published version of the manuscript.

Acknowledgments

The work was supported by the National Natural Science Foundation of China (61675160, 61705173 and 51801142), the Fundamental Research Funds for the Central Universities (No.201813019), Project funded by China Postdoctoral Science Foundation (No.2019M652472), and the 111 Project (B17035).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chang, H.; Yeung, D.Y.; Xiong, Y. Super-resolution through neighbor embedding. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; p. 1. [Google Scholar]
  2. Timofte, R.; De, V.; Gool, L.V. Anchored Neighborhood Regression for Fast Example-Based Super-Resolution. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, NSW, Australia, 1–8 December 2014; pp. 1920–1927. [Google Scholar]
  3. Timofte, R.; Smet, V.D.; Gool, L.V. A+: Adjusted Anchored Neighborhood Regression for Fast Super-Resolution. In Proceedings of the Asian Conference on Computer Vision, Singapore, 1–5 November 2014; pp. 111–126. [Google Scholar]
  4. Jianchao, Y.; John, W.; Thomas, H.; Yi, M. Image Super-Resolution via Sparse Representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  5. Zhao, Y.; Chen, Q.; Sui, X.; Gu, G. A novel infrared image super-resolution method based on sparse representation. Infrared Phys. Technol. 2015, 71, 506–513. [Google Scholar] [CrossRef]
  6. Qi, G.; Wang, H.; Haner, M.; Weng, C.; Chen, S.; Zhu, Z. Convolutional neural network based detection and judgement of environmental obstacle in vehicle operation. CAAI Trans. Intell. Technol. 2019, 4, 80–91. [Google Scholar] [CrossRef]
  7. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image SuperResolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Dong, C.; Chen, C.L.; Tang, X. Accelerating the Super-Resolution Convolutional Neural Network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 391–407. [Google Scholar]
  9. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  10. Kim, J.; Kwon Lee, J.; Mu Lee, K. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  11. Ma, J.; Wang, X.; Jiang, J. Image Super-Resolution via Dense Discriminative Network. IEEE Trans. Ind. Electron. 2019. [Google Scholar] [CrossRef]
  12. Ma, J.; Yu, W.; Liang, P.; Li, C.; Jiang, J. FusionGAN: A generative adversarial network for infrared and visible image fusion. Inf. Fusion 2019, 48, 11–26. [Google Scholar] [CrossRef]
  13. Han, T.Y.; Kim, Y.J.; Song, B.C. Convolutional neural network-based infrared image super resolution under low light environment. In Proceedings of the 25th European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 803–807. [Google Scholar]
  14. Tseng, C.W.; Su, H.R.; Lai, S.H.; Liu, J. Depth image super-resolution via multi-frame registration and deep learning. In Proceedings of the 2016 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA), Jeju, South Korea, 13–16 December 2016; pp. 1–8. [Google Scholar]
  15. Huang, D.; Huang, W.; Gu, P.; Liu, P.; Luo, Y. Image super-resolution reconstruction based on regularization technique and guided filter. Infrared Phys. Technol. 2017, 83, 103–113. [Google Scholar] [CrossRef]
  16. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Fast and accurate image super-resolution with deep laplacian pyramid networks. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 11, 2599–2613. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vision 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  18. Huang, J.B.; Singh, A.; Ahuja, N. Single Image Super-resolution from Transformed Self-Exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  19. Fujimoto, A.; Ogawa, T.; Yamamoto, K.; Matsui, Y.; Aizawa, K. Manga109 dataset and creation of metadata. In Proceedings of the 1st International Workshop on coMics ANalysis, Processing and Understanding, Cancun, Mexico, 4 December 2016; p. 2. [Google Scholar]
  20. Martin, D.R.; Fowlkes, C.; Tal, D.; Malik, J. A Database of Human Segmented Natural Images and its Application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth International Conference On Computer Vision (ICCV-01), Vancouver, BC, Canada, 7–14 July 2001; pp. 1–11. [Google Scholar]
  21. Choi, Y.; Kim, N.; Hwang, S.; Kweon, I.S. Thermal Image Enhancement using Convolutional Neural Network. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, South Korea, 9–14 October 2016; pp. 223–230. [Google Scholar]
  22. Barrera, F.; Lumbreras, S.A.D. Multispectral piecewise planar stereo using Manhattan-world assumption. Pattern Recognit. Lett. 2013, 34, 52–61. [Google Scholar] [CrossRef]
  23. Hwang, S.; Park, J.; Kim, N.; Choi, Y.; Kweon, I.S. Multispectral Pedestrian Detection: Benchmark Dataset and Baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1037–1045. [Google Scholar]
  24. Zou, X. Research on Infrared and Visible Image Registration Algorithm Based on Self-similarity. In Proceedings of the International Conference on Information Science and Technology (IST 2017), Beijing, China, 18–20 October 2017; p. 02002. [Google Scholar]
  25. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Physica D 1992, 60, 259–268. [Google Scholar] [CrossRef]
  26. Zhang, X.; Wang, R.; Jiao, L.C. Partial Differential Equation Model Method Based on Image Feature for Denoising. In Proceedings of the 2011 International Workshop on Multi-Platform/Multi-Sensor Remote Sensing and Mapping, Xiamen, China, 10–12 January 2011; pp. 1–4. [Google Scholar]
  27. Pu, Y.F.; Siarry, P.; Zhou, J.L.; Liu, Y.G.; Zhang, N.; Huang, G.; Liu, Y.Z. Fractional partial differential equation denoising models for texture image. Sci. Chin. Inf. Sci. 2014, 57, 1–19. [Google Scholar] [CrossRef] [Green Version]
  28. Dabov, K.; Foi, A.; Egiazarian, K. Image denoising with block-matching and 3D filtering. In Proceedings of the Image Processing: Algorithms and Systems, Neural Networks, and Machine Learning, San Jose, CA, USA, 15–19 January 2006; p. 606414. [Google Scholar]
  29. Honzátko, D.; Kruliš, M. Accelerating block-matching and 3D filtering method for image denoising on GPUs. J. Real-Time Image Process. 2017, 1–15. [Google Scholar]
  30. Guan, J.; Lai, R.; Xiong, A. Wavelet Deep Neural Network for Stripe Noise Removal. IEEE Access 2019, 7, 44544–44554. [Google Scholar] [CrossRef]
  31. Tan, W.; Zhou, H.; Rong, S.; Qian, K.; Yu, Y. Fusion of multi-focus images via a Gaussian curvature filter and synthetic focusing degree criterion. Appl. Opt. 2018, 57, 10092–10101. [Google Scholar] [CrossRef] [PubMed]
  32. He, K.; Sun, J.; Tang, X. Guided image filtering. In Proceedings of the European Conference on Computer Vision, Crete, Greece, 5–11 September 2010; pp. 1–14. [Google Scholar]
  33. Wu, H.; Shuai, Z.; Zhang, J.; Huang, K. Fast End-to-End Trainable Guided Filter. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1838–1847. [Google Scholar]
  34. Shi, W.; Caballero, J.; Theis, L.; Huszar, F.; Wang, Z. Is the deconvolution layer the same as a convolutional layer? arXiv 2016, 7009. [Google Scholar]
  35. Wang, Z.; Liu, D.; Yang, J.; Han, W.; Huang, T. Deep networks for image super-resolution with sparse prior. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 370–378. [Google Scholar]
  36. Lee, S.; Cho, M.S.; Jung, K.; Kim, J.H. Scene text extraction with edge constraint and text collinearity. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 3983–3986. [Google Scholar]
  37. Morris, N.J.W.; Avidan, S.; Matusik, W.; Pfister, H. Statistics of Infrared Images. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–7. [Google Scholar]
  38. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  39. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  40. Yuyang, W.; Feng, S.; Ye, Q. Super-Resolution of Text Image Based on Conditional Generative Adversarial Network. In Proceedings of the Pacific Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; pp. 270–281. [Google Scholar]
  41. Yu, L.; Long, X.; Tong, C. Single Image Super-Resolution Based on Improved WGAN. In Proceedings of the 2018 International Conference on Advanced Control, Automation and Artificial Intelligence (ACAAI 2018), Shenzhen, China, 21–22 January 2018. [Google Scholar]
Figure 1. LapSRN Network Architecture
Figure 1. LapSRN Network Architecture
Sensors 20 00281 g001
Figure 2. Dataset Section 2.3 of IR-COLOR2000( the scene in outdoor and indoor ), (a) Infrared images from IR-COLOR2000, (b) RGB images from IR-COLOR2000.
Figure 2. Dataset Section 2.3 of IR-COLOR2000( the scene in outdoor and indoor ), (a) Infrared images from IR-COLOR2000, (b) RGB images from IR-COLOR2000.
Sensors 20 00281 g002
Figure 3. RGB-IR cross input and sub-pixel upsampling Network architecture. (the red arrow represents feature extraction, the green arrow represents the addition of the extracted feature and the upsampled feature, and the blue arrow represents the upsampling operation.).
Figure 3. RGB-IR cross input and sub-pixel upsampling Network architecture. (the red arrow represents feature extraction, the green arrow represents the addition of the extracted feature and the upsampled feature, and the blue arrow represents the upsampling operation.).
Sensors 20 00281 g003
Figure 4. The input, output and intermediate layer image of RGB-IR cross input and sub-pixel upsampling network at ×4 scale. (a) input RGB image, (b) feature image ×2, (c) feature image ×4, (d) input IR image, (e) denoising result, and (f) SR image.
Figure 4. The input, output and intermediate layer image of RGB-IR cross input and sub-pixel upsampling network at ×4 scale. (a) input RGB image, (b) feature image ×2, (c) feature image ×4, (d) input IR image, (e) denoising result, and (f) SR image.
Sensors 20 00281 g004
Figure 5. Sub-pixel convolution (r = 2).
Figure 5. Sub-pixel convolution (r = 2).
Sensors 20 00281 g005
Figure 6. Performance of sub-pixel model in feature extraction. (blue curve represents no sub-pixel model, red curve represents sub-pixel model)
Figure 6. Performance of sub-pixel model in feature extraction. (blue curve represents no sub-pixel model, red curve represents sub-pixel model)
Sensors 20 00281 g006
Figure 7. Difference image, which is A+, SRCNN, VDSR, ours SR result subtracting the input noise image. (PSNR values for noise image SR results.)
Figure 7. Difference image, which is A+, SRCNN, VDSR, ours SR result subtracting the input noise image. (PSNR values for noise image SR results.)
Sensors 20 00281 g007
Figure 8. SR experiment results of scale 2 in Teaching building.
Figure 8. SR experiment results of scale 2 in Teaching building.
Sensors 20 00281 g008
Figure 9. SR experiment results of scale 2 in Laborer.
Figure 9. SR experiment results of scale 2 in Laborer.
Sensors 20 00281 g009
Figure 10. SR experiment results of scale 2 in Window.
Figure 10. SR experiment results of scale 2 in Window.
Sensors 20 00281 g010
Figure 11. The visual quality comparison for an image from OutdoorUrban datasets with scale ×2 and ×4.(a) is infrared image as input (b) is the RGB image as input. (c1f1) is SR results for ×2 scale respectively proposed method, vdsr, srcnn, A+. (c2f2) is SR results for ×4 scale respectively proposed method, vdsr, srcnn, A+.
Figure 11. The visual quality comparison for an image from OutdoorUrban datasets with scale ×2 and ×4.(a) is infrared image as input (b) is the RGB image as input. (c1f1) is SR results for ×2 scale respectively proposed method, vdsr, srcnn, A+. (c2f2) is SR results for ×4 scale respectively proposed method, vdsr, srcnn, A+.
Sensors 20 00281 g011
Figure 12. SR experiment results of scale 4 in night from KAIST dataset.
Figure 12. SR experiment results of scale 4 in night from KAIST dataset.
Sensors 20 00281 g012
Figure 13. Test images in Table 4 ((a,b) are from OutdoorUrban dataset; (c,d) are from night videos of KAIST dataset; (e,f) are from daylight videos of KAIST dataset.).
Figure 13. Test images in Table 4 ((a,b) are from OutdoorUrban dataset; (c,d) are from night videos of KAIST dataset; (e,f) are from daylight videos of KAIST dataset.).
Sensors 20 00281 g013
Figure 14. Comparison on the proposed method without denoising and the proposed method with denoising. (a) is the result of the proposed method without denoising; (b) is the result of the proposed method with denoising.
Figure 14. Comparison on the proposed method without denoising and the proposed method with denoising. (a) is the result of the proposed method without denoising; (b) is the result of the proposed method with denoising.
Sensors 20 00281 g014
Table 1. Comparison of CNN-based SR methods: SRCNN, FSRCNN, VDSR, DRCN, and LapSRN.
Table 1. Comparison of CNN-based SR methods: SRCNN, FSRCNN, VDSR, DRCN, and LapSRN.
MethodInputDepthResidual StructureReconstructionLoss
SRCNNLR + bicubic3×-L2
FSRCNNLR8×-L2
VDSRLR + bicubic20-L2
DRCNLR + bicubic5 (recursive)×-L2
LapSRNLR24ProgressiveCharbonnier
Table 2. Convolution structure of network branches for 4 × SR.
Table 2. Convolution structure of network branches for 4 × SR.
PartConvolution Kernel SizeDepthFiltersParametersInput DimensionsOutput Dimensions
Feature Extraction 3 × 3 2264737k164
sub-pixel 3 × 3 26436k6432
2 × 2 23216k3264
Image Reconstruction 4 × 4 264102411
Table 3. Quantitative evaluation of state-of-art SR algorithm.
Table 3. Quantitative evaluation of state-of-art SR algorithm.
AlgorithmScaleBuildingLaborerWindow
RGBIRRGBIRRGBIR
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/ SSIMPSNR/SSIMPSNR/SSIM
Bicubic×232.858/0.90134.755/0.90828.202/0.82637.436/0.93332.137/0.90940.118/0.938
A+32.913/0.91235.01/0.92129.011/0.92137.867/0.94033.973/0.92940.563/0.941
SCSR33.158/0.90635.055/0.91328.502/0.83737.736/0.93432.431/0.92840.417/0.939
SCN31.484/0.71533.468/0.53727.204/0.73035.078/0.65332.312/0.55538.159/0.506
SRCNN35.051/0.93636.144/0.92430.019/0.88438.459/0.94534.888/0.93340.777/0.945
VDSR34.420/0.92235.802/0.91329.573/0.87138.200/0.93634.078/0.92340.609/0.939
LapSRN34.440/0.92935.815/0.92029.580/0.87638.207/0.94334.083/0.93040.616/0.945
Ours34.915/0.94536.190/0.92930.171/0.88038.836/0.94834.682/0.94341.150/0.949
Bicubic×426.042/0.62727.971/0.74823.783/0.52931.886/0.821327.747/0.76735.224/0.867
A+28.912/0.71230.213/0.81024.871/0.59233.248/0.84929.436/0.81836.989/0.899
SCSR26.339/0.64938.311/0.76823.857/0.52832.221/0.82627.902/0.90235.550/0.876
SCN26.151/0.63628.129/0.75924.537/0.56932.408/0.83928.962/0.79035.727/0.889
SRCNN29.365/0.76731.445/0.83425.164/0.64034.310/0.87630.292/0.84437.535/0.904
VDSR29.382/0.75431.101/0.81325.129/0.63834.122/0.86030.253/0.83537.070/0.891
LapSRN29.392/0.76631.110/0.82425.138/0.64134.133/0.87130.264/0.84437.079/0.898
Ours29.989/0.78231.890/0.85025.82/0.65034.697/0.88530.762/0.86437.805/0.910
Table 4. Quantitative evaluation of a few images from KAIST and OutdoorUrban dataset.
Table 4. Quantitative evaluation of a few images from KAIST and OutdoorUrban dataset.
ImageScaleIndexBicubicA+SRCNNSCSRCGANWGANDRCNRDNSRGANProposed
Sensors 20 00281 i001×2PSNR42.40043.443.63542.4244-17.89644.1744.2124.53844.704
SSIM0.95920.9630.9650.9603-0.3640.9760.9770.5030.987
×4PSNR38.95536.09339.72339.245739.98314.7839.9840.4720.8940.31
SSIM0.9380.8870.9440.93850.9500.5870.9420.9490.5810.958
Sensors 20 00281 i002×2PSNR42.61443.243.30342.369-24.92444.10144.4324.58945.087
SSIM0.9640.9710.9670.962-0.3960.9760.9790.5950.986
×4PSNR39.12438.339.85539.31339.97517.87439.5940.5420.45640.06
SSIM0.93770.940.9440.93770.9480.5870.9490.9450.5260.960
Sensors 20 00281 i003×2PSNR38.94639.940.43239.01-18.31338.39241.2129.20741.929
SSIM0.9460.9530.9590.948-0.2050.9410.9700.6550.972
×4PSNR33.96734.635.76234,335.93416.9533.8736.5216.97136.53
SSIM0.8760.8760.8890.8720.9010.2830.8650.890.53860.921
Sensors 20 00281 i004×2PSNR46.61846.546.25346.4-13.04543.89147.3634.9645.50
SSIM0.9750.9740.9780.968-0.0530.9670.9670.8060.916
×4PSNR41.30741.942.94241.642.79710.80841.043.7311.61640.48
SSIM0.9540.9520.9570.9530.9500.0500.9510.9640.4070.920
Sensors 20 00281 i005×2PSNR54.18854.454.8253.9-15.20354.7454.3127.46956.837
SSIM0.9760.9810.9860.976-0.1200.9820.9810.6380.991
×4PSNR46.146.2646.30546.341.28313.73844.7444.5312.75247.196
SSIM0.9780.9800.9820.9810.9680.2440.976o.9800.5820.992
Sensors 20 00281 i006×2PSNR53.554.054.73254.6-17.39354.97954.2332.8257.42
SSIM0.9560.9640.9680.960-0.1570.9710.9690.7410.984
×4PSNR44.27546.0847.2646.343.06215.68644.97945.1216.39747.71
SSIM0.9150.9240.9680.9240.9580.2050.9490.960.5820.979
Average×2PSNR46.37946.947.19746.450-17.79546.71247.62528.93048.579
×4PSNR40.62140.53841.97441.17640.50514.97240.69341.83516.51342.047
Table 5. Comparison of running time for 4× in our datasets.
Table 5. Comparison of running time for 4× in our datasets.
MethodsA+SCNSRCNNVDSRDRCNEDSRRDNSRGANCGANWGANOurs
Running time/ Sec2.1635.3400.7280.2468.2150.6120.4690.2930.2880.1500.098

Share and Cite

MDPI and ACS Style

Du, J.; Zhou, H.; Qian, K.; Tan, W.; Zhang, Z.; Gu, L.; Yu, Y. RGB-IR Cross Input and Sub-Pixel Upsampling Network for Infrared Image Super-Resolution. Sensors 2020, 20, 281. https://doi.org/10.3390/s20010281

AMA Style

Du J, Zhou H, Qian K, Tan W, Zhang Z, Gu L, Yu Y. RGB-IR Cross Input and Sub-Pixel Upsampling Network for Infrared Image Super-Resolution. Sensors. 2020; 20(1):281. https://doi.org/10.3390/s20010281

Chicago/Turabian Style

Du, Juan, Huixin Zhou, Kun Qian, Wei Tan, Zhe Zhang, Lin Gu, and Yue Yu. 2020. "RGB-IR Cross Input and Sub-Pixel Upsampling Network for Infrared Image Super-Resolution" Sensors 20, no. 1: 281. https://doi.org/10.3390/s20010281

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop