Next Article in Journal
High-Throughput Phenotyping Approach for Screening Major Carotenoids of Tomato by Handheld Raman Spectroscopy Using Chemometric Methods
Next Article in Special Issue
Implicit and Explicit Regularization for Optical Flow Estimation
Previous Article in Journal
Convolutional Neural Network and Motor Current Signature Analysis during the Transient State for Detection of Broken Rotor Bars in Induction Motors
Previous Article in Special Issue
Depth-Dependent High Distortion Lens Calibration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Deblurring Using Multi-Stream Bottom-Top-Bottom Attention Network and Global Information-Based Fusion and Reconstruction Network

Department of Biomedical Engineering, School of Life Science and Technology, Ministry of Education Key Laboratory of Molecular Biophysics, Huazhong University of Science and Technology, No 1037, Luoyu Road, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(13), 3724; https://doi.org/10.3390/s20133724
Submission received: 29 May 2020 / Revised: 25 June 2020 / Accepted: 30 June 2020 / Published: 3 July 2020
(This article belongs to the Special Issue Data, Signal and Image Processing and Applications in Sensors)

Abstract

:
Image deblurring has been a challenging ill-posed problem in computer vision. Gaussian blur is a common model for image and signal degradation. The deep learning-based deblurring methods have attracted much attention due to their advantages over the traditional methods relying on hand-designed features. However, the existing deep learning-based deblurring techniques still cannot perform well in restoring the fine details and reconstructing the sharp edges. To address this issue, we have designed an effective end-to-end deep learning-based non-blind image deblurring algorithm. In the proposed method, a multi-stream bottom-top-bottom attention network (MBANet) with the encoder-to-decoder structure is designed to integrate low-level cues and high-level semantic information, which can facilitate extracting image features more effectively and improve the computational efficiency of the network. Moreover, the MBANet adopts a coarse-to-fine multi-scale strategy to process the input images to improve image deblurring performance. Furthermore, the global information-based fusion and reconstruction network is proposed to fuse multi-scale output maps to improve the global spatial information and recurrently refine the output deblurred image. The experiments were done on the public GoPro dataset and the realistic and dynamic scenes (REDS) dataset to evaluate the effectiveness and robustness of the proposed method. The experimental results show that the proposed method generally outperforms some traditional deburring methods and deep learning-based state-of-the-art deblurring methods such as scale-recurrent network (SRN) and denoising prior driven deep neural network (DPDNN) in terms of such quantitative indexes as peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) and human vision.

1. Introduction

Image blurring is one of the major reasons for image quality degradation. It poses a great challenge for understanding and analyzing the high-frequency components in the images [1]. Therefore, deblurring has been a hot research field in image restoration [2,3,4,5,6]. Image deblurring is the process of inferring the latent sharp images in the absence of degradation model information. The blur types can be roughly divided into two categories. The first one is motion blur. For example, camera shake during exposure will blur the captured image. The second one is defocus blur produced by such factors as atmospheric turbulence and aberrations in the optical system. In most situations of practical interest, Gaussian blur is ubiquitous in optical imaging systems. Gaussian blur removal is highly significant in many applications; thus, it has attracted much attention in the field of image deblurring.
Many existing deblurring methods generally describe the blur process as
B = I o k + n
where denotes the 2D convolutional operator, B is the blurry input image, I o is the sharp image, k is the blur kernel, and n is the noise.
To find the sharp image in Equation (1) is an ill-posed problem. Solutions to this ill-posed problem can be divided into regularization methods and iterative schemes via updating and stopping rules. In more detail, the image deblurring methods include the optimization-based and the learning-based methods. Optimization-based algorithms firstly estimate the blur kernel and then realize non-blind deconvolution by using various image priors such as L0 gradients [7], sparse gradients [8,9,10], normalized sparsity [11], and patch priors [12] to regularize the solution space. For example, the nonlocally centralized spare representation (NCSR) [13] model has been presented to improve the performance of sparse representation-based image restoration by using the nonlocal self-similarity to estimate the sparse coding coefficients. Expected patch log likelihood (EPLL) [14] uses the regularization model based on the statistics of image patches for whole-image restoration. The block matching 3D filtering (BM3D) [15] method utilizes the formalized image modeling to realize the iterative deblurring. Michaeli and Irani [16] used self-similarity as an image prior to addressing the deblurring problem. Shan et al. [17] used a piecewise function to fit the gradient distribution of multiple natural images and then iteratively estimated the blur kernel function and the restored image. Pan et al. [18] presented the dark channel prior to image deblurring. This method can perform well in deblurring both natural images and text, face, and low-illumination images under the premise that the dark pixels can be extracted from an image. Chen et al. [19] proposed a deblurring method based on local maximum gradient prior and an effective optimization scheme. Yan et al. [20] proposed a novel image prior named extreme channels prior (ECP) by combining the bright channel prior with the dark channel one to improve the deblurred results. Although these methods have demonstrated good performance in some cases, they involve complicated computation and heuristic parameter-tuning [21]. Meanwhile, due to the utilization of hand-crafted priors that are generally designed based on the limited observations, these methods are not robust enough and may be ineffective for deblurring Gaussian blurred images with the complicated features.
With regard to the learning-based deblurring methods, deep learning has become very popular in the field of image deblurring. Up to now, the different deep learning models, such as generative adversarial network (GAN) and convolutional neural networks (CNN) (e.g., attentional convolutional network [22]) have been proposed. These deep learning-based deblurring methods can be roughly divided into three categories. The first one is to use the CNN to predict the blur kernel and then adopt the traditional non-blind deconvolution methods to restore the images. Sun et al. [23] proposed to predict the probabilistic distribution of motion blur for each patch using the CNN, and a deblurring model using the patch prior was adopted to remove motion blur. Yan et al. [24] used two neural networks to identify the blur type and then estimate its parameters, respectively. Cheng et al. [25] introduced the hand-crafted priors and the implicit deep priors trained by the CNNs as the regularization for the kernel estimation and used half-quadratic splitting optimization for image deconvolution. The performance of these methods greatly depends on the estimation accuracy of blur kernel. The second kind is to directly learn end-to-end mapping to output the deblurred results [26,27,28]. Tao et al. [21] presented a multi-scale recurrent network (SRN) that directly deblurs the wild images by combining the multi-scale images. Zhang et al. [29] presented the GAN-based image deblurring method in which spatial information and temporal information were combined as a generator. Wu et al. [30] presented a dual attention structure to dynamically aggregate the temporal cues for deblurring. These methods generally cannot handle serious image blurring effectively. The last kind is to use the model-driven deep learning methods for image deblurring. Dong et al. [31] presented an iterative deblurring method based on the denoising prior driven deep neural network (DPDNN). Distinctively, the alternating direction method of multipliers (ADMM) [32] technique is used to convert the constrained optimization problem into sub-problems, and the iterative process is unfolded into a deep neural network, which is composed of multiple denoiser modules for exploiting the multi-scale redundancies of natural images and back-projection modules for ensuring the observation consistencies.
Despite the advantage of the deep learning-based deblurring methods over the traditional ones, they still cannot perform well in restoring fine details and reconstructing sharp edges. In particular, small-scale details in the blurred image may be damaged seriously by the deblurring method in the case of the serious image blurring. To address this issue, we proposed two deep learning models to produce the deblurred image with the image details preserved well. For one, we proposed the multi-stream bottom–top–bottom attention network (MBANet) to extract image features at different scales using the multi-scale scheme. In the proposed MBANet, the multi-scale components of original images are firstly extracted and fed into the encoder–decoder ResBlock network simultaneously, which integrate high-level semantic information in the data stream with low-level spatial and channel-wise information. Then, the multi-scale output maps are produced by the multi-stream network with the multi-scale structure, which exploits the significant influence of multi-scales on Gaussian blur. On the other hand, to realize the effective fusion of output maps produced by the MBANet and produce the satisfactory deblurred image, the global information-based fusion and reconstruction network (GIFRNet) was proposed. This network includes the two sub-networks, i.e., the global information-based fusion network (GIFNet) and the global information-based reconstruction network (GIRNet). The GIFRNet integrates multi-scale output maps to greatly improve the global spatial information and continuously refine the outputs of GIFNet to compensate for the unwanted loss of image details in the deblurred image, especially the small-scale ones. Experiments were done on two different dataset to demonstrate the effectiveness and superiority of the proposed method to the state-of-the-art deblurring methods.
The main contributions of this study are summarized as follows:
  • We proposed a novel MBANet by combining the advantages of encoder-decoder structure and attention mechanisms to extract multi-scale and multi-level features effectively, thereby providing a solid foundation for recovering the fine details and reconstructing the sharp edges from the blurry images.
  • The proposed global information-based fusion network can merge the channel-wise and spatial information encoded in the bottom–top data stream at each scale effectively. Moreover, the proposed reconstruction network based on the fused global information can gradually refine the outputs of the GIFNet to finally produce the satisfactory deblurred images.
  • A new loss function was designed by combining the perceptual loss with the mean squared error to accelerate the gradient descent of the network, thereby leading to effective network training.
The remainder of the paper is organized as follows. In Section 2, we present the details of MBANet and the GIFRNet. In Section 3, the experimental results are provided. Finally, conclusions and future research directions are given in Section 4.

2. The Structure of the Proposed Networks

The architecture of the proposed networks is illustrated in Figure 1. As shown in Figure 1, the multi-scale images generated by the resizing operation are firstly fed into the MBANet, which combines a modified encoder-decoder ResBlock network with an attention mechanism. Then, the integration of backward and forward information is performed by the skipping connection. Based on the global information, the squeeze-and-excitation network (SENet) [33] is introduced here to implement the adaptive calibration of the channel-wise feature, highlight the important features in the encoder, and pass them to the decoder. Following the decoder, the resizing operation is again adopted to ensure that the generated multi-scale maps have the same resolution as the original input image. Finally, the GIFRNet is utilized to integrate the global feature information at each scale and refine the output deblurred image. We will provide the detailed description of the construction of the MBANet and the GIFRNet as well as the process of model training.

2.1. MBANet

With regard to the well-established multi-scale networks, there are two main types of network structures. The first one is to process images at different scales separately, which lead to the non-utilization of information between different scales. The second one is to cascade and share the weights among different scales. The integration of features by way of cascading makes network training easier and more stable, but its effectiveness greatly depends on the accuracy at the first scale. In this work, we proposed a parallel bottom-top-bottom network structure and introduced spatial and channel attention mechanisms. As opposed to other multi-scale methods, our method uses multi-scale images as the inputs of the network, so that the errors introduced by the first scale can be prevented from transmission to other scales. In order to greatly reduce training difficulty and increase stability, the network weights is shared across different scales.
The proposed end-to-end MBANet can be viewed as a regression network. From the perspective of network structure, there are several advantages. Firstly, the network is deep enough to generate a large receptive field for extracting different levels of features. Secondly, the combination of spatial and channel-wise information can facilitate the utilization of the global information to highlight the important features. Finally, the network combines the low-level cues and the high-level semantic information to improve the fineness of the deblurred image.

2.1.1. Structure of Bottom-Top-Bottom Attention Network

We chose the encoder-decoder model as the basic structure of the bottom-top-bottom [34] attention network and modified it to realize accurate feature extraction and combination. The encoder-decoder network [35] has a symmetric structure in which the input images are progressively transformed into feature maps with more channels and smaller sizes, and then the feature maps are transformed back to the resolution of the original input images. The skip connections between corresponding feature maps are the operations that combine the forward information, i.e., the extracted features in the encoder. It is easy to understand that the encoder is merely implemented using a series of convolutional (Conv) layers and the decoder contains several deconvolutional (Deconv) layers to further increase the network depth. However, the direct utilization of the existing encoder-decoder structure cannot ensure satisfactory performance. The reason is that we need to add more convolutional layers in the encoder/decoder blocks for deblurring tasks, which makes the network converge slowly. Moreover, it is difficult to obtain the spatial information from the middle feature maps with the small size for reconstruction. Hence, we chose to transform the encoder-decoder model into an encoder-decoder ResBlock network, as shown in Figure 2, which acts as our bottom-top backbone structure. Each backbone structure is comprised of one convolutional layer with a stride of two followed by several ResBlocks. The number of kernels at the convolutional layers is doubled for the successive backbone structure. Each ResBlock includes two convolutional layers with the same number of kernels. The network structure of the decoder is symmetric to that of the encoder. Therefore, the resolution of output images for the transformed encoder-decoder ResBlock network is the same as that of the original input images. It is highly significant to produce a large receptive field for extracting more feature information while making network training simpler. The encoder and the decoder are connected by the SENet, which can realize the information exchange between the channels, enhance the useful features, and suppress the features that are not important for the current task, thereby facilitating extraction of the key information in the feature maps at different channels with a reduction ratio of 64. For the SENet, its core operations include squeeze and excitation. The function of the squeeze operation is to encode the spatial features on a channel into a global feature, thereby allowing information from the global receptive field to be used by different layers. The aggregation is followed by an excitation operation, which adopts a simple self-gating mechanism to produce a collection of per-channel modulation weights. The learned channel weights are multiplied by the original input features to generate the channel feature X c n . It is given by
Z c n = 1 W f e a t u r e × H f e a t u r e i = 1 W f e a t u r e j = 1 H f e a t u r e M c n ( i , j )
S c n = σ ( W 2 θ ( W 1 Z c n ) )
X c n = M c n S c n
where Z c n represents the global spatial information of the n-th scale image at the c-th channel; M c n is the feature map produced by the encoder; W f e a t u r e and H f e a t u r e are the width and height of M c n , respectively; W 1 c r s e × c and W 2 c × c r s e are the full connection operations with r s e denoting the channel reduction ratio; θ means rectified linear unit (ReLu) activation [36]; σ denotes the sigmoid activation function; and S c n is the channel-wise weight. From Equation (4), it can be seen that the channel-wise multiplication between S c n and M c n generates X c n .

2.1.2. Structure of Multi-Stream Bottom-Top-Bottom Attention Network

Considering that Gaussian blur is sensitive to scale, we duplicated the designed bottom–top–bottom attention network in parallel. Specifically, the multi-scale versions of the input blurry images generated by the resizing operation are first encoded in the bottom–top stream. Then, for each input image, its scaled version B n replicates the first scale from the bottom layer to the top layer with a series of residual blocks, and a multi-stream output feature map O n (n = 1, 2, …, Ns with Ns denoting the number of scales) with the same resolution to B n is produced. Finally, these maps are resized to the same resolution to the original input image for the next network (see Figure 1).

2.2. GIFRNet

The GIFRNet consists of two sub-networks: GIFNet and GIRNet. The GIFNet integrates the maps generated by the MBANet to produce the feature map R f i n a l . Then, the GIRNet gradually refines R f i n a l to produce the final deblurred image I f i n a l .

2.2.1. GIFNet

The detailed structure of the GIFNet is shown in Figure 3a. As shown in Figure 3a, the dense spatial information of source images is combined with the multi-stream output maps, which are produced by the multi-stream network with the multi-scale structures to further increase the correlation of the concatenated maps. The original image B and the multi-stream output maps ( O 1 , O 2 , , O N s ) are firstly concatenated into a feature map F 0 with 3Ns + 3 channels. Then we add the global context block (GCBlock) [37], which can help establish a connection between two pixels with a certain distance to capture the spatial information in the feature images, thereby facilitating the fusion of the multi-stream information with the original image. The architecture of the GCBlock is shown in Figure 3c. It includes the context modeling module and the transform. The context modeling module groups the features of all positions together via weighted averaging to extract the global information. In order to reduce the calculation complexity, the channel compression transform is adopted. Furthermore, the layer normalization (LayerNorm) normalizes the mean and variance of all summed inputs to the neurons in one layer to improve the training speed, accuracy, and robustness of the model. The output features are fed to a series of convolutional and ReLU layers. In the convolutional layers, the kernel size is set to 3 × 3, and the channel number is set to 64, 128, 64, and 3, respectively. A merged map R f i n a l with the same resolution as the original image is produced as the final output. The GIFNet can be expressed as
F 0 = c a t ( O 1 , O 2 , , O N s , B )
F 1 = G C B l o c k ( F 0 )
F t = ReLu ( W t F t 1 + b t )
R f i n a l = W T F T 1 + b T
where c a t denotes the concatenation operator; F 1 means the feature information produced by the GCBlock with the channel compression ratio set to 2; F t denotes the feature image generated at the t-th (t = 2, 3, …, T) convolutional layer; W t and b t represent the convolutional filter and bias of t-th convolutional layer, respectively; and the GIFNet fully utilizes the dense spatial information of the original image and nonlinearly fuses the multi-stream output maps.

2.2.2. GIRNet

Although the GIFNet can integrate the feature maps at different scales effectively and improve spatial correlation by introducing the source image, errors may be introduced into the restored image. In order to address this issue, we introduced the GIRNet to further refine the output deblurred image. Figure 3b shows the network structure of the GIRNet, which has the same structure to the GIFNet but with different parameters. At each iteration, we feed both the original image and the input map through the GIRNet to produce the refined map, which in turn serves as the input map at the next iteration. The final restored image I f i n a l can be expressed as
I f i n a l = ( ( ( R f i n a l ; B , W m , b m ) ) )
where R f i n a l is the preceding integrated map as the first input map. The operator means a recursive composition function. W m and b m represent the convolutional filter and the bias of the GIRNet, respectively. The final restored image is produced by continuously correcting its previous errors until the last iteration is implemented.

2.3. Loss

We used Euclidean loss and perceptual loss [38] as the total loss to measure the difference between the network output and the ground truth. Here, the Euclidean loss is the classic mean square error (MSE) loss. Unlike the pixel-level MSE loss, the perceptual loss utilizes high-dimensional features obtained from high-performance convolutional neural networks and can help restore more details for the deblurred images. The proposed total loss is defined as
L M S E = I f i n a l I o 2 2
L c = ϕ k , l ( I f i n a l ) ϕ k , l ( I o ) 2 2
L t o t a l = L M S E + λ L c
where λ is the parameter to balance L M S E and L c , and ϕ k , l denotes the feature extractor for extracting the image features at the l-th convolutional layer (after activation) before the k-th pooling layer in the VGG16 network pre-trained on the ImageNet dataset. The features extracted by ϕ k , l are used to measure the difference between the high-level semantic information of images I f i n a l and I o . The Adam optimizer [39] is adopted to minimize the cost function.

2.4. Model Training

To train the proposed networks, we used the realistic and dynamic scenes (REDS) dataset [40], in which each image was of size 720 × 1280. It was recorded by the GoPro HERO6 Black camera at 120 fps. There were 300 sequences in total, and each one included 100 pairs of sharp and blurry frames. We randomly selected 1120 sharp images as the ground truth for training and manually synthesized the blurry images by convolving the sharp images with Gaussian blur kernels of size 17 × 17. The standard deviations of the blur kernels were changed from 1.6 to 2.4 with steps of 0.2 to generate 5 blur kernels. The blurry images are sampled and randomly cropped into 256 × 256 image patches as training images. The ground-truth patches were generated in the same way. In total, 1,120,000 patches were generated for training.
We realized the proposed networks with TensorFlow 1.9.0. The proposed MBANet and GIFRNet were jointly trained on a Ubuntu 16.04 with an Intel I7-6950 CPU and 96G RAM. The NVIDIA 1080Ti GPU with CUDA 10.0 was used for acceleration. For the proposed networks, the initial learning rate was set to 0.0001, and it was exponentially decayed to 1 × 10−6 at 1000 epochs using power 0.3. At each iteration, we fixed the batch size to be 10. All trainable variables were initialized by Xavier [41]. The training took approximately 1.5 days.

3. Experimental Results and Discussion

In this section, we carried out experiments on the chosen REDS dataset and GoPro dataset [42] to determine the key parameters in the proposed networks and evaluate their performance. The testing data were generated in the same way as the training data. For each type of dataset, 10 images were randomly selected as testing images. To evaluate the deblurring performance of the proposed networks, we adopted such indexes as peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) [43], which are defined as follows:
PSNR = 10 log 10 ( 255 2 1 W H i = 1 W j = 1 H ( I f i n a l ( i , j ) I o ( i , j ) ) 2 )
SSIM = ( 2 I ¯ f i n a l I ¯ o + C 1 ) ( 2 δ I f i n a l I o + C 2 ) ( ( I ¯ f i n a l ) 2 + ( I ¯ o ) 2 + θ 1 ) ( δ I f i n a l 2 + δ I o 2 + θ 2 )
where W and H are the width and height, respectively, of the deblurred image I f i n a l . δ I f i n a l and δ I o are the standard deviations of I f i n a l and I o , respectively. I ¯ f i n a l and I ¯ o are the mean of I f i n a l and I o , respectively; δ I f i n a l I o is the covariance of I f i n a l and I o . C 1 and C 2 are the small constants to stabilize SSIM.

3.1. Parameter Setting

As for the parameters in the proposed networks, they mainly include the number of scales (Ns) in the MBANet, the iterative times (Tr) in the GIRNet, the regularization parameter λ for balancing the content loss, and the perceptual loss and parameters k and l at the feature layer related to the perceptual loss. In the loss function of the proposed networks, the content loss and the perceptual loss are of the same order of magnitude. Following the selection method of parameters in the loss function in other deep learning models for image deblurring, we set λ = 1 and k = l = 3. Here, we discuss how to determine the two parameters Ns and Tr.

3.1.1. Number of Scales

As for the number of scales in the MBANet, it is easy to understand that too large an Ns value brings much calculation burden while too small a value is disadvantageous for extracting the effective features from the input images. Here, we trained the proposed model using three different scales for comparison to analyze the contributions of Ns. The PSNR and SSIM of the three models performed on the REDS dataset and GoPro dataset are shown in Figure 4 and Figure 5, respectively. Clearly, the three-scale model and the two-scale model could only perform well on the REDS dataset and GoPro dataset, respectively. However, the four-scale model could obtain better or competitive results on both the REDS dataset and GoPro dataset in terms of PSNR and SSIM. Therefore, we chose four scales as a compromise.

3.1.2. Iterative Times

For the iterative times Tr in the GIRNet, we found that too large a Tr increased the training time of the proposed networks greatly, while one that was too small could not ensure the satisfactory deblurring performance. To demonstrate the influence of Tr, we trained the proposed model using three different iterative times. The quantitative results are shown in Figure 6 and Figure 7. From Figure 6 and Figure 7, we can see that the proposed model with Tr = 2 performed the worst on the REDS dataset, and our model with Tr = 1 performed the worst on the GoPro dataset. By comparison, our model with Tr = 3 could achieve satisfactory performance on the two datasets. Based on the above analysis, we chose Tr = 3 in the proposed networks.

3.2. Comparison with Different Network Structures

In order to verify the effectiveness of the structure in the proposed networks, we removed GCBlock from GIFNet and GIRNet. The resultant networks were named as fusion network (FNet) and reconstruction network (RNet), respectively. Here, we compare the three models, i.e., MBANet with FNet (MBANet + FNet), MBANet with FNet and RNet (MBANet + FNet + RNet), and MBANet with GIFNet and GIRNet (MBANet + GIFNet + GIRNet). We also used ten images chosen from the REDS dataset as the sharp images for testing. The quantitative results of PSNR and SSIM for the three models are shown in Figure 8. It can be seen from Figure 8 that MBANet + GIFNet + GIRNet achieved the best performance among the three models, which indeed indicates that the adoption of attention mechanism and reconstruction network structure improved the deblurring performance effectively.

3.3. Comparison with State-of-the-Art Deblurring Methods

To demonstrate the advantage of the proposed method, we compared it with five state-of-the-art image deblurring algorithms including three well-known model-based deblurring methods such as EPLL [14], NCSR [13], and ECP [20], and two deep learning methods such as SRN [21] and DPDNN [31]. For the compared algorithms, we firstly followed the parameter settings suggested by the authors and then tuned them to ensure the best results for the corresponding testing images based on the comprehensive consideration of PSNR and SSIM as well as subjective human vision. For fairness, unless noted otherwise, all experiments were conducted for the evaluated methods on the same dataset with the same parameter configuration.

3.3.1. Comparison on REDS Dataset

Experiments were done on the color images in the REDS dataset. Table 1 and Table 2 list the PSNR and SSIM values of all evaluated methods. From Table 1, we can see that the deep learning methods outperformed the ECP, EPLL, and NCSR methods by providing significantly higher PSNR. Compared with the SRN and DPDNN methods, the proposed algorithm, on average, provided PSNR improvements by 0.42 dB and 0.27 dB, respectively. Table 2 shows that the SSIM values of the conventional model-based methods were lower than those of the deep learning methods. Meanwhile, the proposed method could provide slightly higher SSIM than the SRN and DPDNN methods.
Figure 9, Figure 10, Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 show the deblurred results and the corresponding difference images of all evaluated methods. Here, we used the red boxes to indicate where the proposed method was visually superior to other state-of-the-art methods. As shown in Figure 9 and Figure 11, it can be seen that our proposed method could recover the thin lines in Figure 9 and the boundaries of stone steps in Figure 11 better than the other compared methods by providing more complete and sharper edges for them. It is shown in Figure 12 that the multi-scale deblurring approaches such as the proposed method and the SRN method gained an obvious advantage over other methods in recovering small-scale details such as the vertical thin iron rod marked with one red box. For example, the DPDNN method damaged the thin iron rod so seriously that it was almost invisible in the deblurred image while it was preserved significantly better by the SRN method and the proposed method. Furthermore, the proposed method outperformed the SRN method by providing a sharper restored result for the complex structures marked with another red box. As seen in Figure 14, the textures were so complicated and densely arranged that their restoration was very challenging. Clearly, the other compared methods led to the loss of textures in the deblurred images to some extent. However, the proposed method could still recover the textures effectively. Moreover, Figure 15 shows that in the case of large deviations of blur kernels, our method could preserve such image details as the number on the car more effectively than other evaluated methods. Furthermore, from the difference images between the original image and the deblurred images shown in Figure 10, Figure 13 and Figure 16, we can see that the traditional methods performed worse than the deep learning-based methods because there was too much structural information in their difference images. Compared with the SRN and DPDNN methods, the proposed method performed better in that it produced less obvious structural information, which further proves the superiority of the proposed method.
In order to further verify the robustness of the proposed method, we selected four different levels of Gaussian blur kernels with standard deviations ranging from 1.7 to 2.3 with a step of 0.2. These blurred images were not used as training sets for model training. The PSNR and SSIM values of all evaluated methods are listed in Table 3 and Table 4. It can be seen that the deep learning-based methods still performed better than the traditional methods. Compared with the SRN and DPDNN methods, the proposed method provided slightly higher SSIM values and it, on average, provided improvements of PSNR by 0.52 dB and 0.6 dB, respectively. The above comparison indeed demonstrated good robustness of the proposed method.

3.3.2. Comparison with State-of-the-Art Methods on the GoPro Dataset

To further evaluate the performance of the proposed method, we used the GoPro dataset, which is different from the REDS dataset, to compare it with other deblurring methods. The quantitative results computed on the 10 images selected from this dataset are listed in Table 5 and Table 6. The results show that the deep learning-based algorithms still performed better than the traditional algorithms in PSNR and SSIM. The SRN performed worse than the DPDNN and the proposed method. Although the proposed method provided slightly lower PSNR and SSIM results than the DPDNN when the deviation was 2.4, it produced higher average PSNR and SSIM than the latter.
Figure 17 shows the visual comparisons of the deblurred images for three deep learning-based deburring methods performed on the GoPro dataset, where the standard deviation of the Gaussian blur kernel in the test image was 2.2. From Figure 17, it can be seen that the SRN and the DPDNN produced the unwanted artefacts around the horizontal lines in the red box to different extents. By comparison, our method could generate more similar deblurred result as the unblurred sharp image without introducing the artefacts due to the combination of the features from the original image with the global context information. Indeed, the visual comparison demonstrated the superiority of the proposed method to other deep learning-based methods in recovering image details.

3.4. Discussion

In this work, we proposed the novel multi-stream bottom–top–bottom attention network as well as the global information-based fusion and construction network for restoring Gaussian blurred images. From the above experimental results, it can be seen that the proposed method gains an advantage over other compared methods in terms of deblurring performance and robustness. However, this work has some limitations. Firstly, the proposed networks are only used for the deblurring of Gaussian blurred images. Future work will be focused on extending our method to address other kinds of blur such as motion blur and the hybrid blur. Such extension will involve the modification of network structure such as the introduction of dense blocks. Secondly, the training of the proposed networks is very time-consuming. The model training time can be shortened by simplifying the network structure. Finally, the loss function we used is relatively simple, and thus there may be possibilities for further improvement of deblurred results. The effective structural preserving loss item can be explored and additionally introduced to address this problem.

4. Conclusions

In this paper, we proposed the novel multi-stream bottom–top–bottom attention network as well as the global information-based fusion and construction network for restoring Gaussian blurred images. Distinctively, the proposed network fully combines high-level semantics with low-level image features and channel-wise information to generate the deblurred images. Besides, the multi-stream mechanism fully takes advantage of multi-scale information, and this structure is easier to train by sharing the parameters. Additionally, the reconstruction network can further refine the restored image by combining the global context information. Experimental results show that the proposed method can provide better deblurring performance than other state-of-the-art algorithms both quantitatively and qualitatively.

Author Contributions

X.Z. and M.D. conceived the deblurred method and designed the experiments; Q.Z. performed the experiments and wrote the code of the deblurred method as well as the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 61871440.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Flusser, J.; Farokhi, S.; Höschl, C.; Suk, T.; Zitová, B.; Pedone, M. Recognition of Images Degraded by Gaussian Blur. IEEE Trans. Image Process. 2015, 25, 790–806. [Google Scholar] [CrossRef] [PubMed]
  2. Li, J.; Liu, Z. Ensemble Dictionary Learning for Single Image Deblurring via Low-Rank Regularization. Sensors 2019, 19, 1143. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Liu, D.; Chen, X.; Shi, C.; Liu, X. Star Image Prediction and Restoration under Dynamic Conditions. Sensors 2019, 19, 1890. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Li, J.; Gong, W.; Li, W. Combining Motion Compensation with Spatiotemporal Constraint for Video Deblurring. Sensors 2018, 18, 1774. [Google Scholar] [CrossRef] [Green Version]
  5. Yang, Z.; Yang, Z.; Gui, G. A Convex Constraint Variational Method for Restoring Blurred Images in the Presence of Alpha-Stable Noises. Sensors 2018, 18, 1175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Yang, F.; Huang, Y.; Luo, Y.; Li, L.; Li, H. Robust Image Restoration for Motion Blur of Image Sensors. Sensors 2016, 16, 845. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Pan, J.; Hu, Z.; Su, Z.; Yang, M.-H. L-0-regularized intensity and gradient prior for deblurring text images and beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 342–355. [Google Scholar] [CrossRef]
  8. Fergus, R.; Singh, B.; Hertzmann, A.; Roweis, S.T.; Freeman, W.T. Removing camera shake from a single photograph. ACM Trans. Graph. 2006, 25, 787–794. [Google Scholar] [CrossRef]
  9. Levin, A.; Fergus, R.; Durand, F.; Freeman, W.T. Image and depth from a conventional camera with a coded aperture. ACM Trans. Graph. 2007, 26, 787–794. [Google Scholar] [CrossRef]
  10. Wang, Y.; Yang, J.; Yin, W.; Zhang, Y. A New Alternating Minimization Algorithm for Total Variation Image Reconstruction. SIAM J. Imaging Sci. 2008, 1, 248–272. [Google Scholar] [CrossRef]
  11. Krishnan, D.; Tay, T.; Fergus, R. Blind deconvolution using a normalized sparsity measure. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 20–25 June 2011; pp. 233–240. [Google Scholar]
  12. Parameswaran, S.; Deledalle, C.-A.; Denis, L.; Nguyen, T.Q. Accelerating GMM-based patch priors for image restoration: Three ingredients for a 100x speed-up. IEEE Trans. Image Process. 2018, 28, 687–698. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Dong, W.; Zhang, K.; Shi, G.; Li, X. Nonlocally Centralized Sparse Representation for Image Restoration. IEEE Trans. Image Process. 2012, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Zoran, D.; Weiss, Y. From learning models of natural image patches to whole image restoration. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 479–486. [Google Scholar]
  15. Danielyan, A.; Katkovnik, V.; Egiazarian, K. BM3D Frames and Variational Image Deblurring. IEEE Trans. Image Process. 2011, 21, 1715–1728. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Michaeli, T.; Irani, M. Blind Deblurring Using Internal Patch Recurrence. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 783–798. [Google Scholar]
  17. Shan, Q.; Jia, J.; Agarwala, A. High-quality motion deblurring from a single image. ACM Trans. Graph. 2008, 27, 557–566. [Google Scholar] [CrossRef]
  18. Pan, J.; Sun, D.; Pfister, H.; Yang, M. Blind image deblurring using dark channel prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1628–1636. [Google Scholar]
  19. Chen, L.; Fang, F.; Wang, T.; Zhang, G. Blind image deblurring with local maximum gradient prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1742–1750. [Google Scholar]
  20. Yan, Y.; Ren, W.; Guo, Y.; Wang, R.; Cao, X. Image deblurring via extreme channels prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolululu, HI, USA, 21–26 July 2017; pp. 6978–6986. [Google Scholar]
  21. Tao, X.; Gao, H.; Shen, X.; Wang, J.; Jia, J. Scale-recurrent network for deep image deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 8174–8182. [Google Scholar]
  22. Liu, X.; Ma, Y.; Shi, Z.; Chen, J. Griddehazenet: Attention-based multi-scale network for image dehazing. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 October–2 November 2019; pp. 7313–7322. [Google Scholar]
  23. Sun, J.; Cao, W.; Xu, Z.; Ponce, J. Learning a convolutional neural network for non-uniform motion blur removal. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 769–777. [Google Scholar]
  24. Yan, R.; Shao, L. Blind image blur estimation via deep learning. IEEE Trans. Image Process. 2016, 25, 1910–1921. [Google Scholar] [CrossRef]
  25. Cheng, S.; Liu, R.; He, Y.; Fan, X.; Luo, Z. Blind image deblurring via hybrid deep priors modeling. Neurocomputing 2020, 387, 334–345. [Google Scholar] [CrossRef]
  26. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef] [Green Version]
  27. Pan, J.; Dong, J.; Liu, Y.; Zhang, J.; Ren, J.; Tang, J.; Tai, Y.W.; Yang, M.-H. Physics-Based Generative Adversarial Models for Image Restoration and Beyond. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef] [Green Version]
  28. Fei, X.; Zhao, J.; Zhao, H.; Yun, D.; Zhang, Y. Deblurring adaptive optics retinal images using deep convolutional neural networks. Biomed. Opt. Express 2017, 8, 5675–5687. [Google Scholar] [CrossRef] [Green Version]
  29. Zhang, K.; Zhang, Y.; Zhong, Y.; Ma, L.; Liu, W.; Li, H.; Huang, Y.; Du, Y.; Wang, L. Adversarial Spatio-Temporal Learning for Video Deblurring. IEEE Trans. Image Process. 2019, 28, 291–301. [Google Scholar] [CrossRef] [Green Version]
  30. Wu, J.; Yu, X.; Liu, D.; Chandraker, M.; Wang, Z. DAVID: Dual-Attentional Video Deblurring. In Proceedings of the 2020 IEEE Winter Conference on Applications of Computer Vision (WACV), Aspen, CO, USA, 1–5 March 2020; pp. 2365–2374. [Google Scholar]
  31. Dong, W.; Wang, P.; Yin, W.; Shi, G.; Wu, F.; Lu, X. Denoising Prior Driven Deep Neural Network for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2305–2318. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Minaee, S.; Wang, Y. An ADMM Approach to Masked Signal Decomposition Using Subspace Representation. IEEE Trans. Image Process. 2019, 28, 3192–3204. [Google Scholar] [CrossRef] [Green Version]
  33. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Zhao, W.; Zhao, F.; Wang, N.; Lu, H. Defocus Blur Detection via Multi-Stream Bottom-Top-Bottom Network. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 1884–1897. [Google Scholar] [CrossRef] [PubMed]
  35. Shan, H.; Zhang, Y.; Yang, Q.; Kruger, U.; Kalra, M.K.; Sun, L.; Cong, W.; Wang, G. 3D Convolutional Encoder-Decoder Network for Low-Dose CT via Transfer Learning From a 2D Trained Network. IEEE Trans. Med Imaging 2018, 37, 2750. [Google Scholar] [CrossRef]
  36. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the International Conference on Machine Learning (ICML), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  37. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. GCNet: Non-Local Networks Meet Squeeze-Excitation Networks and Beyond. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, 27 October 2019; pp. 1971–1980. [Google Scholar]
  38. Johnson, J.; Alahi, A.; Li, F.-F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 694–711. [Google Scholar]
  39. Kingma, D.P.; Ba, L.J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  40. Nah, S.; Baik, S.; Hong, S.; Moon, G.; Son, S.; Timofte, R.; Lee, K.M. NTIRE 2019 Challenge on Video Deblurring and Super-Resolution: Dataset and Study. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–20 June 2019; pp. 1996–2005. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 11–18 December 2015; pp. 1026–1034. [Google Scholar]
  42. Nah, S.; Kim, T.H.; Lee, K.M. Deep Multi-scale Convolutional Neural Network for Dynamic Scene Deblurring. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 257–265. [Google Scholar]
  43. Badri, H.; Yahia, H.; Aboutajdine, D. Low-Rankness Transfer for Realistic Denoising. IEEE Trans. Image Process. 2016, 25, 5768–5779. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The flowchart of the proposed method. Each colored box is considered to be a feature block, and the arrows between blocks represent the information stream.
Figure 1. The flowchart of the proposed method. Each colored box is considered to be a feature block, and the arrows between blocks represent the information stream.
Sensors 20 03724 g001
Figure 2. The architecture of the bottom–top–bottom attention network.
Figure 2. The architecture of the bottom–top–bottom attention network.
Sensors 20 03724 g002
Figure 3. The architecture of the sub-network models. (ac): GIFNet, GIRNet, and GCBlock, respectively. Here, ( C , H , W ) means a feature map with the number C of channels, height H , and width W ; r denotes channel compression ratio; represent matrix multiplication; means broadcast element-wise addition.
Figure 3. The architecture of the sub-network models. (ac): GIFNet, GIRNet, and GCBlock, respectively. Here, ( C , H , W ) means a feature map with the number C of channels, height H , and width W ; r denotes channel compression ratio; represent matrix multiplication; means broadcast element-wise addition.
Sensors 20 03724 g003
Figure 4. The peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) of the proposed model using different numbers of scales on the realistic and dynamic scenes (REDS) dataset: (a) PSNR; (b) SSIM.
Figure 4. The peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) of the proposed model using different numbers of scales on the realistic and dynamic scenes (REDS) dataset: (a) PSNR; (b) SSIM.
Sensors 20 03724 g004
Figure 5. The PSNR and SSIM of the proposed model using different numbers of scales on the GoPro dataset: (a) PSNR; (b) SSIM.
Figure 5. The PSNR and SSIM of the proposed model using different numbers of scales on the GoPro dataset: (a) PSNR; (b) SSIM.
Sensors 20 03724 g005
Figure 6. The PSNR and SSIM of the proposed model using different iterative times on the REDS dataset: (a) PSNR; (b) SSIM.
Figure 6. The PSNR and SSIM of the proposed model using different iterative times on the REDS dataset: (a) PSNR; (b) SSIM.
Sensors 20 03724 g006
Figure 7. The PSNR and SSIM of the proposed model using different iterative times on the GoPro dataset: (a) PSNR; (b) SSIM.
Figure 7. The PSNR and SSIM of the proposed model using different iterative times on the GoPro dataset: (a) PSNR; (b) SSIM.
Sensors 20 03724 g007
Figure 8. The PSNR and SSIM of different network models: (a) PSNR; (b) SSIM.
Figure 8. The PSNR and SSIM of different network models: (a) PSNR; (b) SSIM.
Sensors 20 03724 g008
Figure 9. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 1.6. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Figure 9. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 1.6. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Sensors 20 03724 g009
Figure 10. Difference images between the region marked with the green box in Figure 9a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Figure 10. Difference images between the region marked with the green box in Figure 9a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Sensors 20 03724 g010
Figure 11. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 1.8. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Figure 11. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 1.8. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Sensors 20 03724 g011
Figure 12. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.0. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Figure 12. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.0. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Sensors 20 03724 g012
Figure 13. Difference images between the region marked with the green box in Figure 12a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Figure 13. Difference images between the region marked with the green box in Figure 12a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Sensors 20 03724 g013
Figure 14. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.2. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Figure 14. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.2. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Sensors 20 03724 g014
Figure 15. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.4. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Figure 15. Deblurred results for test image with 17 × 17 Gaussian blur kernel of deviation 2.4. The first row shows the sharp image and deblurred images for the evaluated methods. The second row shows the enlarged views of the region marked with the green box in the original image and the deblurred images. (a) Original image [37]; (b) ECP; (c) EPLL; (d) NCSR; (e) SRN; (f) DPDNN; (g) proposed method.
Sensors 20 03724 g015
Figure 16. Difference images between the region marked with the green box in Figure 15a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Figure 16. Difference images between the region marked with the green box in Figure 15a and its deblurred results. (a) ECP; (b) EPLL; (c) NCSR; (d) SRN; (e) DPDNN; (f) proposed method.
Sensors 20 03724 g016
Figure 17. Deblurred results for three deep learning methods performed on the GoPro test image with 17 × 17 Gaussian blur kernel of deviation 2.2. The first row shows the sharp image and deblurred images by the evaluated methods. The second row shows the enlarged views of the region marked with green box in the original image and the deblurred images. (a) original image [39]; (b) SRN; (c) DPDNN; (d) proposed method.
Figure 17. Deblurred results for three deep learning methods performed on the GoPro test image with 17 × 17 Gaussian blur kernel of deviation 2.2. The first row shows the sharp image and deblurred images by the evaluated methods. The second row shows the enlarged views of the region marked with green box in the original image and the deblurred images. (a) original image [39]; (b) SRN; (c) DPDNN; (d) proposed method.
Sensors 20 03724 g017
Table 1. The PSNR results of deblurred images for all evaluated methods.
Table 1. The PSNR results of deblurred images for all evaluated methods.
MethodsStandard DeviationAverage
1.61.82.02.22.4
ECP29.7229.4829.2628.9428.1829.12
EPLL33.2532.2931.4530.7330.1231.57
NCSR33.9032.6631.6630.8530.1631.85
SRN36.5535.5334.6633.8533.0534.73
DPDNN36.9335.7834.7633.8933.0434.88
Proposed37.0336.0035.0534.2333.4035.14
Table 2. The SSIM results of the deblurred images for all evaluated methods.
Table 2. The SSIM results of the deblurred images for all evaluated methods.
MethodsStandard DeviationAverage
1.61.82.02.22.4
ECP0.89050.86560.84830.83870.81710.8520
EPLL0.93110.91680.90250.88820.87430.9026
NCSR0.93630.91960.90380.88840.87360.9043
SRN0.96220.95180.94170.93130.91980.9414
DPDNN0.96430.95400.94310.93240.92110.9430
Proposed0.96500.95570.94570.93570.92450.9453
Table 3. The PSNR results of deblurred images for all evaluated methods.
Table 3. The PSNR results of deblurred images for all evaluated methods.
MethodsStandard DeviationAverage
1.71.92.12.3
ECP29.5529.4029.0728.3629.10
EPLL32.0931.2830.5629.9430.97
NCSR33.2432.1531.2430.4931.78
SRN35.0034.3433.6332.9133.97
DPDNN34.9934.1333.6132.8133.89
Proposed35.6134.9834.1733.2034.49
Table 4. The SSIM results of the deblurred images for all evaluated methods.
Table 4. The SSIM results of the deblurred images for all evaluated methods.
MethodsStandard DeviationAverage
1.71.92.12.3
ECP0.87050.85750.84190.82170.8479
EPLL0.92390.90950.89510.88100.9024
NCSR0.92790.911570.89590.88100.9041
SRN0.94970.94090.93120.92160.9359
DPDNN0.95130.94090.93150.92130.9363
Proposed0.95630.94780.93850.92930.9430
Table 5. The PSNR results of the deblurred images in the GoPro dataset.
Table 5. The PSNR results of the deblurred images in the GoPro dataset.
MethodsStandard DeviationAverage
1.61.82.02.22.4
ECP32.0732.0832.0931.5131.2931.81
EPLL32.2431.5230.6829.9629.3630.75
NCSR36.5334.6733.1931.9630.9633.46
SRN37.1734.9734.5232.9932.0234.33
DPDNN37.5836.4935.1033.8233.0835.21
Proposed37.9936.4835.1634.0432.9935.33
Table 6. The SSIM results of the deblurred images in the GoPro dataset.
Table 6. The SSIM results of the deblurred images in the GoPro dataset.
MethodsStandard DeviationAverage
1.61.82.02.22.4
ECP0.88610.85880.84940.82970.81540.8479
EPLL0.96110.95480.94580.93470.92280.9438
NCSR0.97420.96360.95140.93770.92350.9501
SRN0.97770.96870.96480.95290.94300.9614
DPDNN0.98080.97600.96910.95980.95490.9681
Proposed0.98290.97600.96990.96190.95240.9686

Share and Cite

MDPI and ACS Style

Zhou, Q.; Ding, M.; Zhang, X. Image Deblurring Using Multi-Stream Bottom-Top-Bottom Attention Network and Global Information-Based Fusion and Reconstruction Network. Sensors 2020, 20, 3724. https://doi.org/10.3390/s20133724

AMA Style

Zhou Q, Ding M, Zhang X. Image Deblurring Using Multi-Stream Bottom-Top-Bottom Attention Network and Global Information-Based Fusion and Reconstruction Network. Sensors. 2020; 20(13):3724. https://doi.org/10.3390/s20133724

Chicago/Turabian Style

Zhou, Quan, Mingyue Ding, and Xuming Zhang. 2020. "Image Deblurring Using Multi-Stream Bottom-Top-Bottom Attention Network and Global Information-Based Fusion and Reconstruction Network" Sensors 20, no. 13: 3724. https://doi.org/10.3390/s20133724

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop