Next Article in Journal
Low Spatial Peak-to-Average Power Ratio Transmission for Improved Energy Efficiency in Massive MIMO Systems
Previous Article in Journal
A WoT-Based Method for Creating Digital Sentinel Twins of IoT Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Multi-Frame Super-Resolution Based on Adaptive Half-Quadratic Function and Local Structure Tensor Weighted BTV

1
College of Computer Science, Sichuan University, Chengdu 610065, China
2
Enrollment and Employment Department, Sichuan Normal University, Chengdu 610066, China
3
Science and Technology Branch, Southwest Jiaotong University Press, Chengdu 610031, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5533; https://doi.org/10.3390/s21165533
Submission received: 14 July 2021 / Revised: 11 August 2021 / Accepted: 13 August 2021 / Published: 17 August 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
It is difficult to improve image resolution in hardware due to the limitations of technology and too high costs, but most application fields need high resolution images, so super-resolution technology has been produced. This paper mainly uses information redundancy to realize multi-frame super-resolution. In recent years, many researchers have proposed a variety of multi-frame super-resolution methods, but it is very difficult to preserve the image edge and texture details and remove the influence of noise effectively in practical applications. In this paper, a minimum variance method is proposed to select the low resolution images with appropriate quality quickly for super-resolution. The half-quadratic function is used as the loss function to minimize the observation error between the estimated high resolution image and low-resolution images. The function parameter is determined adaptively according to observation errors of each low-resolution image. The combination of a local structure tensor and Bilateral Total Variation (BTV) as image prior knowledge preserves the details of the image and suppresses the noise simultaneously. The experimental results on synthetic data and real data show that our proposed method can better preserve the details of the image than the existing methods.

1. Introduction

Super Resolution (SR) refers to the process of reconstructing a High Resolution (HR) image from a single Low Resolution (LR) image or multiple low resolution images by the software method without modifying the hardware environment. SR is often used in satellite remote sensing [1], medical diagnosis [2] and video surveillance [3] and so forth. According to the number of input low resolution images, it can be divided into Single Image Super Resolution (SISR) and Multi-Frame Super Resolution (MFSR). This paper mainly studies MFSR.
The first MFSR was proposed [4] in 1984, and is based on the frequency domain. The frequency domain method is easy to implement and cheap computationally. However, this method has no way to add image prior knowledge, and can only achieve good results for images without noise and degradation, which is not feasible in practical applications. Therefore, algorithms in the spatial domain are widely studied. There are interpolation-based methods [5,6], reconstruction-based methods [7,8,9,10] and learning-based methods [11,12]. With the popularity of deep learning architectures, many deep learning networks are implemented for MFSR [13,14,15,16]. Learning-based methods need a large number of external training sets; the effect is closely related to the type and number of training sets. Although learning-based SR methods have been hot topics of research in the last few decades, reconstruction-based methods, which do not require training sets, are very popular. The quality of the reconstructed image only depends on our mathematical model and input LR images. This paper mainly studies reconstruction-based MFSR methods that reconstruct an HR image by using the redundancy and supplementary information between multiple LR images in the same scene. In order to find useful information redundancy, an important condition for realizing MFSR is to have sub-pixel displacement as shown in Figure 1.
Because SR itself is an ill-posed problem, regularization techniques are widely used to solve the minimization problem. The methods of regularization mainly contain the fidelity term and the regularization term. The purpose of the fidelity term is to minimize the observation errors between the reconstructed HR image and the input LR images. The purpose of the regularization term is to make the reconstructed HR image reach a robust state.
The widely-used fidelity terms in the SR are the L1 norm and the L2 norm. Both the L1 norm and the L2 norm have their advantages and drawbacks. The L2 norm can produce lower variance than the L1 norm. However, the L2 norm is sensitive to outliers. M-estimators are explored in MFSR later. For example, the Huber function [17] is proposed for the fidelity term. In [18], the Lorentzian norm is applied to MFSR to increase the robustness. In [19], the performances of the Tukey, Lorentzian and Huber norms are studied concerning outliers. A locally adaptive L1–L2 fidelity norm is also proposed in [20]. Xueying Zeng et al. [8] introduce the adaptive fidelity term based on a new M-estimator half-quadratic function and bilateral edge-preserving (BEP) regularization method. Köhler et al. [9] propose a weighted Gaussian observation model which takes into account the spatial variant noise and weighted bilateral total variation to exploit the sparsity of natural images. Xiaohong Liu et al. [10] use the half-quadratic estimation method to adaptively select the error norm and propose an adaptive bilateral total variation regularization method.
Image prior knowledge is used to regularize the image reconstruction. Tikhonov regularization [21] is one of most commonly used methods. The Total Variation (TV) family, such as Bilateral Total Variation (BTV) [7], is another popular regularization technique. The BTV uses the L1 norm to handle outliers, unlike Tikhonov regularization which is based on the L2 norm.
In recent years, many researchers have proposed a large number of MFSR methods, but MFSR is an ill-posed problem and it is very difficult to reconstruct a satisfactory HR image because the result of MFSR is affected by motion estimation, image registration, unknown blur, noise, and so forth. Even if they are studied separately, each affecting factor is extremely challenging.
In order to simultaneously preserve the image edge and texture details and remove the influence of noise, we propose a new robust MFSR method. In this paper, there are three major contributions to improving the quality of the final reconstructed HR image, as follows:
  • We propose a fast and effective method for selecting appropriate LR images, which provide a relatively good input for the reconstruction;
  • A novel fidelity term is proposed. We adopt the half-quadratic function as the error norm adaptively. All parameters are automatically adjusted according to observation errors;
  • We propose a local structure tensor weighted BTV regularization term. The novel image prior can better simultaneously preserve image details and suppress noise by assigning a certain weight to each pixel according to the local structure tensor information of the image.
The rest of this paper is organized as follows. Section 2 introduces the observation model and the basic framework of MFSR. Section 3 describes our proposed algorithm in detail. Section 4 presents our experimental results on synthetic data and real data. Section 5 summarizes our paper.

2. Observation Model and Basic Framework of MFSR

In practical applications, noise, blur, motion and other factors will affect the final SR. Therefore, the observation model of MFSR can be expressed as follows
Y k = D B k M k X + n k ,
where Y k represents the kth LR image, k = 1 , 2 , , K . The size of Y k is m n × 1 ,   K is the number of LR images.   X represents the HR image, the size is r m r n × 1 ,the down sampling factor is r . M k represents the geometric motion matrix (sub-pixel displacement) of the kth LR image, the size is m r n × r m r n . B k represents the blur matrix of the kth LR image, the size is r m r n × r m r n .   D represents the down sampling matrix, the size is m n × r m r n . n k represents the noise of the kth LR image, the size is m n × 1 .
We can simplify Equation (1) by combining D B k M k as a system matrix W k [22], and Equation (1) can be rewritten as follows
Y k = W k X + n k .
The observation error r k of the LR image Y k can be defined as
r k = | W k X ^ Y k | ,
where | · | means the absolute value and X ^ represents the estimated HR image.
The basic framework of MFSR can be expressed as
X ^ = a r g m i n X { k = 1 K | | r k | | p p + λ   γ ( X ) } ,  
where γ ( X ) is the regularization term with respect to X ,   p represents L p norm, and λ is the trade-off parameter between the fidelity term and the regularization term.
BTV is calculated simply and is easy to implement, so it is often selected in the regularization term. The formula of BTV can be expressed as
γ ( X ) = n = P P m = P P ζ | m | + | n | X S x n S y m X 1 ,  
where P represents the size of the sliding window.   ζ is a scaled weight parameter. The range of ζ is 0 < ζ < 1 . S x n ,   S y m denotes shifting X by n pixels in the horizontal direction and by m pixels in the vertical direction, respectively.

3. Proposed MFSR Algorithm

We first select appropriate LR images from a set of LR frames as the input for MFSR. The overall SR framework of this paper is to employ a coarse to fine strategy, enlarging LR images to the target image size gradually in order to adapt to the large scale factor. The half-quadratic function as the loss function is used to control residuals between the unknown HR image and LR images for the fidelity term. For the regularization term, we combine the local structure tensor information of the image and the image sparsity as the image prior knowledge to regularize the estimated HR image.

3.1. Selecting the Appropriate LR Images and Alignment

Many existing MFSR algorithms input all LR images instead of selecting input LR images. Some LR images not only do not improve the quality of the reconstructed HR image, but also consume more time and increase the computational complexity. In order to improve the practical applications of MFSR, it is necessary to select LR images with appropriate quality. In this paper, we can remove some low quality LR images quickly to SR.
The number of images required for SR is at least twice the square of the magnification factor under the ideal environment according to [23]. The number of appropriate frames K can be expressed as
r 2 K M ,
where M is the number of all LR images,   r is the amplification factor.
Although there are many quality evaluation methods [24,25,26] for image selection, most of them are computationally intensive. We need a simple and effective method to quickly select appropriate LR images, so we propose a method of minimum variance. The specific steps for selecting appropriate LR images are as follows:
  • Calculate the variance of all LR images. The variance formula of the kth LR image Y k is as follows
    s k 2 = 1 m n x = 1 m y = 1 n ( Y k Y k ¯ ) 2 ,  
    where   m , n represents the width and height of the LR image Y k , respectively, Y k ¯ is the mean value of the LR image Y k , and the calculation formula is:
    Y k ¯ = 1 m n x = 1 m y = 1 n Y k
  • Calculate the average of the variance of all LR images, which can be expressed as:
    s 2 ¯ = 1 M k = 1 M s k 2 .  
  • Calculate the difference between variance per LR image and the average of the variance which can be defined as:
    Δ s = | s k 2 s 2 ¯ | .
  • Sort Δ s and choose small Δ s corresponding LR images. Appropriate LR images are adaptively selected combining Equation (6) and small Δ s .
This paper uses the optical flow algorithm to register selected LR images.

3.2. Proposed Fidelity Term

Our proposed algorithm is based on the maximum a posteriori (MAP) estimation and Bayesian theory, the expression of the estimated HR image can be written as follows:
X ^ = a r g   max X p ( X | Y 1 , Y 2 , , Y K ) = a r g   max X p ( Y 1 , Y 2 , , Y K | X ) p ( X ) p ( Y 1 , Y 2 , , Y K ) .
Since p ( Y 1 , Y 2 , , Y K ) has no effect on the estimate X , the above formula can be rewritten as:
X ^ = a r g   max X   p ( Y 1 , Y 2 , , Y K | X ) p ( X ) .
Let us suppose that each LR image Y k is independent. The Equation (12) can be simplified to
X ^ = a r g   max X k = 1 K p ( Y k | X ) p ( X ) ,  
where p ( Y k | X ) represents the conditional probability of the LR image Y k . Given the HR image X , p ( X ) is the prior probability for the HR image X .
The Half-Quadratic (HQ) function is first proposed in [27] as a potential function. The formula is as follows:
f ( x , a ) = a a 2 + x 2 a 2 ,
where a is a positive constant.
The L1 norm, L2 norm, half-quadratic function, Huber function, Lorentzian function and Tukey function are shown in Figure 2. We find that the half-quadratic function is close to the L2 norm when observation errors are small and it is close to the L1 norm when observation errors are big. This paper uses the half-quadratic function, which combines the advantages of the L1 norm and the L2 norm as the loss function. The half-quadratic function is strictly convex and twice continuously differentiable to obtain the optimum value easily.
The fidelity term of our robust MFSR by the half-quadratic function is:
X ^ = a r g m i n X { k = 1 K i = 1 I C · ( a k a k 2 + r k , i 2 a k 2 ) } ,  
where a k is the half-quadratic parameter for the kth LR image, r k , i represents the observation error of the ith pixel of the kth LR image.
Figure 3 shows the performance of the half-quadratic function with different a values. We can see from Figure 3 that the larger the a value is, the closer the half-quadratic function is to the L2 norm. The smaller the a value is, the closer it is to the L1 norm. Therefore, the a value is inversely proportional to observation errors.
The observation error of the kth LR image r k can be expressed as:
r k = i = 1 I | r k , i | ,  
where | · | means absolute value, I is the total number of pixels per LR image.
a k can be defined as:
a k = max ( r k ) r k ,
where max ( r k ) represents the maximum value of r k .
C is the confidence matrix composed of confidence weights, C = ( β 1 , , β k , , β K ) T . For each LR image, the observation error is larger and the corresponding confidence weight should be smaller. So, the confidence weight of the kth LR image β k can be expressed as:
β k = { mean ( r k ) r k                                               i f   | r k , i | c σ f t , mean ( r k ) r k   ·   c σ f t | r k , i |                 o t h e r w i s e ,                                            
where mean ( r k ) represents the mean value of r k . c is a positive constant, σ f t is the estimate value of the scale parameter in each iteration t to discriminate inliers and outliers adaptively. c is set to 2 in this paper. We use the median absolute deviation (MAD) [28] to estimate σ f t , the formula can be expressed as:
σ f t = σ 0 · M A D ( r t 1 | β t 1 ) ,
where we set σ 0 = 1.4826 for the Gaussian distribution.

3.3. Proposed Local Structure Tensor Weighted BTV Regularization Term

Due to the BTV regularization term, we cannot distinguish the edge information of the image well. We propose combining the local structure tensor information of the image with the image sparsity as image prior knowledge in order to simultaneously better preserve the edge information of the image and suppress the noise.
The proposed regularization term can be expressed as:
γ ( X ) = n = P P m = P P w T · ζ | m | + | n | X S x n S y m X 1 .  
The value of w T depends on the local structure tensor of the image. The local structure tensor can well describe the local structure information of the image, which can be expressed as:
T = [ I x 2 I x I y I x I y I y 2 ] ,
where I x , I y represents the gradient in horizontal and vertical directions, respectively.
Because the local structure tensor T is a positive semi-definite matrix and has two nonnegative eigenvalues, λ 1 and λ 2 ( λ 1 λ 2 ), the spatial structure information of the image can be divided into three cases: When λ 1 λ 2 0 , it means that the gray level change of the image along any direction is very small; it is a flat area. When λ 1 > λ 2 0 , it means that the gray change rate of the image along a certain direction at this point is larger; it is an edge region. When λ 1 λ 2 > 0 , it means that the gray level of the image changes greatly in both vertical directions; the point is a corner.
In order to describe the local structure information of the image well, we construct a local structure tensor weight matrix w T , which is expressed as:
w T = 1 w + 1 2 e ( λ 1 λ 2 ) 2 ,
where w is a fine tuning parameter; we set w = 0.5 .
As the image prior weight, ζ in Equation (20) related to X S x n S y m   X can be calculated adaptively as:
ζ = { 1                                               i f   | X S x n S y m X | c σ p t , c σ p t | X S x n S y m X |                 o t h e r w i s e ,                                            
where c is a positive constant, c is set to 2 in our paper. σ p t is obtained from the distribution | X S x n S y m X | t 1 and the image prior weight ζ t 1 , the formula similar to σ f t is as follows
σ p t = σ 0 · M A D ( | X S x n S y m X | t 1 | ζ t 1 ) ,
where σ 0 = 1 for the Laplacian distribution.

3.4. Image Reconstruction

In this paper, the formula of the reconstructed HR image can be written as
X ^ = a r g m i n X { k = 1 K i = 1 I C · ( a k a k 2 + r k , i 2 a k 2 ) + λ · n = P P m = P P w T · ζ | m | + | n | X S x n S y m X 1 } .  
The regularization parameter λ is used to balance the fidelity term and the regularization term. We use cross validation, which is used by Köhler et al. [9] to determine the value of λ .
There are many optimization algorithms for solving the minimization problem for MFSR. We employ Scaled Conjugate Gradient (SCG) [29] to solve the problem in the Equation (25) in this paper because the convergence of SCG is fast and the SCG algorithm can adjust the step size adaptively. We set the iteration threshold to 10 3 . In the process of optimization, the first-order derivative function of X can be calculated as:
f = k = 1 K i = 1 I C · a k W k T ( W k X Y k ) a k 2 + ( W k X Y k ) 2 + λ · n = P P m = P P w T · ζ | m | + | n | ( I S x n S y m ) s i g n ( X S x n S y m X ) ,
where S x n , S y m is the transpose matrix of S x n , S y m respectively. I is an identity matrix.

4. Experimental Results and Analysis

We used synthetic data and real data to test our proposed MFSR algorithm. We implemented our method on a laptop computer with Intel(R) Core(TM) i7-8650U CPU and 16 GB RAM, (Intel, Santa Clara, CA, USA). Our algorithm was compared to the following MFSR algorithms: L2 + Tikhonov [21], L2 + BTV [7], L1 + BTV [7], BEP [8], DeepSR [13], IRW [9], and SWHQ + ABTV [10].

4.1. Experiments on Synthetic Data

Our proposed method was first measured on synthetic data quantitatively, because the ground truth HR images were available. We used peak-signal-to-noise ratio (PSNR), structural similarity (SSIM) and Information fidelity criterion (IFC) to assess image quality in this paper. We used common HR images to generate synthetic LR images by using the Set 14 dataset [30], shown in Figure 4. We created 30 LR images from one HR image by adding random motion, blur, down sampling and noise. The range of random translations was from −3 to +3 pixels. The range of random rotation angles was from −1° to +1°. Each LR image was blurred by a Gaussian PSF with σ = 0.4 . We put in mixed noises by Gaussian noise and Poisson noise.
We first selected 25 appropriate LR images by using our method mentioned in Section 3.1. The results of PSNR, SSIM and IFC of these eight synthetic LR images are presented in Table 1, Table 2 and Table 3, respectively. Figure 5 shows the results of the visual comparison of several algorithms for the Cameraman image. The results of the visual comparison between our algorithm and the ground truth for some images are shown in Figure 6.
Table 1, Table 2 and Table 3 show that our proposed method in PSNR, SSIM and IFC is better than other algorithms. Our degradation model and the setting of the weight function are more reasonable. Figure 5 shows the results of the HR image estimated by these algorithms for the Cameraman image. Our algorithm performs better with image details. Figure 6 shows the comparison results of our estimated HR image and the ground truth. The result of our method is very close to the ground truth.

4.2. Experiments on Real Data

Except for synthetic data, we tested our proposed algorithm on real data which came from the Multi-Dimensional Signal Processing Research Group (MDSP) [31]. In practical applications, the image assessment matrices such as PSNR, SSIM and IFC cannot be used to evaluate the quality of the real images in the absence of ground truth images. We only tested our algorithm by visual comparison of reconstructed results for real data. Figure 7 and Figure 8 show the results of the visual comparison of EIA and Alpaca, respectively.
Figure 7 and Figure 8 show that our proposed algorithm still has better results when the magnification factor is relatively large because we employ the coarse to fine strategy which gradually enlarges to the target size. In addition, our estimated HR image, which has no obvious artifacts or noise, can maintain the details of the image and suppress the noise simultaneously, mainly because we propose the local structure tensor weighted BTV regularization term which is universal for natural images.
The experimental results on synthetic data and real data show that our proposed method has a better effect compared with current methods; especially for preserving edge information, our method has better performance.

5. Conclusions

MFSR is a very challenging problem. In this paper, we propose a new robust MFSR algorithm to preserve richer image detail information while suppressing noise. We first select the LR images with appropriate quality simply and quickly instead of all LR images as the input of MFSR, which reduces the computational cost of MFSR and makes it more practical. For the fidelity term, we analyze and use the half-quadratic function as the loss function. We adjust the function parameter and confidence weights adaptively according to the observation errors. We gradually reduce the observation errors of the LR images and the estimated HR image through iteration. We better suppress the noise and preserve the image edge details simultaneously using the local structure tensor weighted BTV regularization term. The combination of the local structure tensor and the sparsity of the image, as the prior knowledge of the image, can reflect the structural characteristics of the natural images.
Our proposed method is a non-blind reconstruction, which assumes that the point spread function is known. The blur kernel is often unknown in practical applications. Our next task is to study the unknown blur in MFSR. One direction of our future work is to extend our method to blind MFSR.

Author Contributions

S.L. is the first author of this paper. Her main contributions include the basic idea and writing of this paper. M.W. is the corresponding author of this paper. His main contributions include analyzing the basic idea. The main contributions of Q.H. and X.L. include checking the experimental results. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets available online: https://users.soe.ucsc.edu/~milanfar/software/sr-datasets.html (accessed on 4 March 2021).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, H.; Zhang, L.; Shen, H. A super-resolution reconstruction algorithm for hyperspectral images. Signal Process. 2012, 92, 2082–2096. [Google Scholar] [CrossRef]
  2. Köhler, T.; Brost, A.; Mogalle, K.; Zhang, Q.; Köhler, C.; Michelson, G.; Hornegger, J.; Tornow, R.P. Multi-frame super-resolution with quality self-assessment for retinal fundus videos. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Boston, MA, USA, 14–18 September 2014; pp. 650–657. [Google Scholar]
  3. Zhang, L.; Zhang, H.; Shen, H.; Li, P. A super-resolution reconstruction algorithm for surveillance images. Signal Process. 2010, 90, 848–859. [Google Scholar] [CrossRef]
  4. Tsaiand, R.Y.; Huang, T.S. Multi-frame image restoration and registration. Adv. Comput. Vis. Image Process. 1984, 1, 317–339. [Google Scholar]
  5. Takeda, H.; Farsiu, S.; Milanfar, P. Kernel regression for image processing and reconstruction. IEEE Trans. Image Process. 2007, 16, 349–366. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Baatz, M.; Eichenseer, A.; Kaup, A. Multi-image super resolution using a dual weighting scheme based on voronoi tessellation. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 2822–2826. [Google Scholar]
  7. Farsiu, S.; Robinson, M.D.; Elad, M.; Milanfar, P. Fast and robust multiframe super resolution. IEEE Trans. Image Process. 2004, 13, 1327–1344. [Google Scholar] [CrossRef]
  8. Zeng, X.; Yang, L. A robust multiframe super-resolution algorithm based on half-quadratic estimation with modified BTV regularization. Digit. Signal Process. 2013, 23, 98–109. [Google Scholar] [CrossRef]
  9. Köhler, T.; Huang, X.; Schebesch, F.; Aichert, A.; Maier, A.; Hornegger, J. Robust Multiframe Super-Resolution Employing Iteratively Re-Weighted Minimization. IEEE Trans. Comput. Imaging 2016, 2, 42–58. [Google Scholar] [CrossRef]
  10. Liu, X.; Chen, L.; Wang, W.; Zhao, J. Robust Multi-Frame Super-Resolution Based on Spatially Weighted Half-Quadratic Estimation and Adaptive BTV Regularization. IEEE Trans. Image Process. 2018, 27, 4971–4986. [Google Scholar] [CrossRef]
  11. Wang, P.; Hu, X.; Xuan, B.; Mu, J. Super resolution reconstruction via multiple frames joint learning. In Proceedings of the 2011 International Conference on Multimedia and Signal Processing (ICMSP), Seville, Spain, 18–21 July 2011; pp. 357–361. [Google Scholar]
  12. Kato, T.; Hino, H.; Murata, N. Multi-frame image super resolution based on sparse coding. Neural Netw. 2015, 66, 64–78. [Google Scholar] [CrossRef] [PubMed]
  13. Liao, R.; Tao, X.; Li, R.; Ma, Z.; Jia, J. Video super-resolution via deep draft-ensemble learning. In Proceedings of the IEEE International Conference on Computer Vision, Berlin, Germany, 11–14 March 2015; pp. 531–539. [Google Scholar]
  14. Noor, D.F.; Li, L.; Li, Z.; Bhattacharyya, S. Multi-frame super resolution with deep residual learning on flow registered non-integer pixel images. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2164–2168. [Google Scholar]
  15. Huang, Y.; Wang, W.; Wang, L. Bidirectional recurrent convolutional networks for multi-frame super-resolution. Adv. Neural Inf. Process. Syst. 2015, 28, 235–243. [Google Scholar]
  16. Huang, Y.; Wang, W.; Wang, L. Video super-resolution via bidirectional recurrent convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 1015–1028. [Google Scholar] [CrossRef] [PubMed]
  17. Patanavijit, V.; Jitapunkul, S. A robust iterative multiframe super resolution reconstruction using a Huber Bayesian approach with Huber Tikhonov regularization. In Proceedings of the IEEE International Symposium on Intelligent Signal Processing and Communications, Yonago, Japan, 12–15 December 2006; pp. 13–16. [Google Scholar]
  18. Patanavijit, V.; Jitapunkul, S. A Lorentzian stochastic estimation for a robust iterative multiframe super-resolution reconstruction with Lorentzian–Tikhonov regularization. J. Adv. Signal Process. 2007, 2, 1–21. [Google Scholar] [CrossRef] [Green Version]
  19. Anastassopoulos, A.; Vassilis, P. Regularized super-resolution image reconstruction employing robust error norms. Opt. Eng. 2009, 48, 1–14. [Google Scholar]
  20. Yue, L.; Shen, H.; Yuan, Q.; Zhang, L. A locally adaptive L1-L2 norm for multi-frame super-resolution of images with mixed noise and outliers. Signal Process. 2014, 105, 156–174. [Google Scholar] [CrossRef]
  21. Elad, M.; Feuer, A. Restoration of a single super resolution image from several blurred, noisy, and undersampled measured images. IEEE Trans. Image Process. 1997, 6, 1646–1658. [Google Scholar] [CrossRef] [Green Version]
  22. Pickup, L.C.; Capel, D.P.; Roberts, S.J.; Zisserma, A.N. Overcoming registration uncertainty in image super-resolution: Maximize or marginalize. EURASIP J. Adv. Signal Process. 2007, 92, 20. [Google Scholar] [CrossRef] [Green Version]
  23. Traonmilin, Y.; Ladjal, S.; Almansa, A. On the amount of regularization for superresolution interpolation. In Proceedings of the 2012 20th European Signal Processing Conference (EUSIPCO), Bucharest, Romania, 27–31 August 2012; pp. 380–384. [Google Scholar]
  24. Gu, K.; Zhai, G.; Lin, W.; Yang, X.; Zhang, W. No-reference image sharpness assessment in autoregressive parameter space. IEEE Trans. Image Process. 2015, 24, 3218–3231. [Google Scholar] [PubMed]
  25. Gu, K.; Li, L.; Lu, H.; Min, X.; Lin, W. A fast reliable image quality predictor by fusing micro-and macro-structures. IEEE Trans. Ind. Electron. 2017, 64, 3903–3912. [Google Scholar] [CrossRef]
  26. Rahimi, A.; Moallem, P.; Shahtalebi, K.; Momeni, M. Preserving quality in minimum frame selection within multi-frame super-resolution. Digit. Signal Process. 2018, 72, 19–43. [Google Scholar] [CrossRef]
  27. Charbonnier, L.; Blanc-Féraud, G.; Aubert, M. Barlaud Deterministic edge-preserving regularization in computed imaging. IEEE Trans. Image Process. 1997, 6, 298–311. [Google Scholar] [CrossRef] [PubMed]
  28. Scales, J.A.; Gersztenkorn, A. Robust methods in inverse theory. Inverse Probl. 1988, 4, 1071–1091. [Google Scholar] [CrossRef]
  29. Nabney, I.T. NETLAB: Algorithms for Pattern Recognition, 1st ed.; Springer: New York, NY, USA, 2002. [Google Scholar]
  30. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces; Springer: Berlin/Heidelberg, Germany, 2010; pp. 711–730. [Google Scholar]
  31. Farsiu, S. MDSP Super-Resolution and Demosaicing Datasets. 2017. Available online: https://users.soe.ucsc.edu/~milanfar/software/sr-datasets.html (accessed on 4 March 2021).
Figure 1. Sub-pixel displacement.
Figure 1. Sub-pixel displacement.
Sensors 21 05533 g001
Figure 2. L1 norm, L2 norm, half-quadratic function, Huber function, Lorentzian function, Tukey function.
Figure 2. L1 norm, L2 norm, half-quadratic function, Huber function, Lorentzian function, Tukey function.
Sensors 21 05533 g002
Figure 3. Half-quadratic function with different a values.
Figure 3. Half-quadratic function with different a values.
Sensors 21 05533 g003
Figure 4. 8 HR images for synthetic LR images. (a) Bridge. (b) Cameraman. (c) Coastguard. (d) Baboon. (e) Comic. (f) Barbara. (g) Pepper. (h) Zebra.
Figure 4. 8 HR images for synthetic LR images. (a) Bridge. (b) Cameraman. (c) Coastguard. (d) Baboon. (e) Comic. (f) Barbara. (g) Pepper. (h) Zebra.
Sensors 21 05533 g004
Figure 5. Visual comparison of MFSR algorithms for Cameraman image with mixed noises (   r = 3 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed. (j) ground truth.
Figure 5. Visual comparison of MFSR algorithms for Cameraman image with mixed noises (   r = 3 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed. (j) ground truth.
Sensors 21 05533 g005aSensors 21 05533 g005b
Figure 6. Visual comparison of MFSR algorithms with mixed noises (   r = 3 ). (a1) the LR image for Zebra image. (a2) our method image for Zebra image. (a3) the ground truth image for Zebra image. (b1) the LR image for Comic image. (b2) our method image for Comic image. (b3) the ground truth image for Comic image. (c1) the LR image for Pepper image. (c2) our method image for Pepper image. (c3) the ground truth image for Pepper image. (d1) the LR image for Coastguard image. (d2) our method image for Coastguard image. (d3) the ground truth image for Coastguard image.
Figure 6. Visual comparison of MFSR algorithms with mixed noises (   r = 3 ). (a1) the LR image for Zebra image. (a2) our method image for Zebra image. (a3) the ground truth image for Zebra image. (b1) the LR image for Comic image. (b2) our method image for Comic image. (b3) the ground truth image for Comic image. (c1) the LR image for Pepper image. (c2) our method image for Pepper image. (c3) the ground truth image for Pepper image. (d1) the LR image for Coastguard image. (d2) our method image for Coastguard image. (d3) the ground truth image for Coastguard image.
Sensors 21 05533 g006aSensors 21 05533 g006b
Figure 7. Visual comparison of MFSR algorithms for EIA frames (   r = 5 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed.
Figure 7. Visual comparison of MFSR algorithms for EIA frames (   r = 5 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed.
Sensors 21 05533 g007aSensors 21 05533 g007b
Figure 8. Visual comparison of MFSR algorithms for Alpaca frames ( r = 5 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed.
Figure 8. Visual comparison of MFSR algorithms for Alpaca frames ( r = 5 ). (a) the LR image. (b) L2 + Tikhonov. (c) L2 + BTV. (d) L1 + BTV. (e) BEP. (f) DeepSR. (g) IRW. (h) SWHQ + ABTV. (i) Proposed.
Sensors 21 05533 g008aSensors 21 05533 g008b
Table 1. PSNR Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
Table 1. PSNR Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
ImagesL2 + TikhonovL2 + BTVL1 + BTVBEPDeepSRIRWSWHQ + ABTVOur Proposed
Baboon26.751927.653927.851528.617929.042832.014532.081732.8181
Barbara26.479527.677628.065928.452828.512728.955829.097829.2285
Bridge27.376727.386427.666528.525728.350328.615829.055529.3100
Cameraman25.931228.444728.183130.318030.574931.230031.565931.7468
Coastguard30.474233.370834.035234.149534.113534.302934.622035.0370
Comic29.012329.203331.076131.222231.129832.030932.336232.4921
Pepper31.266834.067834.374036.126636.274137.151737.282437.3033
Zebra30.556234.423934.082435.043734.782135.185735.406835.5603
average28.481130.278630.666831.557031.597532.435932.681032.9370
Table 2. SSIM Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
Table 2. SSIM Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
ImagesL2 + TikhonovL2 + BTVL1 + BTVBEPDeepSRIRWSWHQ + ABTVOur Proposed
Baboon0.80480.80840.82580.84820.85020.90890.91890.9237
Barbara0.76240.79850.80730.84690.84750.85590.86030.8712
Bridge0.79470.79940.83510.84950.83860.85510.86260.8795
Cameraman0.66610.69610.76420.90120.90230.90490.90880.9150
Coastguard0.81440.90750.90810.91370.91060.92770.92870.9316
Comic0.86790.91780.92130.94580.93980.95560.95570.9562
Pepper0.71680.84080.82360.91460.91380.91420.91520.9158
Zebra0.84010.92410.92460.93990.93220.94260.95190.9532
average0.78340.83660.85120.89500.89190.90810.91280.9183
Table 3. IFC Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
Table 3. IFC Results of our MFSR Algorithms on the 8 synthetic LR images under r = 3 .
ImagesL2 + TikhonovL2 + BTVL1 + BTVBEPDeepSRIRWSWHQ + ABTVOur Proposed
Baboon7.35347.95127.95798.09448.45638.75508.76308.8393
Barbara7.01357.34127.43828.04638.37588.61108.72948.7274
Bridge5.94176.34126.78887.16226.98757.21757.24467.3342
Cameraman4.16734.32694.35044.95574.98765.09985.12775.1951
Coastguard6.07166.25986.22806.35656.17896.72856.85236.8968
Comic7.84568.58848.58089.04078.91329.55699.60009.6259
Pepper4.51224.81354.73965.74845.74785.74845.79265.8216
Zebra8.08528.31438.50888.86998.78919.00159.06629.1005
average6.37386.74216.82417.28437.30457.58987.64707.6926
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, S.; Wang, M.; Huang, Q.; Liu, X. Robust Multi-Frame Super-Resolution Based on Adaptive Half-Quadratic Function and Local Structure Tensor Weighted BTV. Sensors 2021, 21, 5533. https://doi.org/10.3390/s21165533

AMA Style

Liu S, Wang M, Huang Q, Liu X. Robust Multi-Frame Super-Resolution Based on Adaptive Half-Quadratic Function and Local Structure Tensor Weighted BTV. Sensors. 2021; 21(16):5533. https://doi.org/10.3390/s21165533

Chicago/Turabian Style

Liu, Shanshan, Minghui Wang, Qingbin Huang, and Xia Liu. 2021. "Robust Multi-Frame Super-Resolution Based on Adaptive Half-Quadratic Function and Local Structure Tensor Weighted BTV" Sensors 21, no. 16: 5533. https://doi.org/10.3390/s21165533

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop