Next Article in Journal
Effective Incomplete Multi-View Clustering via Low-Rank Graph Tensor Completion
Previous Article in Journal
Public Debt, Governance, and Growth in Developing Countries: An Application of Quantile via Moments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Successive Convex Approximation for Image Super-Resolution

1
School of Electronics and Information Engineering, Liaoning University of Technology, Jinzhou 121001, China
2
SolBridge International School of Business, Woosong University, Daejeon 34613, Republic of Korea
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(3), 651; https://doi.org/10.3390/math11030651
Submission received: 31 December 2022 / Revised: 19 January 2023 / Accepted: 21 January 2023 / Published: 27 January 2023

Abstract

:
Image super-resolution (SR), as one of the classic image processing issues, has attracted increasing attention from researchers. As a highly ill-conditioned, non-convex optimization issue, it is difficult for image SR to restore a high-resolution (HR) image from a given low-resolution (LR) instance. Recent researchers have tended to regard image SR as a regression task and to design an end-to-end convolutional neural network (CNN) to predict the pixels directly, which lacks inherent theoretical analysis and limits the effectiveness of the restoration. In this paper, we analyze image SR from an optimization perspective and develop a deep successive convex approximation network (SCANet) for generating HR images. Specifically, we divide non-convex optimization into several convex LASSO sub-problems and use CNN to adaptively learn the parameters. To boost network representation, we use residual feature aggregation (RFA) blocks and devise a spatial and channel attention (SACA) mechanism to improve the restoration capacity. The experimental results show that the proposed SCANet can restore HR images more effectively than other works. Specifically, SCANet achieves higher PSNR/SSIM results and generates more satisfying textures.

1. Introduction

Image super-resolution (SR) is one of the classic image processing issues [1]. Given a low-resolution (LR) image, the task of image SR is to generate a corresponding high-resolution (HR) image with higher subjective and objective performance [2]. Image SR has been widely considered in various applications, such as iris and ocular recognition [3] and other high-level computer vision tasks [4].
With the rapid development of deep learning, there are numerous convolutional neural networks (CNNs) that have been specifically designed to solve the image SR problem [5,6,7,8]. SRCNN [5] is the first CNN-based image SR method to generate satisfying results from the given instances. After SRCNN, deeper and wider networks have achieved good performance with elaborate block designs [7,8]. However, these works only regard image SR as a regression issue and use an end-to-end network to directly predict the pixels of HR images. This direct regression limits the precision of the restoration methods.
As a highly ill-conditioned problem, image SR can be regarded as an optimization issue [9,10,11,12,13]. From an optimization perspective, image SR aims to find a satisfying result which is in accordance with the LR representation after down-sampling and follows the prior HR image. As such, image SR can be regarded as an unconstrained optimization problem under the consideration of the fidelity term and the prior term. There are also works regarding image SR as an optimization issue [9,10,11,12,13]. Half-quadratic splitting (HQS) and the alternating direction method of multipliers (ADMM) are two representative strategies to iteratively find a feasible solution for non-convex tasks. Recently, works have considered HQS and ADMM for image SR, which have achieved good performances [9,10]. However, both HQS- and ADMM-based CNNs solve the non-convex problem without a specific prior HR image, which makes it hard to find a satisfying solution and limits the effectiveness of the network.
In this paper, we analyze image SR from a non-convex optimization perspective and solve the issue in the successive convex approximation (SCA) manner. Specifically, we divide the non-convex problem into several convex LASSO sub-problems and calculate the solution successively. From this point of view, the prior HR image is explicitly considered as the 1 -norm, which follows the sparsity of the natural images. We design a network, dubbed SCANet, to adaptively learn the parameters and conduct the optimization. To improve the network capacity, we use residual feature aggregation (RFA) blocks [14] and devise a spatial and channel attention (SACA) mechanism in SCANet. The experimental results show that the proposed SCANet achieves better performance than other image SR works, with higher PSNR/SSIM results. Furthermore, SCANet can generate more satisfying textures and details with better visual performances than other works.
Our contributions can be concluded as follows:
  • We analyze image SR from a non-convex optimization perspective and propose an iterative solution by dividing the task into several convex LASSO sub-problems, which is more effective than other iteration solutions. Based on this analysis, we devise a SCANet for image super-resolution.
  • We utilize residual feature aggregation (RFA) blocks and develop a spatial and channel attention (SACA) mechanism in SCANet, which can improve network representation and restoration capacity.
  • The experimental results show that the proposed SCANet achieves better subjective and objective performance than other works, with higher PSNR/SSIM results.

2. Related Works

Recently, convolutional neural networks (CNNs) have found great success in image SR [15]. SRCNN [5] is the first CNN-based image SR method that utilizes three-layer architecture to generate HR details. After SRCNN, numerous well-designed networks have been proposed for effective image SR. FSRCNN [6] modified the SRCNN for faster speed and higher restoration precision with a thinner but deeper design. VDSR [7] used a very deep network design with a residual connection to achieve good restoration performance. EDSR [8] developed a deep residual network with residual blocks [16] for better information transmission and network representation. DRRN [17] and DRCN [18] considered the recursive design for image SR and utilized an iterative design for restoration. Recently, researchers have mainly focused on elaborate block designs for image SR. CARN [19] built a cascading network for efficient image SR. OISR [20] designed the block according to the ODE pattern and achieved good performance. RDN [21] combined the residual connection [16] and the dense connection [22] and developed a residual dense block for image SR. SRDenseNet [23] built a deep network with dense connections for better gradient transmission. RCAN [24] utilized a residual-in-residual design for effective gradient transmission and made the network deeper. SAN [25] developed a second-order attention mechanism for image SR and achieved state-of-the-art performance. However, these works only consider building a deeper and wider network with well-designed blocks for restoration, and neglect to consider the image SR problem from a mathematical perspective.
By considering the image SR issue from a mathematical perspective, it can be regarded as a highly ill-conditioned issue that can be solved using optimization methodologies. IRCNN [9] utilized a half-quadratic splitting (HQS) strategy to first separate the image SR into a closed-form solution and a denoiser. DPSR [10] considered a novel image SR degradation situation, and proposed a deep plug-and-play model to solve the problem. However, these works only aimed to solve the non-convex problem using the iterative mechanism, which lacks the explicit prior definition of HR images and makes it difficult to find a satisfying solution.

3. Methodology

In this section, we introduce the proposed SCANet in the following manner. First, we analyze image SR as an optimization issue and propose a solution with the help of successive convex approximation (SCA). Then, an end-to-end network is designed to find a solution that adaptively learns the parameters from the data pairs. Finally, we discuss the details of the proposed SCANet.

3.1. Successive Convex Approximation for Image Super-Resolution

Given a low-resolution (LR) image I L R , the task of image SR is to find a feasible high-resolution (HR) image I H R , satisfying
I H R = arg min I H R 1 2 | | I L R D ( I H R ) | | 2 + λ Φ ( I H R ) ,
where D ( · ) is the degradation operation and Φ ( · ) is the prior image of the HR images. λ is a scaling factor.
In general, Equation (1) is a non-convex issue, for which it is difficult to find a satisfying solution. To find a feasible result, we utilize the successive convex approximation (SCA) strategy to convert this non-convex problem into several convex tasks and find the solution in an iterative manner.
For a k-th iteration, we aim to solve the task
I k H R = arg min I H R 1 2 | | I L R H I k 1 H R | | 2 2 + λ k 1 | | I k 1 H R | | 1 ,
where H is the matrix representation of the degradation operation D ( · ) . In Equation (2), we choose 1 norm as the prior image, which considers the sparsity of the natural image. It can be seen that Equation (2) embodies the same situation as the LASSO problem. To solve this equation, we use the sub-gradient descent method and find a feasible descent direction for optimization. Then, the descent direction of the k-th iteration can be shown as
d k = ( H T H I k 1 H R H I L R + λ k 1 sign ( I k 1 H R ) ) .
By using this sub-gradient descent direction, the optimization of the k-th step is described as
I k H R = I k 1 H R + γ d k ,
where γ is the step length and d k is the descent direction of the k-th step. The descent direction is a theoretical guidance for the optimization step that decreases the distance between I k H R and the ideal I H R . The optimization steps follow the original steps of successive convex optimization, which have been investigated in previous optimization studies [26,27,28].

3.2. Deep Successive Convex Approximation Network

Based on mathematical analysis of image SR, we develop the deep successive convex approximation network (SCANet) for image SR. The design of SCANet is shown in Figure 1. In this figure, x l r is the feature of the LR image, x i is the feature processed using the i-th LASSO block, and x h r is the restored feature of the HR image. We conduct SCA in the feature space by designing special LASSO blocks. First, one convolution processes the input LR image and generates the corresponding feature maps as
x l r = C o n v ( I L R ) .
After that, K LASSO blocks are used to conduct the optimization. For the k-th step, this is
x k = L A S S O ( x l r , x k 1 ) .
The LASSO blocks follow the operation of Equations (3) and (4).
After K LASSO blocks, a padding structure is used to restore the HR image feature with residual connections as
x h r = P a d d i n g ( x K ) + x l r ,
where P a d d i n g ( · ) is composed of two convolutional layers and a ReLU activation.
Finally, an upscale module restores the HR image from the feature maps as
I H R = U p s c a l e ( x l r ) .
The upscale module is composed of one convolutional layer and a sub-pixel convolution [29].
Specifically, we describe the design of the LASSO block, which is the core of SCANet. Figure 2 shows the block design of the LASSO block. The block follows the analysis of Equations (3) and (4). In this figure, we can see that the H T H in Equation (3) is simulated using four residual feature aggregation blocks (RFAB) and one convolution. The H is simulated using two convolutional layers and a ReLU activation. Then, the learned descent direction of the k-th iteration step is calculated as
d k = C o n v 1 × 1 ( [ H T H x k H x l r , x k ] ) ,
where [ · ] is the concatenate operation. The 1 × 1 convolution is used to adaptively learn the scaling factor λ k 1 . Finally, one 1 × 1 convolution is used to conduct Equation (4) by adaptively learning the step length as
x k + 1 = C o n v 1 × 1 ( [ d k , x k ] ) .
Figure 3 shows the design of RFAB in the LASSO block. RFAB has been proven to be a good component for image SR in recent works [14,30] and has been widely considered in previous SR works. According to previous reports [14,30], RFAB can more effectively explore the image features than other well-designed blocks.
Different from residual blocks, RFAB aggregates the residual information from different blocks to jointly learn multi-scale information and restores the image information. Compared with traditional residual blocks, RFAB has an adaptive ability to describe the residual features. There are three processing modules with the same design in RFAB, which are composed of two convolutional blocks, one ReLU activation, and a spatial and channel attention (SACA) mechanism. Let P r o c R F A B ( · ) be the processing module and x i n be the input of RFAB; then, the operation of the processing modules can be described as
r 1 = P r o c R F A B ( x i n ) , x i n 1 = x i n + r 1 , r 2 = P r o c R F A B ( x i n 1 ) , x i n 2 = x i n + r 2 , r 3 = P r o c R F A B ( x i n 2 ) .
Then, the aggregated residual information is calculated via one 1 × 1 convolution as
r a l l = C o n v 1 × 1 ( [ r 1 , r 2 , r 3 ] ) .
The output of RFAB x o u t is calculated as
x o u t = x i n + r a l l .
SACA in RFAB is shown in Figure 4. SACA focuses on spatial attention and channel attention separately, and utilizes one convolution to combine dual attention information. To this end, the two types of attention can focus on different dimensions of the image and improve the restoration capacity. We calculate the spatial attention and the channel attention via two parallel pathways. In each path, there are two convolutional layers, one ReLU activation and one sigmoid activation, which follow the design of various attention mechanisms [24,25]. There is one global average pool in the channel attention mechanism. After the spatial and channel attentions, one 1 × 1 convolution combines the two dual information sources, and a residual connection is introduced for efficient gradient transmission.

3.3. Implementation Details

There are K = 4 LASSO blocks in SCANet. All convolutional layers have c = 64 channels, and the kernel sizes are set as 3 × 3 , except for the 1 × 1 convolutions. The filter in the upscale module is set as c u p = 3 × s × s , where s is the scaling factor of the image SR.

4. Experiment

4.1. Settings

We train our network with the DIV2K [31] dataset. The DIV2K dataset contains 900 high-resolution images for training and validation, with a resolution up to 2K. In this work, we first choose 800 images for training, and the last 5 images for validation, which are the same settings used in recent works [21,24]. The testing datasets chosen are Set5 [32], Set14 [33], B100 [34], Urban100 [35], and Manga109 [36], which are commonly used in different SR works. The scaling factors of image SR chosen are × 2 , × 3 and × 4 . The learning rate chosen is l r = 10 4 . The patch size chosen is 48 × 48 for LR images, and the batch size is the same as the RDN [21]. We update our network using the Adam optimizer [37] for 1000 epochs, and halve the learning rate for every 200 epochs. The metrics of image SR are chosen as PSNR and SSIM.

4.2. Ablation Study

In SCANet, we develop RFAB and SACA for effective image SR. To investigate the effectiveness of these blocks, we compare the PSNR and the SSIM results against three different benchmarks: B100 [34], Urban100 [35], and Manga109 [36]. Table 1 shows the results of the different designs.
In this table, we can see that RFAB brings a large improvement to PSNR and SSIM. By comparing the first and second lines, RFAB brings a 0.4 dB improvement to the Urban100 and Manga109 benchmarks. The model also benefits from SACA. In the second and third lines, SACA results in 0.05 db PSNR and 0.02 SSIM for the Urban100 and Manga109 benchmarks. From this point of view, RFAB and SACA are effective components for the proposed SCANet.

4.3. Results

We compare our SCANet with several CNN-based methods: SRCNN [5], FSRCNN [6], VDSR [7], DRCN [18], CNF [38], LapSRN [39], DRRN [17], BTSRN [40], MemNet [41], SelNet [42], CARN [19], RAN [43], DNCL [44], FilterNet [45], MRFN [46], and DEGREE [47]. The scaling factors chosen are × 2 , × 3 , and × 4 . The testing benchmarks chosen are Set5, Set14, B100, Urban100, and Manga109. The metrics chosen are PSNR and SSIM.
Table 2 shows the PSNR/SSIM comparisons against five testing benchmarks with three different scaling factors. In the table, we can see that our SCANet achieves competitive or better performances than those in other works. SCANet achieves a 0.2 dB improvement over Urban100 and a 0.3 dB improvement over Manga109, with a scaling factor of × 2 . Similarly, there is a 0.2 dB PSNR improvement over Set5 compared with CARN when the scaling factor is × 3 . When the scaling factor is × 4 , SCANet achieves the highest PSNR/SSIM results when compared with other works. From this point of view, SCANet is effective for image SR.
Furthermore, we compare the visualization quality with VDSR [7]. Figure 5 shows the restored images of the Urban100 benchmark with a scaling factor ×4. In the figure, we can see that the proposed SCANet can restore the circles and lines more clearly than other works. Furthermore, the results of SCANet are the closest to the HR images. As such, the PSNR results of SCANet are better than other works. From this point of view, SCANet can effectively restore LR images with more satisfying results.
We also investigate the subjective qualities of different methods. To this end, we use the no-reference indicators and measure the visual quality of different methods. The indicators are chosen as NIQE and BRISQUE. Table 3 shows the subjective comparisons. It should be noted that the lower NIQE and BRISQUE values mean a better visual performance. In the figure, it can be seen that SCANet achieves the best performance with the lowest indicator scores. From this point of view, SCANet obtains better visual quality than other works.
To demonstrate the significance of our method, we conduct the two-sample t-test on the NIQE results of the three methods. Table 4 shows the results. The figure “0” in the table means there is no significant difference between the two methods; “1” means the upper method outperforms the left method; and “−1 ” means the left method outperforms the upper one. We can see from the table that SCANet performs significantly better than other works, with good visual quality.

4.4. Discussion

In this work, we compare our method with others, with a scaling factor of × 2 , × 3 , and × 4 . There are works with much larger scaling factors, such as 16, 32, or even 64 [50,51]. However, these methods only focus on some specific kinds of images, such as facial SR [50]. These methods only work in the specific situation, but cannot be generalized to different applications. Our work is a general SR method that can be directly utilized in various scenarios. The most commonly compared scaling factors are 2, 3, and 4 for general SR works, such as SRCNN [5], LapSRN [39], and CARN [19].
There are also cases of failure for SCANet. SCANet regards image super-resolution as an optimization task, which does not specifically focus on edge restoration. From this point of view, works aiming at structural restoration, such as SeaNet [48], achieve higher SSIM scores on the Urban100 [35] and Manga109 [36] datasets, which contain plentiful edges and lines. In further work, we will consider special edges and structural priors in the optimization step and thus improve restoration performance.
Different from vanilla optimization methods, the deep learning-based optimization method has a fixed number of iteration steps. Convergence is guaranteed by the convex optimization algorithm, and the network adaptively learns the parameters during the training step. Successive convex approximation has proved to be convergent in the previous literature [26,27,28]. The LASSO sub-problem also has a converged solution [52]. From this point of view, SCANet can be converged during iteration steps.
In general, this network provides an effective restoration method for image SR and achieves a better performance than other related works. Furthermore, this network provides a new perspective on combining deep learning and the optimization problem, which can be further investigated for different computer vision tasks.

5. Conclusions

Currently, researchers usually regard image SR as a regression task and use an end-to-end convolutional neural network (CNN) to predict the pixels directly, which lacks inherent theoretical analysis and limits the effectiveness of restoration. In this paper, we propose a deep successive convex approximation network for image super-resolution (SR), dubbed SCANet. In SCANet, the SR task was investigated from an optimization perspective, which was converted into LASSO sub-problems and solved using network blocks. The network blocks were established with residual feature aggregation blocks (RFAB) and the spatial and channel attention (SACA) mechanism for better performance. The experimental results show that the proposed SCANet achieved a better objective performance than other works and restored more satisfying results with higher subjective quality.

Author Contributions

Methodology, X.L. (Xiaohui Li); writing, J.W. and X.L. (Xinbo Liu); data curation, X.L. (Xiaohui Li). All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Liaoning Applied Basic Research Program (2022JH2/101300278).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available in DIV2K at https://data.vision.ee.ethz.ch/cvl/DIV2K/, reference number [31].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, Y.W.; Kim, J.S.; Park, K.R. Ocular Biometrics with Low-Resolution Images Based on Ocular Super-Resolution CycleGAN. Mathematics 2022, 10, 3818. [Google Scholar] [CrossRef]
  2. Batchuluun, G.; Nam, S.H.; Park, C.; Park, K.R. Super-Resolution Reconstruction-Based Plant Image Classification Using Thermal and Visible-Light Images. Mathematics 2023, 11, 76. [Google Scholar] [CrossRef]
  3. Lee, Y.W.; Park, K.R. Recent Iris and Ocular Recognition Methods in High-and Low-Resolution Images: A Survey. Mathematics 2022, 10, 2063. [Google Scholar] [CrossRef]
  4. Lee, S.J.; Yoo, S.B. Super-resolved recognition of license plate characters. Mathematics 2021, 9, 2494. [Google Scholar] [CrossRef]
  5. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the European Conference on Computer Vision, Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar]
  6. Dong, C.; Loy, C.C.; Tang, X. Accelerating the Super-Resolution Convolutional Neural Network. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 391–407. [Google Scholar]
  7. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  8. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  9. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning Deep CNN Denoiser Prior for Image Restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2808–2817. [Google Scholar]
  10. Zhang, K.; Zuo, W.; Zhang, L. Deep Plug-And-Play Super-Resolution for Arbitrary Blur Kernels. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1671–1681. [Google Scholar]
  11. Wei, Y.; Li, Y.; Ding, Z.; Wang, Y.; Zeng, T.; Long, T. SAR Parametric Super-Resolution Image Reconstruction Methods Based on ADMM and Deep Neural Network. IEEE Trans. Geosci. Remot. Sens. 2021, 59, 10197–10212. [Google Scholar] [CrossRef]
  12. Ma, Q.; Jiang, J.; Liu, X.; Ma, J. Deep Unfolding Network for Spatiospectral Image Super-Resolution. IEEE Trans. Comput. Imaging 2022, 8, 28–40. [Google Scholar] [CrossRef]
  13. Ran, Y.; Dai, W. Fast and Robust ADMM for Blind Super-Resolution. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 6–12 June 2021; pp. 5150–5154. [Google Scholar] [CrossRef]
  14. Liu, J.; Zhang, W.; Tang, Y.; Tang, J.; Wu, G. Residual Feature Aggregation Network for Image Super-Resolution. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020, Seattle, WA, USA, 13–19 June 2020; pp. 2356–2365. [Google Scholar]
  15. Anwar, S.; Khan, S.H.; Barnes, N. A Deep Journey into Super-resolution: A Survey. ACM Comput. Surv. 2021, 53, 60:1–60:34. [Google Scholar] [CrossRef]
  16. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  17. Tai, Y.; Yang, J.; Liu, X. Image Super-Resolution via Deep Recursive Residual Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2790–2798. [Google Scholar] [CrossRef]
  18. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-Recursive Convolutional Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  19. Ahn, N.; Kang, B.; Sohn, K. Fast, Accurate, and Lightweight Super-Resolution with Cascading Residual Network. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; Volume 11214, pp. 256–272. [Google Scholar]
  20. He, X.; Mo, Z.; Wang, P.; Liu, Y.; Yang, M.; Cheng, J. ODE-Inspired Network Design for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 1732–1741. [Google Scholar]
  21. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 43, 2480–2495. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar] [CrossRef] [Green Version]
  23. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image Super-Resolution Using Dense Skip Connections. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4809–4817. [Google Scholar]
  24. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; Volume 11211, pp. 294–310. [Google Scholar]
  25. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.; Zhang, L. Second-Order Attention Network for Single Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11057–11066. [Google Scholar]
  26. Fang, B.; Qian, Z.; Zhong, W.; Shao, W. Iterative precoding for MIMO wiretap channels using successive convex approximation. In Proceedings of the 2015 IEEE 4th Asia-Pacific Conference on Antennas and Propagation (APCAP), Bali Island, Indonesia, 30 June–3 July 2015; pp. 65–66. [Google Scholar] [CrossRef]
  27. Pang, C.; Au, O.C.; Zou, F.; Zhang, X.; Hu, W.; Wan, P. Optimal dependent bit allocation for AVS intra-frame coding via successive convex approximation. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 1520–1523. [Google Scholar] [CrossRef]
  28. Yang, Y.; Pesavento, M.; Chatzinotas, S.; Ottersten, B. Successive Convex Approximation Algorithms for Sparse Signal Estimation With Nonconvex Regularizations. IEEE J. Sel. Top. Signal Process. 2018, 12, 1286–1302. [Google Scholar] [CrossRef] [Green Version]
  29. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-Time Single Image and Video Super-Resolution Using an Efficient Sub-Pixel Convolutional Neural Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar] [CrossRef] [Green Version]
  30. Liu, J.; Tang, J.; Wu, G. Residual Feature Distillation Network for Lightweight Image Super-Resolution. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 41–55. [Google Scholar] [CrossRef]
  31. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1122–1131. [Google Scholar]
  32. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi Morel, M.L. Low-Complexity Single-Image Super-Resolution based on Nonnegative Neighbor Embedding. In Proceedings of the British Machine Vision Conference (BMVC), Surrey, UK, 3–7 September 2012; pp. 135.1–135.10. [Google Scholar]
  33. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; pp. 711–730. [Google Scholar]
  34. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  35. Huang, J.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  36. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based manga retrieval using manga109 dataset. Multimed. Tools Appl. (MTA) 2017, 76, 21811–21838. [Google Scholar] [CrossRef] [Green Version]
  37. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  38. Ren, H.; El-Khamy, M.; Lee, J. Image Super Resolution Based on Fusing Multiple Convolution Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1050–1057. [Google Scholar]
  39. Lai, W.; Huang, J.; Ahuja, N.; Yang, M. Deep Laplacian Pyramid Networks for Fast and Accurate Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5835–5843. [Google Scholar]
  40. Fan, Y.; Shi, H.; Yu, J.; Liu, D.; Han, W.; Yu, H.; Wang, Z.; Wang, X.; Huang, T.S. Balanced Two-Stage Residual Networks for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1157–1164. [Google Scholar]
  41. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A Persistent Memory Network for Image Restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  42. Choi, J.; Kim, M. A Deep Convolutional Neural Network with Selection Units for Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1150–1156. [Google Scholar] [CrossRef]
  43. Wang, Y.; Wang, L.; Wang, H.; Li, P. Resolution-Aware Network for Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 1259–1269. [Google Scholar] [CrossRef]
  44. Xie, C.; Zeng, W.; Lu, X. Fast Single-Image Super-Resolution via Deep Network With Component Learning. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 3473–3486. [Google Scholar] [CrossRef]
  45. Li, F.; Bai, H.; Zhao, Y. FilterNet: Adaptive Information Filtering Network for Accurate and Fast Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 1511–1523. [Google Scholar] [CrossRef]
  46. He, Z.; Cao, Y.; Du, L.; Xu, B.; Yang, J.; Cao, Y.; Tang, S.; Zhuang, Y. MRFN: Multi-Receptive-Field Network for Fast and Accurate Single Image Super-Resolution. IEEE Trans. Multimed. 2020, 22, 1042–1054. [Google Scholar] [CrossRef]
  47. Yang, W.; Feng, J.; Yang, J.; Zhao, F.; Liu, J.; Guo, Z.; Yan, S. Deep Edge Guided Recurrent Residual Learning for Image Super-Resolution. IEEE Trans. Image Process. 2017, 26, 5895–5907. [Google Scholar] [CrossRef] [Green Version]
  48. Fang, F.; Li, J.; Zeng, T. Soft-Edge Assisted Network for Single Image Super-Resolution. IEEE Trans. Image Process. 2020, 29, 4656–4668. [Google Scholar] [CrossRef]
  49. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight Image Super-Resolution with Information Multi-Distillation Network. In Proceedings of the ACM International Conference on Multimedia (MM), Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  50. He, J.; Shi, W.; Chen, K.; Fu, L.; Dong, C. GCFSR: A Generative and Controllable Face Super Resolution Method Without Facial and GAN Priors. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 1879–1888. [Google Scholar] [CrossRef]
  51. Menon, S.; Damian, A.; Hu, S.; Ravi, N.; Rudin, C. PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 2434–2442. [Google Scholar] [CrossRef]
  52. Kim, Y.; Hao, J.; Mallavarapu, T.; Park, J.; Kang, M. Hi-LASSO: High-Dimensional LASSO. IEEE Access 2019, 7, 44562–44573. [Google Scholar] [CrossRef]
Figure 1. Network design of SCANet. x l r is the feature of the LR image, x i is the feature processed using i-th LASSO blocks, and x h r is the restored feature of the HR image.
Figure 1. Network design of SCANet. x l r is the feature of the LR image, x i is the feature processed using i-th LASSO blocks, and x h r is the restored feature of the HR image.
Mathematics 11 00651 g001
Figure 2. Block design of the LASSO block.
Figure 2. Block design of the LASSO block.
Mathematics 11 00651 g002
Figure 3. Block design of RFAB.
Figure 3. Block design of RFAB.
Mathematics 11 00651 g003
Figure 4. Block design of SACA.
Figure 4. Block design of SACA.
Mathematics 11 00651 g004
Figure 5. Visual comparisons against the Urban100 benchmark.
Figure 5. Visual comparisons against the Urban100 benchmark.
Mathematics 11 00651 g005
Table 1. Investigation of RFAB and SACA with scaling factor × 4 .
Table 1. Investigation of RFAB and SACA with scaling factor × 4 .
RFABSACAB100 [34]Urban100 [35]Manga109 [36]
27.50/0.733825.76/0.776330.05/0.9027
27.58/0.736426.12/0.787630.45/0.9079
27.58/0.736526.17/0.789330.50/0.9083
Table 2. Average PSNR/SSIM with degradation model BI  × 2 , × 3 , and × 4 against five benchmarks. (+) means the result is better than our method, and (-) means the result is worse than our method.
Table 2. Average PSNR/SSIM with degradation model BI  × 2 , × 3 , and × 4 against five benchmarks. (+) means the result is better than our method, and (-) means the result is worse than our method.
ScaleModelSet5 [32]Set14 [33]B100 [34]Urban100 [35]Manga109 [36]
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
× 2 SRCNN [5]36.66(-)/
0.9542(-)
32.42(-)/
0.9063(-)
31.36(-)/
0.8879(-)
29.50(-)/
0.8946(-)
35.74(-)/
0.9661(-)
FSRCNN [6]37.00(-)/
0.9558(-)
32.63(-)/
0.9088(-)
31.53(-)/
0.8920(-)
29.88(-)/
0.9020(-)
36.67(-)/
0.9694(-)
VDSR [7]37.53(-)/
0.9587(-)
33.03(-)/
0.9124(-)
31.90(-)/
0.8960(-)
30.76(-)/
0.9140(-)
37.22(-)/
0.9729(-)
DRCN [18]37.63(-)/
0.9588(-)
33.04(-)/
0.9118(-)
31.85(-)/
0.8942(-)
30.75(-)/
0.9133(-)
37.63(-)/
0.9723(-)
CNF [38]37.66(-)/
0.9590(-)
33.38(-)/
0.9136(-)
31.91(-)/
0.8962(-)
--
LapSRN [39]37.52(-)/
0.9590(-)
33.08(-)/
0.9130(-)
31.80(-)/
0.8950(-)
30.41(-)/
0.9100(-)
37.27(-)/
0.9740(-)
DRRN [17]37.74(-)/
0.9591(-)
33.23(-)/
0.9136(-)
32.05(-)/
0.8973(-)
31.23(-)/
0.9188(-)
37.92(-)/
0.9760(-)
BTSRN [40]37.75(-)/-33.20(-)/-32.05(-)/-31.63(-)/--
MemNet [41]37.78(-)/
0.9597(-)
33.28(-)/
0.9142(-)
32.08(-)/
0.8978(-)
31.31(-)/
0.9195(-)
37.72(-)/
0.9740(-)
SelNet [42]37.89(-)/
0.9598(-)
33.61(-)/
0.9160(-)
32.08(-)/
0.8984(-)
--
CARN [19]37.76(-)/
0.9590(-)
33.52(-)/
0.9166(-)
32.09(-)/
0.8978(-)
31.92(-)/
0.9256(-)
38.36(-)/
0.9765(-)
RAN [43]37.58(-)/
0.9592(-)
33.10(-)/
0.9133(-)
31.92(-)/
0.8963(-)
--
DNCL [44]37.65(-)/
0.9599(-)
33.18(-)/
0.9141(-)
31.97(-)/
0.8971(-)
30.89(-)/
0.9158(-)
-
FilterNet [45]37.86(-)/
0.9610(-)
33.34(-)/
0.9150(-)
32.09(-)/
0.8990(-)
31.24(-)/
0.9200(-)
-
MRFN [46]37.98(-)/
0.9611(-)
33.41(-)/
0.9159(-)
32.14(-)/
0.8997(-)
31.45(-)/
0.9221(-)
38.29(-)/
0.9759(-)
SeaNet-baseline [48]37.99(-)/
0.9607(-)
33.60(-)/
0.9174(-)
32.18(-)/
0.8995(-)
32.08(-)/
0.9276(-)
38.48(-)/
0.9768(-)
DEGREE [47]37.58(-)/
0.9587(-)
33.06(-)/
0.9123(-)
31.80(-)/
0.8974(-)
--
SCANet (Ours)38.05/0.960733.54/0.917432.16/0.899532.10/0.928138.62/0.9770
× 3 SRCNN [5]32.75(-)/
0.9090(-)
29.28(-)/
0.8209(-)
28.41(-)/
0.7863(-)
26.24(-)/
0.7989(-)
30.59(-)/
0.9107(-)
FSRCNN [6]33.16(-)/
0.9140(-)
29.43(-)/
0.8242(-)
28.53(-)/
0.7910(-)
26.43(-)/
0.8080(-)
30.98(-)/
0.9212(-)
VDSR [7]33.66(-)/
0.9213(-)
29.77(-)/
0.8314(-)
28.82(-)/
0.7976(-)
27.14(-)/
0.8279(-)
32.01(-)/
0.9310(-)
DRCN [18]33.82(-)/
0.9226(-)
29.76(-)/
0.8311(-)
28.80(-)/
0.7963(-)
27.15(-)/
0.8276(-)
32.31(-)/
0.9328(-)
CNF [38]33.74(-)/
0.9226(-)
29.90(-)/
0.8322(-)
28.82(-)/
0.7980(-)
--
DRRN [18]34.03(-)/
0.9244(-)
29.96(-)/
0.8349(-)
28.95(-)/
0.8004(-)
27.53(-)/
0.8378(-)
32.74(-)/
0.9390(-)
BTSRN [40]34.03(-)/-29.90(-)/-28.97(-)/-27.75(-)/--
MemNet [41]34.09(-)/
0.9248(-)
30.00(-)/
0.8350(-)
28.96(-)/
0.8001(-)
27.56(-)/
0.8376(-)
32.51(-)/
0.9369(-)
SelNet [42]34.27(-)/
0.9257(-)
30.30(-)/
0.8399(-)
28.97(-)/
0.8025(-)
--
CARN [19]34.29(-)/
0.9255(-)
30.29(-)/
0.8407(-)
29.06(-)/
0.8034(-)
28.06(-)/
0.8493(-)
33.50(-)/
0.9440(-)
IMDN [49]34.36(-)/
0.9270(-)
30.32(-)/
0.8417(-)
29.09(-)/
0.8046(-)
28.17(+)/
0.8519(-)
33.61(+)/
0.9445(+)
RAN [43]33.71(-)/
0.9223(-)
29.84(-)/
0.8326(-)
28.84(-)/
0.7981(-)
--
DNCL [44]33.95(-)/
0.9232(-)
29.93(-)/
0.8340(-)
28.91(-)/
0.7995(-)
27.27(-)/
0.8326(-)
-
FilterNet [45]34.08(-)/
0.9250(-)
30.03(-)/
0.8370(-)
28.95(-)/
0.8030(-)
27.55(-)/
0.8380(-)
-
MRFN [46]34.21(-)/
0.9267(-)
30.03(-)/
0.8363(-)
28.99(-)/
0.8029(-)
27.53(-)/
0.8389(-)
32.82(-)/
0.9396(-)
SeaNet-baseline [48]34.36(-)/
0.9280(+)
30.34(+)/
0.8428(+)
29.09(-)/
0.8053(+)
28.17(+)/
0.8527(+)
33.40(-)/
0.9444(+)
DEGREE [47]33.76(-)/
0.9211(-)
29.82(-)/
0.8326(-)
28.74(-)/
0.7950(-)
--
SCANet (Ours)34.42/0.927230.32/0.842129.09/0.805128.14/0.852433.55/0.9443
× 4 SRCNN [5]30.48(-)/
0.8628(-)
27.49(-)/
0.7503(-)
26.90(-)/
0.7101(-)
24.52(-)/
0.7221(-)
27.66(-)/
0.8505(-)
FSRCNN [6]30.71(-)/
0.8657(-)
27.59(-)/
0.7535(-)
26.98(-)/
0.7150(-)
24.62(-)/
0.7280(-)
27.90(-)/
0.8517(-)
VDSR [7]31.35(-)/
0.8838(-)
28.01(-)/
0.7674(-)
27.29(-)/
0.7251(-)
25.18(-)/
0.7524(-)
28.83(-)/
0.8809(-)
DRCN [18]31.53(-)/
0.8854(-)
28.02(-)/
0.7670(-)
27.23(-)/
0.7233(-)
25.14(-)/
0.7510(-)
28.98(-)/
0.8816(-)
CNF [38]31.55(-)/
0.8856(-)
28.15(-)/
0.7680(-)
27.32(-)/
0.7253(-)
--
LapSRN [39]31.54(-)/
0.8850(-)
28.19(-)/
0.7720(-)
27.32(-)/
0.7280(-)
25.21(-)/
0.7560(-)
29.09(-)/
0.8845(-)
DRRN [18]31.68(-)/
0.8888(-)
28.21(-)/
0.7720(-)
27.38(-)/
0.7284(-)
25.44(-)/
0.7638(-)
29.46(-)/
0.8960(-)
BTSRN [40]31.85(-)/-28.20(-)/-27.47(-)/-25.74(-)/--
MemNet [41]31.74(-)/
0.8893(-)
28.26(-)/
0.7723(-)
27.40(-)/
0.7281(-)
25.50(-)/
0.7630(-)
29.42(-)/
0.8942(-)
SelNet [42]32.00(-)/
0.8931(-)
28.49(-)/
0.7783(-)
27.44(-)/
0.7325(-)
--
SRDenseNet [23]32.02(-)/
0.8934(-)
28.50(-)/
0.7782(-)
27.53(-)/
0.7337(-)
26.05(-)/
0.7819(-)
-
CARN [19]32.13(-)/
0.8937(-)
28.60(-)/
0.7806(-)
27.58(-)/
0.7349(-)
26.07(-)/
0.7837(-)
30.47(-)/
0.9084(+)
IMDN [49]32.21(+)/
0.8948(+)
28.58(-)/
0.7811(-)
27.56(-)/
0.7353(-)
26.04(-)/
0.7838(-)
30.45(-)/
0.9075(-)
RAN [43]31.43(-)/
0.8847(-)
28.09(-)/
0.7691(-)
27.31(-)/
0.7260(-)
--
DNCL [44]31.66(-)/
0.8871(-)
28.23(-)/
0.7717(-)
27.39(-)/
0.7282(-)
25.36(-)/
0.7606(-)
-
FilterNet [45]31.74(-)/
0.8900(-)
28.27(-)/
0.7730(-)
27.39(-)/
0.7290(-)
25.53(-)/
0.7680(-)
-
MRFN [46]31.90(-)/
0.8916(-)
28.31(-)/
0.7746(-)
27.43(-)/
0.7309(-)
25.46(-)/
0.7654(-)
29.57(-)/
0.8962(-)
SeaNet-baseline [48]32.18(-)/
0.8948(-)
28.61(-)/
0.7822(+)
27.57(-)/
0.7359(-)
26.05(-)/
0.7896(+)
30.44(-)/
0.9088(+)
DEGREE [47]31.47(-)/
0.8837(-)
28.10(-)/
0.7669(-)
27.20(-)/
0.7216(-)
--
SCANet (Ours)32.20/0.894728.61/0.782127.58/0.736526.17/0.789330.50/0.9083
Table 3. Subjective quality comparisons with different no-reference image quality indicators.
Table 3. Subjective quality comparisons with different no-reference image quality indicators.
IndicatorsSRCNNVDSRSCANet
NIQE ↓7.175.855.48
BRISQUE ↓58.8741.6439.51
Table 4. Subjective quality comparisons with different no-reference image quality indicators.
Table 4. Subjective quality comparisons with different no-reference image quality indicators.
MethodsSRCNNVDSRSCANet
SRCNN011
VDSR−101
SCANet−1−10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Wang, J.; Liu, X. Deep Successive Convex Approximation for Image Super-Resolution. Mathematics 2023, 11, 651. https://doi.org/10.3390/math11030651

AMA Style

Li X, Wang J, Liu X. Deep Successive Convex Approximation for Image Super-Resolution. Mathematics. 2023; 11(3):651. https://doi.org/10.3390/math11030651

Chicago/Turabian Style

Li, Xiaohui, Jinpeng Wang, and Xinbo Liu. 2023. "Deep Successive Convex Approximation for Image Super-Resolution" Mathematics 11, no. 3: 651. https://doi.org/10.3390/math11030651

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop