Next Article in Journal
Distortion Correction and Denoising of Light Sheet Fluorescence Images
Next Article in Special Issue
AcquisitionFocus: Joint Optimization of Acquisition Orientation and Cardiac Volume Reconstruction Using Deep Learning
Previous Article in Journal
Quantitative Prediction of Surface Hardness in Cr12MoV Steel and S136 Steel with Two Magnetic Barkhausen Noise Feature Extraction Methods
Previous Article in Special Issue
A Survey of Deep Learning Road Extraction Algorithms Using High-Resolution Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Burst-Enhanced Super-Resolution Network (BESR)

1
Aerospace Information Research Institute, Chinese Academy of Sciences, No. 9 Dengzhuang South Road, Haidian District, Beijing 100094, China
2
School of Optoelectronics, University of Chinese Academy of Sciences, No. 19(A) Yuquan Road, Shijingshan District, Beijing 100049, China
3
Department of Key Laboratory of Computational Optical Imagine Technology, Chinese Academy of Sciences, No. 9 Dengzhuang South Road, Haidian District, Beijing 100094, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(7), 2052; https://doi.org/10.3390/s24072052
Submission received: 3 March 2024 / Revised: 20 March 2024 / Accepted: 22 March 2024 / Published: 23 March 2024

Abstract

:
Multi-frame super-resolution (MFSR) leverages complementary information between image sequences of the same scene to increase the resolution of the reconstructed image. As a branch of MFSR, burst super-resolution aims to restore image details by leveraging the complementary information between noisy sequences. In this paper, we propose an efficient burst-enhanced super-resolution network (BESR). Specifically, we introduce Geformer, a gate-enhanced transformer, and construct an enhanced CNN-Transformer block (ECTB) by combining convolutions to enhance local perception. ECTB efficiently aggregates intra-frame context and inter-frame correlation information, yielding an enhanced feature representation. Additionally, we leverage reference features to facilitate inter-frame communication, enhancing spatiotemporal coherence among multiple frames. To address the critical processes of inter-frame alignment and feature fusion, we propose optimized pyramid alignment (OPA) and hybrid feature fusion (HFF) modules to capture and utilize complementary information between multiple frames to recover more high-frequency details. Extensive experiments demonstrate that, compared to state-of-the-art methods, BESR achieves higher efficiency and competitively superior reconstruction results. On the synthetic dataset and real-world dataset of BurstSR, our BESR achieves PSNR values of 42.79 dB and 48.86 dB, respectively, outperforming other MFSR models significantly.

1. Introduction

As a crucial low-level computer vision task, image super-resolution helps to drive research in advanced tasks such as target detection [1,2] and image segmentation [3], in in addition to reconstructing image details. Traditional image super-resolution methods are mainly based on interpolation and filtering techniques. Although these methods can increase the resolution of the image, they cannot restore the lost detail information. In recent years, with the rise of deep learning, image super-resolution methods based on neural networks have made significant breakthroughs in improving reconstruction accuracy and visual effects [4]. Super-resolution technology has been widely used in video surveillance [5], remote sensing [6], medical image diagnosis [7], and other fields [8,9,10]. In particular, with the increasing popularity and professionalization of mobile photography, burst super-resolution technology is attracting more and more attention. Burst is a common photography mode, where multiple images are captured in rapid succession within a short period of time. Due to camera shake, there exists sub-pixel displacement among these multiple images. Burst super-resolution [11] aims to restore image details by utilizing the complementary information from multiple displaced images.
SRCNN [12] is a three-layer convolutional network, as well as the first attempt at deep learning methods for super-resolution problems. The residual network structure addresses the issue of gradient vanishing that arises from increased network depth, leading to improved convergence and enhanced model performance. VDSR [13] uses residual learning to increase the number of layers of the super-resolution network to 20 for the first time. A generative adversarial network (GAN) consists of a generator and a discriminator. SRGAN [14] introduces GAN into the super-resolution task, enabling the generation of more realistic textures in the reconstruction. ESRGAN [15] and SFTGAN [16] optimize the adversarial loss function and incorporate prior information, respectively. Other convolutional network-based super-resolution algorithms, such as [17,18,19], achieve better training results through deeper and more intricate network structures.
In recent years, Transformer has attracted widespread attention in the field of NLP. Thanks to its parallel computing and long-distance dependency modeling capabilities, it has significantly improved the model’s ability to learn rich semantic representations. ViT [20] introduces a Transformer into the computer vision field and has achieved remarkable success. For pixel-level visual tasks, such as image segmentation and super-resolution, the introduction of the Transformer model enhances the network’s understanding of global semantics and can capture the correlation between different pixels or semantics. TTSR [21] introduces the Transformer model to the super-resolution task for the first time, learning the deep correspondence between the low-resolution (LR) image and the reference image through self-attention, and transferring similar textures in the reference image to restoring the high-resolution (HR) image. However, the original Transformer has a complexity of O(n2), which comes with high computational costs.
Currently, in the task of super-resolution, Transformer models with linear complexity mainly fall into two categories. One approach involves computing self-attention within a window, as seen in SwinIR [22], which divides the input image into non-overlapping windows and achieves cross-window information interaction through shifting. HAT [23] introduces overlapped cross-attention modules to further enhance the aggregation of information across windows. Another approach involves implicitly encoding global information across feature dimensions, as demonstrated by Restormer [24], which computes cross-channel covariance to implicitly model global information. SAFMN [25] introduces a convolutional channel mixer to simultaneously encode local context and blend channels, achieving global information interaction. We further optimize the implicit global information encoding method and propose a more efficient self-attention mechanism.
Multiple frames of images are sampled from different sub-pixel positions, exhibiting sub-pixel displacement relationships. Multi-frame super-resolution can additionally leverage inter-frame complementary information, recovering more details compared to single-frame super-resolution, and has broad application prospects in fields such as computational photography [26,27], remote sensing satellite imaging [28,29], etc. Commonly used MFSR datasets include PROBA-V [30] for remote sensing tasks and BurstSR [31] for computational photography tasks. The performance improvement of multi-frame super-resolution relies on the sub-pixel-level alignment of multiple frames, and the accuracy of alignment directly impacts the reconstruction results.
Inter-frame alignment and feature fusion are key challenges in burst super-resolution. Frame alignment methods are mainly divided into two types: one involves explicit frame alignment through optical flow estimation and motion compensation [32], and the other employs deformable convolution to learn dynamic sampling positions of adjacent frames for implicit frame alignment [33]. However, existing methods have certain limitations in extracting spatiotemporal correlation information. Therefore, we propose two optimized modules to extract more complementary information from aligned multi-frame images to recover high-frequency details.
To more effectively capture global information, we propose a gate-enhanced Transformer, Geformer. By computing self-attention along the channel dimension, Geformer can implicitly encode global information while maintaining linear complexity. The gating mechanism enables Geformer to effectively model correlations between domain-specific features. We leverage reference features to facilitate inter-frame communication, simultaneously serving as an implicit image co-registration condition to enhance spatiotemporal information consistency. By combining Geformer with convolution, we construct an enhanced CNN-Transformer block (ECTB), providing enhanced feature representation for the two crucial processes of inter-frame alignment and feature fusion. Consequently, we introduce an efficient burst-enhanced super-resolution network, BESR.
We conduct extensive experiments on synthetic and real-world datasets of BurstSR. The results demonstrate a significant improvement in perceptual quality and detail fidelity compared to previous MFSR methods. In terms of PSNR and SSIM metrics, our BESR achieves a gain of 0.35 dB and 0.06 dB, respectively, over the state-of-the-art method RBSR on the two datasets, while having lower parameters. The extensive experimental results provide strong evidence for the effectiveness of our approach.
The main contributions of this paper are summarized as follows:
1.
We propose Geformer, a gate-enhanced Transformer model, and combine Geformer with convolution to construct an enhanced CNN-Transformer block (ECTB), enabling the network to simultaneously perceive global information and local details, thereby achieving enhanced feature representation.
2.
We enhance the interaction between inter-frame information and boost the correlation of spatiotemporal features by constructing reference features, providing more effective information for the entire super-resolution process.
3.
Addressing the critical issues of inter-frame alignment and feature fusion, we introduce optimized pyramid alignment (OPA) and hybrid feature fusion (HFF) modules to fully leverage non-redundant information between frames. Our BESR achieves the optimal reconstruction performance on both synthetic and real-world datasets of BurstSR.

2. Related Work

2.1. Single-Image Super-Resolution (SISR)

SRCNN is the first method to use deep learning to construct a nonlinear mapping between LR and HR, which achieves SOTA performance on multiple public datasets with only three convolutional layers. Subsequent research has shown that by increasing network depth and introducing residual learning, such as [13,34] and other methods, image quality and perceptual quality can be further improved. Research on generative adversarial networks (GANs), such as [14,35], is introduced to generate more realistic and detailed high-resolution images through adversarial training. GANs not only enhance quantitative evaluation metrics but also improve subjective perceptual quality.
Recent studies suggest that attention mechanisms, through feature selection, can strengthen the extraction of important information while suppressing less relevant details, thereby enhancing the network’s feature extraction capabilities. For instance, RCAN [36] combines residual learning with channel attention mechanisms and trains a super-resolution network with over 400 layers. HAN [37] introduces layer attention and channel-wise spatial attention modules to learn the direct correlation between layer, channel, and spatial features, resulting in an enhanced feature representation.
Due to the inherent advantage of Transformer in modeling long-distance dependencies, Transformer-based super-resolution models have achieved significant performance improvements. SwinIR [22] divides the input image into windows, computing self-attention between patches within each window and using a moving window to model distant dependencies, thus reducing computational complexity linearly. HAT [23] introduces overlapping cross-attention modules to enhance information interaction between adjacent windows. However, the local windows limit the utilization of crucial global information in image super-resolution. Restormer [24] implicitly encodes global information by computing self-attention across feature channels, resulting in a more lightweight super-resolution network. We have improved its limitations in spatial modeling by introducing a gating mechanism to dynamically modulate self-attention, enabling better modeling of the correlation of neighborhood features and extracting more effective global information.

2.2. Multi-Frame Super-Resolution (MFSR)

The research on multi-frame super-resolution is first proposed by Tsai and Huang [38]. Compared with SISR methods, MFSR is able to utilize additional inter-frame complementary information to achieve better high-frequency detail recovery. Farsiu et al. [39] propose a bilateral total variation regularization method based on the total variation super-resolution approach. Wronski et al. [27] utilized kernel regression to align input frames, introduced radial kernels for local adaptive detail enhancement, and achieved mosaic removal through multi-frame super-resolution. Deudon et al. [40] focus on MFSR in satellite imagery, addressing subtasks such as joint learning of registration, fusion, and upsampling. They introduced the first deep learning-based multi-frame super-resolution network, HighRes-net.
Bhat et al. [11] introduce the baseline DBSR for burst super-resolution, aligning multiple frames through optical flow and utilizing attention modules to fuse aligned frame information, addressing real-world MFSR challenges. Later, Bhat et al. [41] optimized the multi-frame fusion mechanism by modeling the image formation process in the latent space and conducting a deep reparameterization of the classical MAP formulation. Dudhane et al. [42] propose a pseudo-burst fusion strategy for inter-frame information interaction, enhancing it with multiscale context. Luo et al. [43] employ optical flow-guided deformable convolution for inter-frame alignment, using Swin Transformer as the backbone network for deep feature extraction and image reconstruction.
These models often come with complex parameters and computational requirements. In comparison, the lightweight MFSR model is more practical. Wu et al. [44] proposed a recursive model, RBSR, which merges inter-frame complementary information frame by frame and introduces implicit weighted loss to handle sequences of varying lengths. We have further improved frame alignment by constructing a feature pyramid to integrate multi-scale aligned features, achieving more precise alignment effects, and providing more complementary information for feature fusion.

2.3. Efficient Vision Transformers

In recent years, Vision Transformers have experienced rapid development, demonstrating outstanding performance in various tasks. However, constrained by high computational costs, they are not ideal for real-time applications. Therefore, researchers have started to explore lightweight Vision Transformer models. For example, more efficient attention mechanisms have been introduced into the model, such as sparse attention [45,46,47] and local attention [48,49,50]. CvT [51] uses convolutional mapping instead of position-wise linear projection and undersamples K and V matrices to improve attention efficiency. Researchers have also explored methods such as model pruning [52] and parameter quantization [53] to compress model size, reducing memory requirements while maintaining high performance. The research on efficient Vision Transformers is expected to provide feasible solutions for scenarios with limited computational resources. In this study, we propose an efficient Transformer model, Geformer. Introducing a gating mechanism enhances the modeling of spatial relationships in the neighborhood, allowing for the implicit encoding of global information with higher computational efficiency.

3. Methodology

3.1. Network Architecture

The proposed BESR framework, as illustrated in Figure 1, is primarily based on four key components: shallow feature extraction, optimized pyramid alignment (OPA), hybrid feature fusion (HFF), and image reconstruction.
Specifically, the input of the network is the RAW burst sequences LR, whose size is N × H × W × C0. Among them, N, H, W, and C0, respectively, represent the frames, height, width, and channels of the LR. The specific values are 14, 48, 48, and 4, respectively. The output of the super-resolution network is a high-resolution RGB image SR with a size of 8H × 8W × 3. We generate the reference feature by repeating the 0-th frame of the input feature N times, and then concatenate the input feature and the reference feature along the channel dimension to obtain the feature F C , whose size is N × H × W × 2C0.
LR ref [ i ] = LR   [ 0 ] ,   i [ 0 , N 1 ]
F C [ n , h , w , c ] = LR [ n , h , w , c ] ,   c [ 0 , C 1 ] LR ref n , h , w , c C ,   c [ C , 2 C 1 ]
where i denotes the i-th frame of the reference feature. We employ a single layer of 3 × 3 convolution to aggregate information from concatenated features, mapping them to a high-dimensional feature representation, thus obtaining shallow feature F S with a size of N × H × W × C. Here, C represents the channels in the intermediate features, which we set to 48.
F S = C o n v 3 ( F C )
where C o n v 3 represents a 3 × 3 convolution. The enhancement of reference features can not only promote information interaction between frames, but also serve as an implicit multi-frame co-registration condition, helping to reduce the impact of alignment errors.
In the optimized pyramid alignment (OPA) module, we employ a combination of deformable convolution and enhanced CNN-Transformer blocks (ECTB) for multi-scale feature alignment. Subsequently, through the refinement of a cascaded alignment network, we can capture fine-grained inter-frame displacements, achieving more precise feature alignment and obtaining enhanced aligned features, denoted as F A , with dimensions N × H × W × C.
F A = OPA ( F S )
Our BESR can flexibly handle different numbers of inputs. To capture more effective information from the inputs, we use all 14 bursts from the same scene as the input to the model. For convenience in subsequent feature transformations, we adjust the size of the aligned features by applying the average pooling operator in the AvgPool module to obtain the pooled features F P as the input to the fusion module, with a size of 8 × H × W × C.
F P = AvgPool ( F A )
In the hybrid feature fusion (HFF) module, we integrate inter-frame complementary information using ECTBs and facilitate inter-frame communication with reference features, enhancing the spatiotemporal correlations among multiple frames. Subsequently, we combine 1 × 1 convolution and Pixel Shuffle to upsample the fused features, achieving the restoration of high-frequency details. Finally, through a 3 × 3 convolution layer, we map the fused high-level semantic feature F U (with dimensions 8H × 8W × C) to a high-resolution RGB image SR, sized 8H × 8W × 3.
F U = HFF ( F P )
SR = C o n v 3 ( F U )

3.2. Enhanced CNN-Transformer Block

For pixel-level visual tasks such as image segmentation and super-resolution, when the input image size is H × W pixels, the complexity of the Transformer model is O(H2W2). The quadratic complexity results in high computational costs. In super-resolution tasks, the key lies in capturing similar features. Features in spatial proximity exhibit higher similarity, while those at spatially distant locations demonstrate lower similarity. Standard Transformers acquire attention maps through global computations, which results in a large number of redundant calculations. At the same time, it also lacks local perception capabilities and does not capture high-frequency information well. To solve these problems, we propose an enhanced CNN-Transformer block (ECTB), as shown in Figure 2.
X F = Geformer ( Geformer ( X ) )
X E = X F + Local   Enh ( X F     X ) )
Here, X represents the input feature, while X F and X E , respectively, denote the intermediate and output features of ECTB, all with dimensions H × W × C. By adopting a CNN-Transformer hybrid structure, ECTB can simultaneously perceive global context and local detailed features, resulting in an enhanced feature representation.

3.2.1. Geformer

Inspired by Restormer [24], we model global correlations by computing the covariance across feature channels, and integrate gating mechanisms in the attention and FFN components to model neighborhood correlations. Consequently, we propose a gate-enhanced Transformer called Geformer, whose structure is illustrated in Figure 3. It comprises two core components: Transposed Gating Attention (TGA) and Spatial-Gate Modulated Network (SGMN).
X N = X i n + TGA ( L N ( X i n ) )
X M = X N + SGMN ( L N ( X N ) )
Among them, X i n represents the input feature, while X N and X M represent the intermediate and output features of Geformer, respectively, and their sizes are H × W × C. LN represents the LayerNorm layer. Through the introduced gating mechanism, Geformer achieves dynamic spatial modulation of self-attention, which can better capture the correlation of neighborhood features and provide more effective contextual information for super-resolution tasks.
Transposed Gating Attention (TGA) is a gate-enhanced self-attention mechanism we proposed, as illustrated in Figure 4. It enhances the ability to extract global correlations by calculating cross-covariance across feature channels and introducing a gating mechanism to model spatial neighborhood correlations.
Specifically, given an input feature X 0 of size H × W × C, we expand its channels through two 1 × 1 convolutions. Subsequently, we enhance local contextual awareness through a 3 × 3 depth-wise separable convolution. After reshaping the features, we obtain the query projection (Q), the value projection (V), and two key projections ( K 1 and K 2 ), where Q, V, K 1 , and K 2 are all of size HW × C.
Q , V = S p l i t ( D W C o n v 31 ( C o n v 11 ( X 0 ) ) )
K 1 , K 2 = S p l i t ( D W C o n v 32 ( C o n v 12 ( X 0 ) ) )
Among them, C o n v 11 and D W C o n v 31 represent the first 1 × 1 convolution and 3 × 3 depth-separable convolution, and C o n v 12 and D W C o n v 32 represent the second 1 × 1 convolution and 3 × 3 depth-separable convolution. By transposing Q, we obtain the transposed projection Q T with dimensions C × HW. We perform element-wise multiplication between Q T and K 1 as well as Q T and K 2 to assess the similarity between the input features, yielding two weight matrices, W 1 and   W 2 . Subsequently, we apply GELU to activate W 1 , serving as a gating modulation unit for W 2 . By element-wise multiplication of two matrices and applying softmax mapping, the modulated attention map Attn is obtained. Then, it is multiplied with V to enhance similar features. Finally, through 1 × 1 convolution, the TGA produces output features X A with a size of H × W × C.
W 1 = Q T K 1 ,   W 2 = Q T K 2
A t t n = S o f t m a x ( GELU ( W 1 ) · W 2 / α )
X A = C o n v 1 ( V A t t n )
Here, α is a learnable scaling parameter, and ⊗ represents matrix multiplication. Compared to the original self-attention, TGA can effectively capture the correlations between spatial neighborhood features, enhancing the exploration of global information.
The Feedforward Neural Network (FFN) introduces non-linearity to the Transformer, but its limitations in feature fitting arise due to the implicit modeling of spatial and channel relationships. Therefore, we propose the Spatial-Gate Modulated Network (SGMN), a gate-enhanced efficient feedforward network, as illustrated in Figure 5.
Specifically, the network takes the enhanced feature X T obtained from TGA as input. We perform high-dimensional mapping on X T using a 1 × 1 convolution, followed by nonlinear activation with GELU. Along the channel dimension, we partition the features into two parts, resulting in two features X 1 and X 2 of size H × W × Ch, where Ch denotes the channels in the high-dimensional feature mapping. One branch undergoes local contextual enhancement via depth-wise separable convolution, while the other branch undergoes an identity mapping. Spatial modulation is achieved by element-wise multiplication of these two features. Finally, we aggregate information along the spatial dimension using a 1 × 1 convolution to obtain the nonlinearly enhanced feature representation X N L with a size of H × W × C.
X 1 , X 2 = S p l i t ( GELU ( C o n v 1 ( X T ) ) )
X N L = C o n v 1 ( D W C o n v 3 ( X 1 ) · X 2 )
Compared with standard FFN, our SGMN can explicitly model spatial information while reducing redundant information between channels and has a smaller number of parameters.
Compared to existing Transformer models, our proposed Geformer maintains linear instead of quadratic computational complexity by establishing inter-feature channel information interaction. Benefiting from the gating mechanisms introduced in TGA and SGMN, Geformer can model spatial neighborhood correlations while capturing global information, thus providing powerful feature learning capabilities for super-resolution reconstruction tasks.

3.2.2. Local Enhancement

To enhance the perception of local features, we construct a local enhancement (Local Enh) module (as shown in Figure 2) for extracting local residuals, which are then added to the output of the Geformer to obtain augmented hybrid features. The Local Enh module comprises two cascaded layers of 3 × 3 convolutions and GELU, designed to excite and compress features, thereby capturing more local detail information. Given an input feature X G , the output feature X C of Local Enh can be expressed as:
X C = GELU ( C o n v 3 ( GELU ( C o n v 3 ( X G ) ) ) )
The first layer of convolution doubles the channels in the feature map, while the second layer of convolution recovers the number of channels by reduction. The size of both X G and X C is H × W × C.

3.3. Optimized Pyramid Alignment

Compared with SISR, MFSR can utilize the complementary information between frames by aligning multiple frames with sub-pixel displacements to provide more effective information for the reconstruction of high-frequency details of the image. Inter-frame alignment is a key issue in multi-frame super-resolution. Effective alignment can strengthen the spatiotemporal correlation between frames, allowing the model to capture sub-pixel differences between frames. Enhancing through the fusion of adjacent frame information is helpful in mitigating motion-induced blurriness and distortion.
Deformable alignment uses deformable convolution to estimate the offset and resample adjacent frames to achieve alignment with the reference frame. Deformable convolution extends traditional convolution by introducing a learnable offset for each convolution sampling point to adjust the sampling points at each position in the convolutional kernel. These offsets are generated by the input feature map and another convolution.
Specifically, given a convolutional kernel with K sampling positions, where w k and n k represent the weight and pre-set offset of the k-th position, and let x(n) and y(n) denote the features at position n in the input feature map x and output feature map y, deformable convolution can be defined as follows:
  y = k = 1 K w k · x n + n k + Δ n k · Δ m k
where Δ n k and Δ m k represent the learnable offset and modulation scalar at the k-th position, with Δ m k constrained within the range [0, 1]. As the coordinates after incorporating Δ n k are typically non-integer, resulting in irregular sampling positions, we employ bilinear interpolation to resample the feature map and obtain the displaced features.
Based on deformable convolution, we propose an optimized pyramid alignment (OPA) module, as shown in Figure 6. In each layer of the alignment network, we first aggregate intra-frame context and inter-frame correlation information through the proposed ECTB to obtain enhanced feature representation, and then utilize deformable convolution to perform motion estimation and compensation on the aggregated features. In the reference enhancement (Ref Enh) module, we concatenate and fuse the input feature F D and the reference feature F r e f along the channel dimension to promote inter-frame communication and enhance the mining of spatiotemporal associated information.
F C [ n , h , w , c ] = F D [ n , h , w , c ] ,   c [ 0 , C 1 ] F r e f n , h , w , c C ,   c [ C , 2 C 1 ]
F R = GELU ( C o n v 3 ( F C ) )
where F C and F R represent concatenated features and fused features, respectively. OPA is a top-down three-level feature pyramid, where each alignment network layer consists of ECTBs on both sides, along with a deformable convolution and a reference enhancement module. To generate the input feature at the (i + 1)-th level, we utilize a convolution with a 3 × 3 kernel and a stride of 2 to downsample the input at the i-th pyramid level.
F i + 1 = C o n v S 2 ( F i )
where F i and F i + 1 represent the input features of i-th level and (i + 1)-th level, respectively. The input of the i-th deformable convolution consists of two parts: the refined feature output by the i-th level ECTB and the offset output by the (i + 1)-th level deformable convolution. The aligned features and offsets output by the (i + 1)-th level are upsampled through Pixel Shuffle and deconvolution, respectively, and then input to the i-th level. For each level of alignment network,
F i = ECTB R ( Ref   Enh ( D e C o n v ( E C T B L ( F i ) ) ) )
F M = Up ( Up ( F 2 ) + F 1 ) + F 0
Among them, F i represents the output features and alignment features of the i-th level. And ECTB L and ECTB R , respectively, represent the ECTB blocks on both sides of each layer of the aligned network. By integrating the output of the three-layer network, we obtain the multi-scale aligned feature F M . By repeating the 0-th frame of F M 8 times, we construct reference features to provide more contextual information for feature fusion.
R e f [ i ] = F M [ 0 ] ,   i [ 0 , 7 ]
To further enhance the spatiotemporal consistency of inter-frame features, we cascade another layer of alignment network after the pyramid structure to refine the F M and obtain an enhanced alignment feature F A , which further improves the alignment accuracy and can provide more supplementary information for reconstruction.
F A = ECTB R ( Ref   Enh ( D e C o n v ( E C T B L ( F M ) ) ) )

3.4. Hybrid Feature Fusion

For feature fusion, we propose a hybrid feature fusion module (HFF), as shown in Figure 7. The three-layer network enables the model to extract more effective information from the fused features. We use reference features to enhance inter-frame correlations and add them to the aligned features as input to the fusion module. The introduction of reference features provides additional reference information for feature fusion while facilitating the interaction of interframe information. By introducing an ECTB to capture and aggregate global and local contextual information in input features, the network can more fully exploit inter-frame complementary information, resulting in finer high-frequency detail recovery.
In HFF model, the input of the (i + 1)-th level network is obtained by adding the output features of the i-th level network and the reference features. At each level of the network, the input features are fused with complementary information between frames through ECTB. We then utilize 1 × 1 convolution for channel expansion and upsampling via pixel shuffling to recover the lost details.
Y i + 1 = Pixel   Shuffle ( C o n v 1 ( ECTB ( Y i + R e f i ) ) )
where Y i and R e f i represent the output feature and reference feature of the i-th level, respectively. And Y i + 1 represents the output feature of the (i + 1)th-level. The stepwise feature fusion and upsampling strategy helps to preserve the spatial information of the image while mitigating the impact of noise and blurring factors, thereby enhancing the reconstruction effect. To reduce the model’s parameters, we use ECTB only in the first two layers and share parameters between the three upsampling layers of reference features to achieve a lightweight model while maintaining high performance.

4. Experiments

4.1. BurstSR Dataset

BurstSR is a benchmark for the MFSR task proposed by the NTIRE 2022 Burst Super-Resolution Challenge, which includes synthetic and real-world datasets. Among them, the synthetic dataset consists of 46,839 bursts for training and 300 bursts for validation. Each burst contains 14 RAW images of size 48 × 48 pixels. These images are obtained by converting the raw images from sRGB to linear space by inverse camera pipeline, then randomly translating, rotating, and downsampling each image by bilinear interpolation, and finally, mosaicking using Bayer mode. The real-world dataset includes 5405 bursts for training and 882 bursts for validation. The LR images and HR images are captured by a smartphone and a DSLR camera with a zoom lens, respectively. They exhibit certain misalignments in both spatial and color aspects. Therefore, we employ aligned L1 loss and perceptual loss for model training and evaluate the model performance using aligned PSNR and SSIM.

4.2. Evaluation Metrics

Quantitative evaluation based on objective assessment methods is the mainstream evaluation metric in the current field of super-resolution reconstruction. Among them, peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) are two of the most commonly used metrics for image quality evaluation. PSNR is the most widely used image reconstruction quality evaluation metric in the current task of super-resolution reconstruction. It measures the quality of an image by calculating the differences between corresponding pixels in the reconstructed image and the ground truth image. A higher PSNR value indicates less distortion in the reconstructed image. SSIM is a full-reference image quality assessment metric used to measure the similarity between a reconstructed image and a ground truth image in terms of brightness, contrast, and structure. It provides a more accurate reflection of human perception of image quality, and a value closer to 1 indicates higher similarity between the reconstructed image and the ground truth image. Learned Perceptual Image Patch Similarity (LPIPS) is a deep learning-based image quality assessment metric. It evaluates the quality of images by learning human perception, where a lower value indicates greater similarity between two images. We use LPIPS along with PSNR and SSIM as evaluation metrics for the super-resolution models.

4.3. Training Details

We choose the Adam optimizer for model training and set the decay rate parameters β1 and β2 to 0.9 and 0.999. During model training, we use the L1 norm as the loss function and set the batch size to 4. The learning rate is gradually reduced during the training phase according to the decay strategy of cosine annealing, from the initial value of 1 × 10−4 to 1 × 10−6. We trained the proposed BESR model for 300 epochs using 4 NVIDIA A5000 GPUs (NVIDIA, Beijing, China) based on the PyTorch 1.10.2 framework. For real-world data, we use the model weights trained on synthetic data to fine-tune it, with a period of 75 epochs, and the learning rate is reduced from 1 × 10−5 to 3 × 10−6.

4.4. Comparison with State-of-the-Art Methods

4.4.1. Quantitative Results

We quantitatively compare the proposed BESR with other state-of-the-art MFSR methods, including DBSR, HighRes-net, MFIR, BIPNet, BSRT-S, and RBSR, on benchmark synthetic datasets and real-world datasets. We use PSNR and SSIM as objective evaluation indicators of the reconstruction performance of different methods and report the parameters and inference time of the models, and the comparison results are shown in Table 1. The time in the table represents the inference time required to generate a single SR image of size 384 × 384.
The comparison results indicate that, compared to other MFSR methods, our BESR achieves the best reconstruction results on almost all evaluation metrics, fully demonstrating the effectiveness of the proposed approach. Thanks to the efficient self-attention mechanism in Geformer, BESR maintains lower model parameters compared to other methods while enhancing performance. By adopting a CNN-Transformer hybrid structure, our model can simultaneously capture local details and global context, providing more effective information for the super-resolution task. Compared to the latest model RBSR, our method achieves a PSNR gain of 0.35 dB and 0.06 dB on the synthetic dataset and real-world dataset of BurstSR, respectively, while also obtaining competitive LPIPS results.

4.4.2. Visual Comparisons

The visual results of the synthetic and real-world datasets are shown in Figure 8, Figure 9 and Figure 10 and Figure 11, Figure 12 and Figure 13, respectively. The red box represents the target area that we have selected. Comparison results across different scenes indicate that, compared to other MFSR methods, our proposed BESR can better restore high-frequency details lost in the images and exhibit better robustness to noise, resulting in visually appealing HR images.
For example, as shown in Figure 8 and Figure 10, our super-resolution results reconstruct sharp and natural edges, while the generated results from other methods exhibit significant texture distortions. As depicted in Figure 9, our method can recover more details of the wall, capturing relatively complete local structures, whereas the reconstruction results from other methods are blurry and suffer from severe detail distortions. These results demonstrate that our method effectively restores more high-frequency information.
From the visual comparisons in Figure 11 and Figure 13, it is evident that the reconstruction results from other methods exhibit significant blurring and detail distortions, whereas our reconstruction results have fewer artifacts and sharper texture details. Especially in the scene img_0077_0012, only our method effectively restores the details of the windows. In Figure 12, it can be observed that our reconstructed results have sharper lines and higher clarity.
The visual comparison results indicate that our method, while enhancing the overall structure, also focuses on local texture details, resulting in better super-resolution results, further demonstrating the exceptional reconstruction performance of the proposed method.

4.4.3. Ablation Study

In this section, we discuss the effectiveness of key modules in BESR, including the enhanced CNN-Transformer block, the optimized pyramid alignment module, and the hybrid feature fusion module. We select the corresponding baselines for the ablation study of the proposed modules and report the results of the ablation experiments on the benchmark dataset in Table 2, Table 3, Table 4 and Table 5. ✓ and ✗ in the table indicate the presence or absence of the module respectively.
  • Effectiveness of ECTB.
We use the Transformer Block proposed in Restormer as the corresponding baseline for the core component ECTB in our network. Our proposed Geformer can dynamically modulate self-attention through a gating mechanism, providing better modeling of the correlation of neighborhood features. ECTB adopts a CNN-Transformer hybrid structure, which can effectively capture and process both local and global information in the image, thereby providing more effective information for the recovery of high-frequency details.
Compared to the baseline, ECTB is able to obtain enhanced feature representations, resulting in a 0.47 dB gain in PSNR for improved reconstruction performance. We report the impact of different core components on the model’s reconstruction performance in Table 3. The results show that our model maintains high performance while having lower parameters.
  • Effectiveness of OPA.
We chose to use the PCD alignment module proposed in EDVR [55] as the baseline for the OPA module ablation study. By adopting a CNN-Transformer hybrid structure, ECTB has the ability to simultaneously perceive local features and model global information and can effectively capture the spatiotemporal correlation of multi-frame features. In Table 4, ECTB L and ECTB R , respectively, represent the ECTBs on the left and right sides of each layer in the alignment network, playing a crucial role in improving the alignment effect. The OPA module performs inter-frame alignment at different scales and obtains more refined alignment features through aggregation. We fuse the reference frame features with the current frame features to further enhance the correlation of inter-frame features, thereby extracting more inter-frame complementary information. Compared with PCD, the OPA module brings a PSNR gain of 1.13 dB to improve model performance.
  • Effectiveness of HFF.
We construct a baseline for HFF module ablation research using three upsampling layers consisting of 1 × 1 convolution and Pixel Shuffle. The ECTB proposed in this study can perform fine-grained feature fusion on the obtained alignment features to fully utilize the complementary information between frames for high-frequency detail reconstruction. In addition, the introduction of reference features helps to extract more different information from aligned features, providing an additional reference for reconstruction while alleviating the problem of edge information loss in image features. Compared with the baseline, the model using two layers of ECTB achieved a PSNR gain of 0.68 dB, demonstrating its effectiveness for feature fusion.

5. Conclusions

In this study, we propose an efficient burst-enhanced super-resolution network, BESR. Based on the efficient Geformer introduced in the paper, we construct an enhanced CNN-Transformer block that effectively aggregates intra-frame context and inter-frame correlation information from multiple frames. Additionally, we leverage reference features to facilitate inter-frame communication, enhancing spatiotemporal coherence among multiple frames. To address the challenges of inter-frame alignment and feature fusion, we introduce optimized pyramid alignment and hybrid feature fusion modules to extract and utilize complementary information between frames, providing more effective information for the restoration of high-frequency details. Extensive experiments conducted on two benchmark datasets demonstrate that BESR achieves state-of-the-art reconstruction performance, producing high-resolution images with rich details and clear textures. In the future, our work will be dedicated to researching super-resolution algorithms for unsupervised models.

Author Contributions

Conceptualization, J.L. and Z.T.; methodology, J.L.; software, J.L.; investigation, J.L., W.Z. and Y.Z.; writing—original draft preparation, J.L.; writing—review and editing, J.L. and Z.T.; project administration, Q.L.; funding acquisition, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key Program Project of Science and Technology Innovation of the Chinese Academy of Sciences (no. KGFZD-135-20-03-02) and by the Innovation Foundation of the Key Laboratory of Computational Optical Imaging Technology, CAS (no. CXJJ-23S016).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, J.; Lei, J.; Xie, W.; Fang, Z.; Li, Y.; Du, Q. SuperYOLO: Super resolution assisted object detection in multimodal remote sensing imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  2. Li, R.; Shen, Y. YOLOSR-IST: A deep learning method for small target detection in infrared remote sensing images based on super-resolution and YOLO. Signal Process. 2023, 208, 108962. [Google Scholar] [CrossRef]
  3. Li, Y.; Luo, L.; Lin, H.; Heng, P.-A.; Chen, H. Scale-aware Super-resolution Network with Dual Affinity Learning for Lesion Segmentation from Medical Images. arXiv 2023, arXiv:2305.19063. [Google Scholar]
  4. Gendy, G.; He, G.; Sabor, N. Lightweight image super-resolution based on deep learning: State-of-the-art and future directions. Inf. Fusion 2023, 94, 284–310. [Google Scholar] [CrossRef]
  5. Hijji, M.; Khan, A.; Alwakeel, M.M.; Harrabi, R.; Aradah, F.; Cheikh, F.A.; Sajjad, M.; Muhammad, K. Intelligent Image Super-Resolution for Vehicle License Plate in Surveillance Applications. Mathematics 2023, 11, 892. [Google Scholar] [CrossRef]
  6. Ali, A.M.; Benjdira, B.; Koubaa, A.; Boulila, W.; El-Shafai, W. TESR: Two-Stage Approach for Enhancement and Super-Resolution of Remote Sensing Images. Remote Sens. 2023, 15, 2346. [Google Scholar] [CrossRef]
  7. Wu, Z.; Zhu, F.; Guo, K.; Sheng, R.; Chao, L.; Fang, H. Modal adaptive super-resolution for medical images via continual learning. Signal Process. 2024, 217, 109342. [Google Scholar] [CrossRef]
  8. Ariav, I.; Cohen, I. Fully cross-attention transformer for guided depth super-resolution. Sensors 2023, 23, 2723. [Google Scholar] [CrossRef] [PubMed]
  9. Ju, Y.; Jian, M.; Wang, C.; Zhang, C.; Dong, J.; Lam, K.-M. Estimating high-resolution surface normals via low-resolution photometric stereo images. IEEE Trans. Circuits Syst. Video Technol. 2023; early access. [Google Scholar] [CrossRef]
  10. Xiao, J.; Jia, W.; Lam, K.-M. Feature redundancy mining: Deep light-weight image super-resolution model. In Proceedings of the ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1620–1624. [Google Scholar]
  11. Bhat, G.; Danelljan, M.; Van Gool, L.; Timofte, R. Deep burst super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 19–25 June 2021; pp. 9209–9218. [Google Scholar]
  12. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV 13. pp. 184–199. [Google Scholar]
  13. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  14. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  15. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  16. Wang, X.; Yu, K.; Dong, C.; Loy, C.C. Recovering realistic texture in image super-resolution by deep spatial feature transform. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 606–615. [Google Scholar]
  17. Liu, J.; Zhang, W.; Tang, Y.; Tang, J.; Wu, G. Residual feature aggregation network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2359–2368. [Google Scholar]
  18. Zhang, K.; Gool, L.V.; Timofte, R. Deep unfolding network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3217–3226. [Google Scholar]
  19. Jiang, K.; Wang, Z.; Yi, P.; Jiang, J. Hierarchical dense recursive network for image super-resolution. Pattern Recognit. 2020, 107, 107475. [Google Scholar] [CrossRef]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  21. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning texture transformer network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5791–5800. [Google Scholar]
  22. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  23. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Xiamen, China, 13–15 October 2023; pp. 22367–22377. [Google Scholar]
  24. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  25. Sun, L.; Dong, J.; Tang, J.; Pan, J. Spatially-Adaptive Feature Modulation for Efficient Image Super-Resolution. arXiv 2023, arXiv:2302.13800. [Google Scholar]
  26. Wei, P.; Sun, Y.; Guo, X.; Liu, C.; Li, G.; Chen, J.; Ji, X.; Lin, L. Towards Real-World Burst Image Super-Resolution: Benchmark and Method. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 13233–13242. [Google Scholar]
  27. Wronski, B.; Garcia-Dorado, I.; Ernst, M.; Kelly, D.; Krainin, M.; Liang, C.-K.; Levoy, M.; Milanfar, P. Handheld multi-frame super-resolution. ACM Trans. Graph. (ToG) 2019, 38, 28. [Google Scholar] [CrossRef]
  28. Salvetti, F.; Mazzia, V.; Khaliq, A.; Chiaberge, M. Multi-image super resolution of remotely sensed images using residual attention deep neural networks. Remote Sens. 2020, 12, 2207. [Google Scholar] [CrossRef]
  29. Valsesia, D.; Magli, E. Permutation invariance and uncertainty in multitemporal image super-resolution. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–12. [Google Scholar] [CrossRef]
  30. Märtens, M.; Izzo, D.; Krzic, A.; Cox, D. Super-resolution of PROBA-V images using convolutional neural networks. Astrodynamics 2019, 3, 387–402. [Google Scholar] [CrossRef]
  31. Bhat, G.; Danelljan, M.; Timofte, R.; Cao, Y.; Cao, Y.; Chen, M.; Chen, X.; Cheng, S.; Dudhane, A.; Fan, H. NTIRE 2022 burst super-resolution challenge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 1041–1061. [Google Scholar]
  32. Wang, L.; Guo, Y.; Liu, L.; Lin, Z.; Deng, X.; An, W. Deep video super-resolution using HR optical flow estimation. IEEE Trans. Image Process. 2020, 29, 4323–4336. [Google Scholar] [CrossRef]
  33. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable convnets v2: More deformable, better results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 9308–9316. [Google Scholar]
  34. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual dense network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Munich, Germany, 8–14 September 2018; pp. 2472–2481. [Google Scholar]
  35. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  36. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  37. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. pp. 191–207. [Google Scholar]
  38. Tsai, R.Y.; Huang, T.S. Multiframe image restoration and registration. Multiframe Image Restor. Regist. 1984, 1, 317–339. [Google Scholar]
  39. Farsiu, S.; Robinson, M.D.; Elad, M.; Milanfar, P. Fast and robust multiframe super resolution. IEEE Trans. Image Process. 2004, 13, 1327–1344. [Google Scholar] [CrossRef]
  40. Deudon, M.; Kalaitzis, A.; Goytom, I.; Arefin, M.R.; Lin, Z.; Sankaran, K.; Michalski, V.; Kahou, S.E.; Cornebise, J.; Bengio, Y. Highres-net: Recursive fusion for multi-frame super-resolution of satellite imagery. arXiv 2020, arXiv:2002.06460. [Google Scholar]
  41. Bhat, G.; Danelljan, M.; Yu, F.; Van Gool, L.; Timofte, R. Deep reparametrization of multi-frame super-resolution and denoising. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 2460–2470. [Google Scholar]
  42. Dudhane, A.; Zamir, S.W.; Khan, S.; Khan, F.S.; Yang, M.-H. Burst image restoration and enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5759–5768. [Google Scholar]
  43. Luo, Z.; Li, Y.; Cheng, S.; Yu, L.; Wu, Q.; Wen, Z.; Fan, H.; Sun, J.; Liu, S. BSRT: Improving burst super-resolution with swin transformer and flow-guided deformable alignment. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 998–1008. [Google Scholar]
  44. Wu, R.; Zhang, Z.; Zhang, S.; Zhang, H.; Zuo, W. Rbsr: Efficient and flexible recurrent network for burst super-resolution. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Xiamen, China, 13–15 October 2023; pp. 65–78. [Google Scholar]
  45. Child, R.; Gray, S.; Radford, A.; Sutskever, I. Generating long sequences with sparse transformers. arXiv 2019, arXiv:1904.10509. [Google Scholar]
  46. Wei, C.; Duke, B.; Jiang, R.; Aarabi, P.; Taylor, G.W.; Shkurti, F. Sparsifiner: Learning Sparse Instance-Dependent Attention for Efficient Vision Transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Xiamen, China, 13–15 October 2023; pp. 22680–22689. [Google Scholar]
  47. Chen, X.; Li, H.; Li, M.; Pan, J. Learning A Sparse Transformer Network for Effective Image Deraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Xiamen, China, 13–15 October 2023; pp. 5896–5905. [Google Scholar]
  48. Zhou, J.; Wang, P.; Wang, F.; Liu, Q.; Li, H.; Jin, R. Elsa: Enhanced local self-attention for vision transformer. arXiv 2021, arXiv:2112.12786. [Google Scholar]
  49. Chen, C.-F.; Panda, R.; Fan, Q. Regionvit: Regional-to-local attention for vision transformers. arXiv 2021, arXiv:2106.02689. [Google Scholar]
  50. Hassani, A.; Walton, S.; Li, J.; Li, S.; Shi, H. Neighborhood attention transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Xiamen, China, 13–15 October 2023; pp. 6185–6194. [Google Scholar]
  51. Wu, H.; Xiao, B.; Codella, N.; Liu, M.; Dai, X.; Yuan, L.; Zhang, L. Cvt: Introducing convolutions to vision transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 22–31. [Google Scholar]
  52. Kwon, W.; Kim, S.; Mahoney, M.W.; Hassoun, J.; Keutzer, K.; Gholami, A. A fast post-training pruning framework for transformers. Adv. Neural Inf. Process. Syst. 2022, 35, 24101–24116. [Google Scholar]
  53. Li, Z.; Gu, Q. I-ViT: Integer-only quantization for efficient vision transformer inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 17065–17075. [Google Scholar]
  54. Zhou, Y.; Li, Z.; Guo, C.-L.; Bai, S.; Cheng, M.-M.; Hou, Q. SRFormer: Permuted Self-Attention for Single Image Super-Resolution. arXiv 2023, arXiv:2303.09735. [Google Scholar]
  55. Wang, X.; Chan, K.C.; Yu, K.; Dong, C.; Change Loy, C. Edvr: Video restoration with enhanced deformable convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
Figure 1. The overall architecture of BESR.
Figure 1. The overall architecture of BESR.
Sensors 24 02052 g001
Figure 2. The structure of ECTB.
Figure 2. The structure of ECTB.
Sensors 24 02052 g002
Figure 3. The structure of Geformer.
Figure 3. The structure of Geformer.
Sensors 24 02052 g003
Figure 4. The illustration of TGA.
Figure 4. The illustration of TGA.
Sensors 24 02052 g004
Figure 5. The illustration of SGMN.
Figure 5. The illustration of SGMN.
Sensors 24 02052 g005
Figure 6. The structure of OPA and Ref Enh.
Figure 6. The structure of OPA and Ref Enh.
Sensors 24 02052 g006
Figure 7. The structure of HFF.
Figure 7. The structure of HFF.
Sensors 24 02052 g007
Figure 8. Visual comparison of img_0012 on the synthetic dataset.
Figure 8. Visual comparison of img_0012 on the synthetic dataset.
Sensors 24 02052 g008
Figure 9. Visual comparison of img_0060 on the synthetic dataset.
Figure 9. Visual comparison of img_0060 on the synthetic dataset.
Sensors 24 02052 g009
Figure 10. Visual comparison of img_0116 on the synthetic dataset.
Figure 10. Visual comparison of img_0116 on the synthetic dataset.
Sensors 24 02052 g010
Figure 11. Visual comparison of img_0010_0013 on the real-world dataset.
Figure 11. Visual comparison of img_0010_0013 on the real-world dataset.
Sensors 24 02052 g011
Figure 12. Visual comparison of img_0060_0015 on the real-world dataset.
Figure 12. Visual comparison of img_0060_0015 on the real-world dataset.
Sensors 24 02052 g012
Figure 13. Visual comparison of img_0077_0012 on the real-world dataset.
Figure 13. Visual comparison of img_0077_0012 on the real-world dataset.
Sensors 24 02052 g013
Table 1. Quantitative comparison with state-of-the-art MFSR methods on benchmark datasets. Red and blue represent the best and second-best reconstruction performance, respectively.
Table 1. Quantitative comparison with state-of-the-art MFSR methods on benchmark datasets. Red and blue represent the best and second-best reconstruction performance, respectively.
MethodsParameters (M)Time (ms)SyntheticReal-World
PSNR SSIMLPIPSPSNR SSIMLPIPS
SingleImage [11]13.014036.860.9190.11346.600.9790.039
HighResNet [40]34.7846.337.450.9240.10646.640.9800.038
DBSR [11]13.0143139.170.9460.08147.700.9840.029
MFIR [41]12.1342041.550.9640.04548.320.9850.023
BIPNet [42]6.6713041.930.9670.03548.490.9850.023
BSRT-S [43]4.9219842.720.9710.03148.480.9850.021
RBSR [44]6.4233642.440.9700.03548.800.9870.022
BESR (Ours)3.816142.790.9710.03148.860.9870.022
Table 2. The ablation study of the enhanced CNN-Transformer block (ECTB).
Table 2. The ablation study of the enhanced CNN-Transformer block (ECTB).
ComponentsBaseline
TGA
SGMN
Local Enh
BenchmarkMetrics (PSNR/SSIM)
Synthetic42.32/0.96842.64/0.96942.71/0.97042.47/0.96842.58/0.96942.79/0.971
Real-world48.43/0.98548.67/0.98648.78/0.98648.52/0.98648.64/0.98648.86/0.987
Table 3. Comparison of different core blocks of the network.
Table 3. Comparison of different core blocks of the network.
BlockParameters (M)SyntheticReal-World
PSNR/SSIMPSNR/SSIM
ECTB42.79/0.97148.86/0.9873.81
RSTB [22]42.75/0.97148.81/0.9866.09
RHAG [23]42.76/0.97148.83/0.9877.33
PAB [54]42.69/0.97048.77/0.9865.78
Table 4. The ablation study of the optimized pyramid alignment (OPA) module.
Table 4. The ablation study of the optimized pyramid alignment (OPA) module.
ComponentsBaseline
ECTB L
ECTB R
Ref Enh
BenchmarkMetrics (PSNR/SSIM)
Synthetic41.66/0.96441.78/0.96642.32/0.96842.23/0.96742.64/0.97042.79/0.971
Real-world47.93/0.98348.08/0.98448.47/0.98648.35/0.98548.77/0.98748.86/0.987
Table 5. The ablation study of the hybrid feature fusion (HFF) module.
Table 5. The ablation study of the hybrid feature fusion (HFF) module.
ComponentsBaseline
NECTB010122
Ref
BenchmarkMetrics (PSNR/SSIM)
Synthetic42.11/0.96642.34/0.97042.18/0.96742.40/0.96942.67/0.97042.79/0.971
Real-world48.21/0.98548.43/0.98648.25/0.98548.47/0.98648.79/0.98748.86/0.987
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Lv, Q.; Zhang, W.; Zhang, Y.; Tan, Z. Burst-Enhanced Super-Resolution Network (BESR). Sensors 2024, 24, 2052. https://doi.org/10.3390/s24072052

AMA Style

Li J, Lv Q, Zhang W, Zhang Y, Tan Z. Burst-Enhanced Super-Resolution Network (BESR). Sensors. 2024; 24(7):2052. https://doi.org/10.3390/s24072052

Chicago/Turabian Style

Li, Jiaao, Qunbo Lv, Wenjian Zhang, Yu Zhang, and Zheng Tan. 2024. "Burst-Enhanced Super-Resolution Network (BESR)" Sensors 24, no. 7: 2052. https://doi.org/10.3390/s24072052

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop