Next Article in Journal
iCanClean Removes Motion, Muscle, Eye, and Line-Noise Artifacts from Phantom EEG
Previous Article in Journal
Opto-Electronic Hybrid Network Based on Scattering Layers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spatial and Channel Aggregation Network for Lightweight Image Super-Resolution

College of Mechanical Engineering and Automation, Fuzhou University, Fuzhou 350108, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8213; https://doi.org/10.3390/s23198213
Submission received: 29 August 2023 / Revised: 19 September 2023 / Accepted: 27 September 2023 / Published: 1 October 2023
(This article belongs to the Section Sensor Networks)

Abstract

:
Advanced deep learning-based Single Image Super-Resolution (SISR) techniques are designed to restore high-frequency image details and enhance imaging resolution through the use of rapid and lightweight network architectures. Existing SISR methodologies face the challenge of striking a balance between performance and computational costs, which hinders the practical application of SISR methods. In response to this challenge, the present study introduces a lightweight network known as the Spatial and Channel Aggregation Network (SCAN), designed to excel in image super-resolution (SR) tasks. SCAN is the first SISR method to employ large-kernel convolutions combined with feature reduction operations. This design enables the network to focus more on challenging intermediate-level information extraction, leading to improved performance and efficiency of the network. Additionally, an innovative 9 × 9 large kernel convolution was introduced to further expand the receptive field. The proposed SCAN method outperforms state-of-the-art lightweight SISR methods on benchmark datasets with a 0.13 dB improvement in peak signal-to-noise ratio (PSNR) and a 0.0013 increase in structural similarity (SSIM). Moreover, on remote sensing datasets, SCAN achieves a 0.4 dB improvement in PSNR and a 0.0033 increase in SSIM.

1. Introduction

Single Image Super-Resolution (SISR) refers to the technique of restoring Low-Resolution (LR) images to High-Resolution (HR) clear images. SISR not only improves the perceptual quality of images but also helps to enhance the performance of other computer vision tasks such as object detection and image denoising [1,2,3]. As a result, it has attracted wide attention from researchers.
Since the success of the first SISR method based on Convolutional Neural Networks (CNN) [4], many excellent CNN-based SISR methods have emerged [5,6,7,8,9,10,11], many of which are built upon classic architectures such as residual networks [12], dense networks [13], hybrid network [14] and visual attention mechanisms [15,16,17], among others.
Edge computing devices have made remarkable strides [18,19], enabling the deployment of super-resolution algorithms. However, the existing algorithms, due to their high computational complexity, struggle to strike a balance between performance and speed on edge devices. Therefore, there is a pressing need to develop SISR methods that are lightweight and efficient, ensuring their viability for resource-constrained devices.
Accordingly, there have been many outstanding works on lightweight SISR networks [20,21,22,23,24,25,26,27], most of which employ more compact network architectures and utilize ingenious lightweight strategies. These lightweight strategies include the use of group convolutions [28], depth-wise separable convolutions [29], dilated convolutions [30], and cross convolution [31] to replace regular convolutions. In addition, there are also lightweight strategies such as neural architecture search [32,33], structural reparameterization [34], efficient attention mechanisms [35] and so on. Despite the increasing availability of lightweight SISR networks, their performance is often severely com-promised, making it difficult for them to meet the demands of complex practical applications. There remains room for improvement in the field of lightweight SISR.
Recent research suggests that the remarkable performance of ViT is primarily at-tributed to its macro architecture [36,37]. By utilizing advanced training configurations and adopting ViT-inspired architectural enhancements, CNN can achieve performance on par with, or even surpass, that of ViT, especially when employing large kernel convolutions [38,39]. Following that, Yu et al. [40] present MetaFormer as a general architecture abstracted from transformers, as shown in Figure 1, MetaFormer architecture consists of normalization layers, Spatial Mixture (SM) modules, and Channel Mixture (CM) modules. Further investigation by Yu et al. [41] reveals that a pure CNN network with the MetaFormer architecture is more efficient compared to the ViT-based network in image classification task. The Multi-scale Attention Network (MAN) [42] is an exceptional work of SISR that employs the MetaFormer architecture and large kernel convolutions. MAN employ large convolutional kernels in their spatial and channel mixture modules to extract information, resulting in more efficient performance compared to ViT-based SR methods.
Furthermore, Deng et al. [43] found that deep neural networks have a tendency to encode interactions that are either too complex or too simple, rather than interactions with a moderate level of complexity. MAN also suffer from this drawback. A recent work focusing on the Multi-Order Gated Aggregation Network (MogaNet) [44] proposed the use of feature reduction operations, which compel the network to focus more on challenging intermediate-level information, and achieved excellent results in image classification tasks.
Inspired by the Multi-Order Gated Aggregation Network (MogaNet) [44], we introduced the strategy of large-kernel convolutions coupled with feature reduction in the field of super-resolution for the first time and, based on this design, developed the Triple-Scale Spatial Aggregation Attention Module to aggregate multi-scale information. Building upon the MetaFormer architecture, we further proposed a Spatial and Channel Aggregation Block (SCAB) to aggregate multi-order spatial and channel information. Furthermore, an ingenious introduction of a 9 × 9 large kernel convolutional layer is made at the end of the SCAB module to further expand the receptive field.
As shown in Figure 2, the proposed SCAN can aggregate more contextual information compared to the light versions of MAN [42] and Image Restoration Using Swin Transformer (SwinIR) [45].
The main contributions of this work can be summarized as follows:
(1)
The Triple-Scale Spatial Aggregation (TSSA) attention module was innovatively introduced for the first time, enabling the aggregation of triple-scale spatial information.
(2)
The Spatial and Channel Aggregation Block (SCAB) is innovatively introduced for the first time, capable of aggregating both multi-scale spatial and channel information.
(3)
The Spatial and Channel Aggregation Network (SCAN), a lightweight and efficient pure CNN-based SISR network model that combines the advantages of both CNN and Transformer is proposed.
(4)
Quantitative and qualitative evaluations are conducted on benchmark datasets and remote sensing datasets to investigate the proposed SCAN. As shown in Figure 3, the proposed SCAN achieves a good trade-off between model performance and complexity.

2. Related Work

Classical Deep Learning-based SISR models. With the rapid development of deep learning techniques, researchers have been actively exploring and studying deep learning-based SISR methods. Compared to traditional approaches, deep learning-based methods can extract more expressive image features from the dataset and adaptively learn the mapping relationship between low-resolution and high-resolution images. Consequently, they have achieved remarkable breakthroughs in this field.
Dong et al. [4] first introduced the application of deep learning in the field of SISR by proposing the pioneering CNN-based SISR network model called Super-Resolution CNN (SRCNN), which utilizes three convolutional layers to achieve an end-to-end mapping between low-resolution and high-resolution images, resulting in superior performance compared to traditional methods. Since then, numerous out-standing deep learning-based SISR methods have emerged. Kim et al. [5] introduced Residual Network (ResNet) [12] into the field of SISR inspired by the deep convolutional neural network VGG-net, and proposed the Very Deep SR (VDSR) with 20 weight layers. Kim et al. [48] introduced Recurrent Neural Network (RNN) into SISR for the first time and proposed the Deep Recursive Convolutional Network (DRCN) with up to 16 recursive layers, combining residual learning to control parameter count and address the overfitting issues associated with increasing network depth. Tong et al. [8] pioneered the application of Dense Convolutional Network (DenseNet) [13] in SR and proposed the SR DenseNet network (SRDenseNet) model, which effectively fuses low and high-level features through dense skip connections and enhances the reconstruction of image details using deconvolutional layers. Lim et al. [6] introduced the Enhanced Deep SR (EDSR) network, which removes Batch Normalization (BN) layers from the SRResNet [49] network and incorporates techniques like residual scaling. Their work demonstrated a remarkable enhancement in the quality of image reconstruction. Addressing the oversight of high-frequency information recovery in existing deep learning-based SR methods, Wu et al. [14] introduced a new convolutional block, known as the Spatial-Frequency Hybrid Convolution Block (SFBlock). This block is engineered to extract features from both the spatial and frequency domains. It enhances the capturing of high-frequency details, while simultaneously preserving low-frequency information.
The attention mechanism [50], by assigning different weights to image features based on their importance, enables the network to prioritize crucial information with higher weights. As a result, it is widely employed in various visual tasks to enhance performance and focus on important information. Zhang et al. [11] introduced the attention mechanism into SR for the first time and proposed the Residual Channel Attention Network (RCAN), which utilizes channel attention to adaptively recalibrate the features of each channel based on their interdependencies. This approach enhances the network’s representation capability by learning more informative channel features. Subsequently, Dai et al. [16] pointed out that the channel attention mechanism introduced in RCAN only utilizes the first-order statistics of features through global average pooling, which neglects higher-order statistics and hinders the network’s discriminative ability. They proposed a second-order attention mechanism and developed the Second-order Attention Network (SAN), which achieved better performance than RCAN. Wu et al. [17] introduced a novel Feedback Pyramid Attention Network (FPAN) for SISR. By leveraging the unique feedback connection structure proposed in their work, the FPAN is capable of enhancing the representative capacity of the network while facilitating information flow within it. Furthermore, it exhibits a proficiency in capturing long-range spatial context information across multiple scales.
Recently, the transformer [39] has made significant breakthroughs in the field of computer vision. Inspired by this, Chen et al. [51] proposed a pre-trained network model called Image Processing Transformer (IPT) for handling various low-level computer vision tasks such as SR, denoising, and rain removal. The IPT model proves to be effective in performing the desired tasks, outperforming most existing methods across different tasks. Liang et al. [45] introduced Image Restoration Using Swin Transformer (SwinIR), a high-quality image SR model based on the Swin Transformer [30] architecture. By leveraging the window-based attention mechanism of Swin Transformers, SwinIR effectively handles spatial relationships in images and learns the mapping between low-resolution and high-resolution images, resulting in superior image SR performance.
Guo et al. [39] proposed the Large Kernel Attention (LKA) module and developed the Visual Attention Network (VAN), which outperforms state-of-the-art VIT-based networks in multiple visual tasks. Inspired by LKA, Li et al. [42] proposed the Multi-Scale Large Kernel Attention (MLKA) and developed the Multi-Scale Large Kernel Attention Net-work (MAN) for SISR. MAN outperforms SwinIR with improved performance and reduced computational cost, thanks to the integration of MLKA.
Although these classic SISR methods have achieved impressive performance, they often come with a significant computational cost, making it challenging to apply them in resource-constrained scenarios.
Lightweight SISR models. To facilitate the practical application of SISR methods, researchers have also proposed numerous lightweight SISR methods. Dong et al. [20] introduced Fast SRCNN (FSRCNN), which replaces the pre-upsampling model frame-work in SRCNN with a post-upsampling model framework. By utilizing a deconvolution layer at the end of the network for upsampling, FSRCNN effectively addresses the high computational complexity of SRCNN. Shi et al. [52] proposed Efficient Sub-Pixel CNN (ESPCN), which, like FSRCNN, adopts a post-upsampling model framework. However, ESPCN employs sub-pixel convolutional layers for image upsampling, resulting in superior reconstruction performance compared to the FSRCNN network model.
Ahn et al. [21] introduced the Cascading Residual Network (CARN) which enhances network efficiency through the utilization of group convolutions and parameter sharing blocks. Hui et al. [22] also employed the strategy of group convolutions and introduced the Information Distillation block to construct the Information Distillation Network (IDN), which further improves the quality of image reconstruction while enhancing the speed. Building upon IDN, Hui et al. [23] improved the information distillation blocks and designed the Information Multi-Distillation Block (IMDB) for constructing a lightweight Information Multi-Distillation Network (IMDN), which achieves higher efficiency performance compared to IDN. Li et al. [53] proposed the linearly assembled pixel-adaptive regression network (LAPAR), which transforms the problem of learning the LR-to-HR image mapping into a linear regression task based on multiple predefined filter dictionaries. This approach achieves good performance while maintaining speed. Luo et al. [54] introduced the lattice block network (LatticeNet), which utilizes multiple cascaded lattice blocks based on a lattice filter bank, as well as backward feature fusion strategy. Inspired by edge detection methodologies, Liu et al. [31] developed a novel cross convolution approach that enables more effective exploration of the structural information of the features.
Song et al. [26] proposed an efficient residual dense block search algorithm for image SR. It employs a genetic algorithm to search for efficient SR network structures. Huang et al. [32] introduced a lightweight image SR method with a fully differentiable neural architecture search (DLSR), which incorporates a fully differentiable neural architecture search. A key innovation of their work lies in the creation of a cell-level and network-level search space, which enables the discovery of optimal lightweight models. Zhang et al. [25] introduced the reparameterization concept to SISR and proposed edge-oriented convolution block for real-time SR (ECBSR). During training, they utilized a multi-branch module to enhance the model’s performance. During inference, they transformed the multi-branch module into a single-branch structure to improve runtime speed while maintaining performance. Zhu et al. [35] introduced a light-weight SISR network known as Expectation-Maximization Attention SR (EMASRN). The distinctive aspect of their work is the incorporation of a novel high-resolution expectation-maximization attention mechanism, which effectively captures long-range dependencies in high-resolution feature maps.
Although the aforementioned methods achieve high processing speeds, they often struggle to achieve satisfactory image reconstruction quality. In recent lightweight SISR research, there has been a growing trend to combine ViT and LKA to achieve improved reconstruction quality.
Lu et al. [55] proposed the Efficient SR Transformer (ESRT), a hybrid network composed of a CNN backbone and a transformer backbone to address the significant computational cost and high GPU memory consumption of transformers. Sun et al. [56] introduced ShuffleMixer, which employs deep convolutions with large kernel sizes to aggregate spatial information from large regions. Sun et al. [57] introduced the Spatially Adaptive Feature Modulation Network (SAFMN), a CNN SR network based on the ViT architecture, which achieved a balance between performance and model complexity. Wu et al. [58] proposed TCSR, which introduces the neighborhood attention module to achieve more efficient performance than LKA.
While these lightweight methods have achieved a good balance between model complexity and performance, there is still room for improvement.

3. Proposed Method

3.1. Network Architecture

As illustrated in Figure 4, the proposed SCAN consists of the following three components: the Shallow Feature (SF) extraction module, the Deep Feature (DF) extraction module based on cascaded Spatial and Channel Aggregation Groups (SCAG), and the high-quality image reconstruction module.
Shallow feature extraction module (SF): given an input Low-Resolution (LR) image I L R R 3 × H × W , where H and W are the height and width of the LR image. The shallow feature extraction module was applied, denoted by ƒ S F (⋅), which consists of only a single 3 × 3 convolution, to extract the shallow feature F p R C × H × W . The process is expressed as follows:
F p = ƒ S F ( I L R )
Deep feature extraction module (DF): then the shallow features are sent to the deep feature extraction module to extract deeper and more abstract high-level features. The process is expressed as follows:
F n = D W D   C o n v 9 × 9 , d = 4 ( F n 1 ) .
Where F r denotes the deep feature maps. ƒ D F (⋅) denotes the deep feature extraction module, which consists of multiple cascaded SCAGs and a single 9 × 9 depth-wise-dilated convolutional layer with dilation ratios d = 4. More specifically, intermediate features F 1 , F 2 , … F n are extracted step by step, as shown in the following formula:
F i = ƒ S C A G i ( F i 1 ) ,   i = 1 , 2 , , n F n = D W D   C o n v 9 × 9 ,   d = 4 ( F n 1 ) .
where ƒ S C A G i (⋅) denotes the ith SCAG and D W D   C o n v 9 × 9 , d = 4 (⋅) denotes 9 × 9 depth-wise-dilated convolutional layer with dilation ratios d = 4. n denotes the number of SCAG.
Image reconstruction module:  F r and F p are sent to the image reconstruction module to complete the super resolution reconstruction of the image. The process can be described as below:
I S R = ƒ R C ( F p + F r ) + B i c u b i c ( I L R ) .
where ƒ R C denotes the up-sampling module, which consists of a pixel shuffle operation and a single 3 × 3 convolution. B i c u b i c (⋅) denotes bicubic interpolation up-sampling operation. Incorporating interpolation at this juncture serves to enhance network performance and expedite network convergence.

3.2. Spatial and Channel Aggregation Groups (SCAG)

As discussed in the first section, neural networks utilizing Metaformer [40] architectures similar to ViT have emerged with tremendous potential. Flowing to Metaformer, we propose a spatial and channel aggregation module which used a Metaformer-style design.
As shown in Figure 5, SCAG consists of multiple cascaded spatial aggregation and channel aggregation blocks (SCAB) and a single 9 × 9 depth-wise-dilated convolutional layer with dilation ratios d = 4.
Spatial aggregation and channel aggregation block (SCAB): SCAB consists of the following three components: the triple-scale spatial aggregation attention (TSSA) module, the channel aggregation (CA) module, and the layer normalization layer.
Given the input feature X, the whole process of SCAB is formulated as:
N = L N ( X ) X 1 = X + T S S A ( N ) N 1 = L N ( X 1 ) X 2 = X 1 + C A ( N 1 )
where LN (·) are layer normalization and is employed to enhance network training stability, accelerate convergence speed, and improve model generalization performance. TSGA (·) and CA (·) denote the triple-scale gated aggregation attention (TSGA) module and the channel aggregation (CA) module, which will be introduced in the next section.

3.2.1. Triple-Scale Spatial Aggregation (TSSA) Attention Module

As mentioned in the first section, methods based on large kernel convolutions have gradually surpassed ViT-based methods in some computer vision domains. However, there still exists a bottleneck in representation with the following current methods: deep neural networks have a tendency to encode interactions that are either too complex or too simple, rather than interactions with a moderate level of complexity.
In MogaNet [44], the authors innovatively introduced feature decomposition operations to encourage the network to pay more attention to intermediate-level information that is often overlooked by deep neural networks. Inspired by MogaNet [44] and CNN methods that employ large kernel convolutions [39,42], we propose triple-scale spatial aggregation attention module (TSSA) to aggregate triple-scale context information. As shown in Figure 6, the whole process of TSSA is formulated as:
Y = F D ( X ) Z = T S G A ( Y )
where FD (·) indicates a Feature Decomposition (FD) module used to eliminate redundant feature interactions. TSGA (·) is a Triple-Scale Gated Aggregation (TSGA) module used to aggregate triple-scale contextual information.
The Feature Decomposition (FD) module can be formulated as:
Y = C o n v 1 × 1 ( X ) . Z = G E L U ( Y + γ S ( Y G A P ( Y ) ) )
where C o n v 1 × 1 (·) and G A P (·) are 1 × 1 convolutional layer and global average pooling layer which can extract common local texture and complex global shape, separately. Y G A P ( Y ) can increase the impact of mid-level information. γ S R C × 1 denotes a scaling factor initialized as zeros, which can increases spatial feature diversities by re-weighting. GELU indicates gaussian error linear unit [59], a high-performing neural network activation function, used for channel information gathering and redistribution.
Triple-Scale Gated Aggregation (TSGA) module can be formulated as:
Y = D W C o n v 5 × 5 ( X ) Y l , Y m , Y h = S p l i t ( Y ) Y l = D W D C o n v 5 × 5 , d = 2 ( Y l ) Y m = D W D C o n v 7 × 7 , d = 3 ( Y m ) Y h = D W D C o n v 9 × 9 , d = 4 ( Y h ) Y = C o n c a t ( Y l , Y m , Y h ) Z = S i L U ( C o n v 1 × 1 ( X ) S i L U ( C o n v 1 × 1 ( Y ) ) ) .
where D W C o n v 5 × 5 (·) are 5 × 5 depth-wise convolutional layer. Then, by S p l i t (·) the output Y is equally divided into Y l R C l × H W , Y m R C m × H W , Y h R C h × H W along the channel dimension, where C l = C m = C h = C / 3 . To extract triple scale feature, Y l , Y m , Y h are sent to the 5 × 5 depth-wise-dilated convolutional layer with a dilation ratio of d = 2, 7 × 7, the depth-wise-dilated convolutional layer with a dilation ratio of d = 3, 9 × 9, and the depth-wise-dilated convolutional layer with a dilation ratio of d = 4, separately. Then, the output Y l , Y m , Y h are concatenated to form triple-scale contexts by C o n c a t (·). S i L U are sigmoid-weighted linear units [60], which possesses both the gating effects of Sigmoid and the stable training properties. C o n v 1 × 1 is 1 × 1 convolutional layer. is element-wise multiplication. Finally, spatial gates are leveraged to learn more local information.

3.2.2. Channel Aggregation (CA) Module

Metaformer-style architectures, as illustrated in first section, perform Channel Mixing (CM) usually by 2-layer channel MLP or MLP equipped with 3 × 3 depth-wise convolution [61,62,63]. The suboptimal efficiency of traditional MLPs can be attributed to redundant channels. To address this issue, a lightweight channel aggregation module was introduced, which was inspired by MogaNet. As illustrated in Figure 7, the Channel Aggregation (CA) module is formulated as:
Y = G E L U ( D W C o n v 3 × 3 ( C o n v 1 × 1 ( X ) ) ) Z = C o n v 1 × 1 ( M S F R ( Y ) ) .
where C o n v 1 × 1 (·) indicates a 1 × 1 convolutional layer, D W C o n v 3 × 3 (·) indicates 3 × 3 depth-wise convolutional layer, G E L U (·) indicates gaussian error linear unit [62], a high-performing neural network activation function.
M S F R (·) indicates multi-scale feature reallocation module, which is realized through a channel-reducing projection and gaussian error linear unit [62] to aggregate and redistribute channel-specific information.
The Multi-Scale Feature Reallocation (MSFR) module is formulated as:
M S F R ( X ) = X + γ c ( X G E L U ( X W r ) ) .
W r indicates the channel-reducing projection, which is formulated as:
R C × H W R 1 × H W .
The emphasis on mid-level features is enhanced by performing the X G E L U ( X W r ) operation, which subtracts the global channel information. This operation effectively reduces the influence of global channel statistics and enables a stronger focus on mid-level information. γ c indicates the channel-wise scaling factor initialized as zeros, which increases channel feature diversities.

4. Experiments

4.1. Experimental Setup

Datasets and Evaluation Metrics. The training images consist of 2650 images fromFlickr2K [6] and 800 images from DIV2K [64]. We evaluated our models on the following widely used benchmark datasets: Set5 [47], Set14 [65], BSD100 [66], Urban100 [67], and Manga109 [68]. The commonly used data augmentation methods are applied in the training dataset. Specifically, we used a random combination of random rotations of 0°, 90°, 180°, 270° and horizontal flipping for data augmentation. The average Peak Signal-to-Noise Ratio (PSNR) and the Structural Similarity (SSIM) on the luminance (Y) channel are used as the evaluation metrics.
Implementation Details. For a more comprehensive evaluation of the proposed methods, two different versions of SCAN were trained to resolve the lightweight SR tasks under different complexity. 1/6 SCAGs and 5/4 SCABs were stacked, and the channel width was set to 48/60 in the corresponding tiny/light SCAN.
Training Details. The model was trained using the Adam optimizer [69] with β1 = 0.9 and β2 = 0.99. The learning rate was initialized as 5× 10 4 and scheduled by cosine annealing learning during the whole 1× 10 6 training iterations. For the ablation studies, we trained all models in 4× 10 5 iterations. The weight of the exponential moving average (EMA) [70] was set to 0.999. Only the L1 loss was used to optimize the model. The patch size/batch size was set to 192 × 192/64 and 192 × 192/32 in the corresponding tiny/light SCAN.

4.2. Comparison with SCAN-Tiny SR Method

Quantitative comparisons. To evaluate the performance of the proposed SCAN-tiny, a comparison was made with state-of-the-art tiny SR methods with a parameter count of around 200 k, including Bicubic, SRCNN [4], FSRCNN [20], ShuffleMixer-tiny [56], ECBSR [25], LAPAR-B [53], MAN-tiny [42], and SAFMN [57]. Table 1 shows the quantitative comparison on benchmark datasets for the upscale factors of ×2, ×3 and ×4. The number of parameters (Params) and Floating-point Operations (FLOPs) are also provided, calculated on the 1280 × 720 output. Benefiting from its simple and efficient architecture, the proposed SCAN-tiny achieved comparable performance with fewer parameters and memory consumption.
As shown in Table 1, the proposed SCAN surpassed all methods with parameter counts less than 200 k. Specifically, compared to MAN-tiny ×4, the proposed SCAN-tiny ×4 achieves average 0.1 dB PSNR gain and 0.00216 SSIM gain on five benchmark datasets, with parameter and computational complexity almost equivalent to MAN-tiny ×4. Similarly, compared to ShuffleMixer-tiny ×4, the proposed SCAN-tiny ×4 achieved on average a 0.244 dB PSNR gain and a 0.00634 SSIM gain on the five benchmark datasets, with only slightly higher parameter and computational complexity than ShuffleMixer-tiny.
In addition to networks with fewer than 200 k parameters, comparisons were also made with many networks having more than 200 k parameters. Specifically, the proposed SCAN-tiny ×4 uses 53% parameter and 18% computational complexity of LAPAR-B ×4, achieves average 0.168 dB PSNR gain and 0.00368 SSIM gain on five benchmark datasets. Moreover, SAFMN, which ranks in the Top 3 for model complexity in the NTIRE2023 ESR Challenge, is also listed. The proposed SCAN-tiny ×4 demonstrates comparable performance to SAFMN ×4, while utilizing only 69% of the parameters and 68% of the computational complexity.
These results validate that the proposed SCAN achieves superior performance with fewer parameters, significantly enhancing computational efficiency.
Qualitative comparisons. In addition to the quantitative evaluations, a visual comparison is presented between the SCAN-tiny and six state-of-the-art tiny SR methods, including Bicubic, SRCNN [4], FSRCNN [20], ECBSR [25], LAPAR-B [53], and MAN-tiny [42]. Figure 8 presents a visual comparison of state-of-the-art methods on the Urban100 (×4) and Set14 (×4) datasets for upscale factors of ×4. The image within the red box is cropped and zoomed in on.
For img096 from Urban100, none of the six methods compared to the proposed SCAN were able to recover realistically sharp edges, and the resulting images exhibit blurriness and artifacts. The proposed SCAN reconstructed images closely resemble the HR images. Similarly, for barbara from Set14, only the proposed SCAN is capable of restoring authentic and clear images of books. These visual results demonstrate the information extraction capabilities of the proposed SCAN.

4.3. Comparison with Light SR Method

Quantitative comparisons. To assess the performance of the proposed SCAN, a comparison is conducted with state-of-the-art lightweight SR methods with a parameter count of approximately 1000 k, including CARN [21], LatticeNet [54], LAPAR-A [53], IMDN [23], ESRT [55], SwinIR-light [45], TCSR-L [58] and MAN-light [42]. Table 2 shows the quantitative comparison on benchmark datasets for upscale factors of ×2, ×3 and ×4. The number of parameters (Params) and floating-point operations (FLOPs) are provided, calculated on the 1280 × 720 output.
Thanks to its straightforward and efficient architecture, the proposed SCAN model achieves comparable performance while utilizing fewer parameters and consuming less memory. As shown in Table 2, the proposed approach surpasses all methods with a parameter count of around 1000 k.
Specifically, compared with several state-of-the-art transformer-based methods such as ESRT and SwinIR-light. The proposed SCAN ×4 uses 77% floating-point operations of ESRT ×4, while achieving a 0.36 dB PSNR gain and 0.00624 SSIM gain on five benchmark datasets. Moreover, compared to SwinIR-light ×4, the proposed SCAN ×4 achieves an average of 0.204 dB PSNR gain and 0.00324 SSIM gain on five benchmark datasets, with parameters and computational complexity almost equivalent to SwinIR-light ×4.
In addition to transformer-based methods, many CNN-based methods were also compared. Specifically, the proposed SCAN ×4 uses 88% parameter and 56% computational complexity of TCSR-L ×4, achieves an average of 0.096 dB PSNR gain and 0.00172 SSIM gain on the five benchmark datasets. Moreover, compared to MAN-light ×4, the proposed SCAN ×4 achieves an average of 0.085 dB PSNR gain and 0.00016 SSIM gain on the five benchmark datasets, with only slightly higher parameters and computational complexity than MAN-light ×4.
Compared to CNN-based and transformer-based methods, the proposed SCAN achieves superior performance with lower computational complexity, demonstrating the superiority of the SCAN approach.
Qualitative comparisons. In addition to the quantitative evaluations, a visual com parison of the proposed SCAN and six state-of-the-art light SR methods is presented, including Bicubic, LAPAR-A [53], IMDN [23], ESRT [55], SwinIR-light [45] and MAN-light [42]. Figure 9 presents a visual comparison of state-of-the-art methods on the Urban100 (×4) and Set14 (×4) datasets for up-scale factors of ×4. The image within the red box is cropped and zoomed in on.
For img092 from Urban100, the proposed SCAN method can restore images that are nearly indistinguishable from the HR, while other methods lack this capability, whereas other methods fail to achieve this level of quality, resulting in unacceptable reconstructions. Similarly, For BokuHaSitatakaKun from Manga109, compared to the other methods evaluated, none of the models were able to reliably recover a clear and accurate representation of the complex letter M.

4.4. Remote Sensing Image Super-Resolution

In comparison to conventional images, remote sensing images exhibit numerous small targets and complex backgrounds, which place higher demands on image super-resolution algorithms. To validate the effectiveness of the proposed SCAN-light, experiments were conducted on publicly available remote sensing datasets, DIOR [71] and DOTA [72].
Two hundred images were randomly selected from the DIOR dataset, and another two hundred images were chosen from the DOTA dataset for conducting the transfer of learning on the proposed model. Sixty images were extracted from the test datasets of both the DIOR and DOTA datasets to evaluate the performance of the proposed SCAN-light.
Quantitative comparisons. In order to assess the performance of the proposed SCAN on remote sensing datasets, a comparison is conducted with state-of-the-art lightweight SR methods, including Bicubic, RFDN [24], LAPAR-A [53], IMDN [23], SwinIR-light [45], and MAN-light [42]. Table 3 shows the quantitative comparison on benchmark datasets for upscale factors of ×2, ×3 and ×4. The number of parameters (Params) and floating-point operations (FLOPs) are provided, calculated on the 1280 × 720 output.
Thanks to its efficient architecture, the proposed SCAN model achieves comparable performance while consuming less memory. As shown in Table 3, the proposed approach surpasses all methods in PSNR and SSIM.
Specifically, compared with MAN-light. The proposed SCAN ×4 employs parameters and floating-point operations similar to MAN-light model, achieves an average PSNR improvement of 0.4 dB and an SSIM improvement of 0.0013 on remote sensing datasets.
Qualitative comparisons. In addition to the quantitative evaluations, a visual comparison of the proposed SCAN and five state-of-the-art light SR methods are presented, RFDN [24], LAPAR-A [53], IMDN [23], SwinIR-light [45] and MAN-light [42]. Figure 10 presents a visual comparison of state-of-the-art methods on the DIOR and DOTA datasets for up-scale factors of ×4. The images within the red box are cropped and zoomed in on.
For img05916 from DIOR, the proposed SCAN demonstrates the ability to restore clear road surfaces and car contours, a feat that other methods struggle to achieve due to the small size of these regions. Similarly, for imgP0047 from DOTA, the proposed SCAN excels in restoring accurate car contours, while other methods often produce indistinct contours that are difficult to recognize as cars.

4.5. Ablation Studies

In this section, we conduct ablation studies on some of the designs involved in our SCAN.
Study on TSSA. As previously discussed in Section 3, the FD module and TSGA module are utilized to aggregate triple-scale context information. To substantiate the effectiveness of the FD and TSGA modules, experiments were conducted by selectively removing either of the two from the SCAN-tiny model, and subsequently observing the resultant impact on model performance. As shown in Table 4, it is evident that employing any of them lead to an improvement in performance.
For further demonstration, this is depicted in Figure 11, with the feature maps at different stages within the TSSA module of the model being visualized. Notably, following the collaborative deployment of the TSGA and FD modules, a substantial augmentation in feature richness is observed.
Study on TSGA. As previously discussed in Section 3, regarding the extraction of multi-scale information, the input is partitioned into three equal portions, and each of these portions is subsequently fed into three depth-wise-dilated convolutional layers of varying scales. To validate the efficacy of this approach, the performance of the triple-scale approach was compared with single-scale approaches of different sizes. The comparison results are presented in Table 5, revealing that employing a triple-scale approach facilitates the attainment of a balance between performance and computational costs.
Study on Activation Functions. As previously discussed in Section 3, a sigmoid-weighted linear unit (SiLU) [60] is utilized in the gating branch. In order to authenticate the effectiveness of SiLUs, experiments were conducted in which the SiLU was replaced with a ReLU [73], PReLU [74], and GELU [59] in the proposed SCAN-tiny model, subsequently comparing their respective performances. The comparison results are shown in Table 6, revealing that leveraging SiLU enables the attainment of superior performance with minimal computational costs.
The Study on CA. As previously discussed in Section 3, the introduced CA module consists of a 2-layer channel MLP, a 3 × 3 depth-wise convolutional layer, and a multi-scale feature reallocation module. In Table 7, the results of deploying a 3 × 3 depth-wise convolutional layer and an MSFR module on the proposed tiny SCAN are presented. When comparing the base models without a 3 × 3 depth-wise convolutional layer and multi-scale feature reallocation modules, it is evident that the employment of a 3 × 3 depth-wise convolutional layer can enhance performance. Furthermore, it can also be observed that employing MSFR without the use of large convolutional kernels negatively impacts performance. This is attributed to the pooling operation along the channel dimension, which results in the loss of essential information required for effective processing.
For further clarification, this is demonstrated in Figure 12, where feature maps from various stages within the CA module of the model were visualized, revealing a significant increase in feature richness after the collaborative utilization of the DWConv and MSRA modules.
Study on tail in spatial and channel aggregation groups (SCAG) and deep feature (DF) extraction module. As an innovation, a 9 × 9 depth-wise-dilated convolutional layer with a dilation ratio of d = 4 was employed in the tail of the Deep Feature (DF) extraction module and the spatial and channel aggregation groups (SCAG) module for the first time.
In Table 8, a comparison was made between the base models without a tail in the spatial and channel aggregation groups (SCAG) and the Deep Feature (DF) extraction module. It was observed that the inclusion of DWDConv (9 × 9, d = 4) as the tail resulted in a notable improvement in performance.

5. Conclusions

In this paper, a Spatial and Channel Aggregation Network (SCAN) for lightweight SISR was proposed. SCAN incorporates the Triple-Scale Spatial Aggregation Attention module (TSSA) to aggregate spatial information at multiple scales. Additionally, the Channel Aggregation (CA) module is used to aggregate channel information. Feature reduction operations are applied in both the TSSA and CA modules to encourage the network to focus on the mid-level information which is challenging for deep neural networks to aggregate. The core concept of SCAN is the utilization of large-kernel convolutions and feature reduction strategies to extract intermediate features that are challenging to capture in both spatial and channel dimensions. Moreover, an innovative approach is introduced by using 9 × 9 large convolutional kernels at the end of the attention modules for the first time, aiming to further enlarge the receptive field. As a result, the proposed SCAN achieves highly efficient SR performance on both public benchmark datasets and remote sensing datasets. Extensive experiments demonstrate that, compared to state-of-the-art lightweight SISR methods with similar parameters and FLOPs, the proposed SCAN provides a significant improvement of 0.13 dB in PSNR metric on benchmark dataset and an even more impressive enhancement of 0.4 dB in PSNR on remote sensing datasets. The proposed SCAN provides a 0.13 dB improvement in peak signal-to-noise ratio (PSNR) and a 0.0013 increase in structural similarity (SSIM). Moreover, on remote sensing datasets, SCAN achieves a 0.4 dB improvement in PSNR and a 0.0033 increase in SSIM.

Author Contributions

Conceptualization, X.W.; formal analysis, X.W.; methodology, X.W., L.Z. and F.H.; resources, X.W.; software, L.Z.; validation, X.W.; writing—original draft, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The project was supported by the Natural Science Foundation of Fujian Province of China (No. 2023J01130137) and Fuzhou University (GXRC-18066).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The super-resolution algorithm codes are available online at: https://github.com/linzezuo/SCAN (accessed on 31 Augest 2023).

Acknowledgments

The authors would like to express their gratitude to the anonymous reviewers and editors who worked selflessly to improve our manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dai, D.; Wang, Y.; Chen, Y.; Van Gool, L. Is image super-resolution helpful for other vision tasks? In Proceedings of the 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Placid, NY, USA, 7–10 March 2016; pp. 1–9. [Google Scholar]
  2. Bai, Y.; Zhang, Y.; Ding, M.; Ghanem, B. Sod-mtgan: Small object detection via multi-task generative adversarial network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 206–221. [Google Scholar]
  3. Bei, Y.; Damian, A.; Hu, S.; Menon, S.; Ravi, N.; Rudin, C. New techniques for preserving global structure and denoising with low information loss in single-image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 874–881. [Google Scholar]
  4. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
  5. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  6. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  7. Lan, R.; Sun, L.; Liu, Z.; Lu, H.; Su, Z.; Pang, C.; Luo, X. Cascading and enhanced residual networks for accurate single-image super-resolution. IEEE Trans. Cybern. 2020, 51, 115–125. [Google Scholar] [CrossRef]
  8. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image super-resolution using dense skip connections. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4799–4807. [Google Scholar]
  9. Jiang, K.; Wang, Z.; Yi, P.; Jiang, J.J.P.R. Hierarchical dense recursive network for image super-resolution. Pattern Recognit. 2020, 107, 107475. [Google Scholar] [CrossRef]
  10. Wei, P.; Xie, Z.; Lu, H.; Zhan, Z.; Ye, Q.; Zuo, W.; Lin, L. Component divide-and-conquer for real-world image super-resolution. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part VIII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 101–117. [Google Scholar]
  11. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  13. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  14. Wu, Z.; Liu, W.; Li, J.; Xu, C.; Huang, D. SFHN: Spatial-Frequency Domain Hybrid Network for Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2023. early access. [Google Scholar] [CrossRef]
  15. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  16. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.-T.; Zhang, L. Second-order attention network for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  17. Wu, H.; Gui, J.; Zhang, J.; Kwok, J.T.; Wei, Z. Feedback pyramid attention networks for single image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2023. early access. [Google Scholar] [CrossRef]
  18. Imran, A.; Zhu, Q.; Sulaman, M.; Bukhtiar, A.; Xu, M. Electric-Dipole Gated Two Terminal Phototransistor for Charge-Coupled Device. Adv. Opt. Mater. 2023, 2023, 2300910. [Google Scholar] [CrossRef]
  19. Jang, J.-W.; Lee, S.; Kim, D.; Park, H.; Ardestani, A.S.; Choi, Y.; Kim, C.; Kim, Y.; Yu, H.; Abdel-Aziz, H. Sparsity-aware and re-configurable NPU architecture for Samsung flagship mobile SoC. In Proceedings of the 2021 ACM/IEEE 48th Annual International Symposium on Computer Architecture (ISCA), Valencia, Spain, 14–18 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 15–28. [Google Scholar]
  20. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  21. Ahn, N.; Kang, B.; Sohn, K.-A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  22. Hui, Z.; Wang, X.; Gao, X. Fast and accurate single image super-resolution via information distillation network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 723–731. [Google Scholar]
  23. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight image super-resolution with information multi-distillation network. In Proceedings of the 27th ACM International Conference on Multimedia, New York, NY, USA, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  24. Liu, J.; Tang, J.; Wu, G. Residual feature distillation network for lightweight image super-resolution. In Proceedings of the Computer Vision–ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 41–55. [Google Scholar]
  25. Zhang, X.; Zeng, H.; Zhang, L. Edge-oriented convolution block for real-time super resolution on mobile devices. In Proceedings of the 29th ACM International Conference on Multimedia, New York, NY, USA, 20–24 October 2021; pp. 4034–4043. [Google Scholar]
  26. Song, D.; Xu, C.; Jia, X.; Chen, Y.; Xu, C.; Wang, Y. Efficient residual dense block search for image super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12007–12014. [Google Scholar]
  27. Huang, F.; Chen, Y.; Wang, X.; Wang, S.; Wu, X. Spectral Clustering Super-Resolution Imaging Based on Multispectral Camera Array. IEEE Trans. Image Process. 2023, 32, 1257–1271. [Google Scholar] [CrossRef] [PubMed]
  28. Ioannou, Y.; Robertson, D.; Cipolla, R.; Criminisi, A. Deep roots: Improving cnn efficiency with hierarchical filter groups. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1231–1240. [Google Scholar]
  29. Sifre, L.; Mallat, S. Rigid-motion scattering for texture classification. arXiv 2014, arXiv:1403.1687. [Google Scholar]
  30. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  31. Liu, Y.; Jia, Q.; Fan, X.; Wang, S.; Ma, S.; Gao, W. Cross-SRN: Structure-preserving super-resolution network with cross convolution. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 4927–4939. [Google Scholar] [CrossRef]
  32. Huang, H.; Shen, L.; He, C.; Dong, W.; Liu, W. Differentiable Neural Architecture Search for Extremely Lightweight Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2022, 33, 2672–2682. [Google Scholar] [CrossRef]
  33. Chu, X.; Zhang, B.; Ma, H.; Xu, R.; Li, Q. Fast, accurate and lightweight super-resolution with neural architecture search. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 59–64. [Google Scholar]
  34. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. Repvgg: Making vgg-style convnets great again. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, BC, Canada, 11–17 October 2021; pp. 13733–13742. [Google Scholar]
  35. Zhu, X.; Guo, K.; Ren, S.; Hu, B.; Hu, M.; Fang, H. Lightweight image super-resolution with expectation-maximization attention mechanism. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1273–1284. [Google Scholar] [CrossRef]
  36. Tolstikhin, I.O.; Houlsby, N.; Kolesnikov, A.; Beyer, L.; Zhai, X.; Unterthiner, T.; Yung, J.; Steiner, A.; Keysers, D.; Uszkoreit, J. Mlp-mixer: An all-mlp architecture for vision. Adv. Neural Inf. Process. Syst. 2021, 34, 24261–24272. [Google Scholar]
  37. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do vision transformers see like convolutional neural networks? Adv. Neural Inf. Process. Syst. 2021, 34, 12116–12128. [Google Scholar]
  38. Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31 × 31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11963–11975. [Google Scholar]
  39. Guo, M.-H.; Lu, C.-Z.; Liu, Z.-N.; Cheng, M.-M.; Hu, S.-M. Visual attention network. Comput. Vis. Media 2023, 9, 733–752. [Google Scholar] [CrossRef]
  40. Yu, W.; Luo, M.; Zhou, P.; Si, C.; Zhou, Y.; Wang, X.; Feng, J.; Yan, S. Metaformer is actually what you need for vision. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10819–10829. [Google Scholar]
  41. Yu, W.; Si, C.; Zhou, P.; Luo, M.; Zhou, Y.; Feng, J.; Yan, S.; Wang, X. Metaformer baselines for vision. arXiv 2022, arXiv:2210.13452. [Google Scholar]
  42. Wang, L.; Shen, J.; Tang, E.; Zheng, S.; Xu, L. Multi-scale attention network for image super-resolution. J. Vis. Commun. Image Represent. 2021, 80, 103300. [Google Scholar] [CrossRef]
  43. Deng, H.; Ren, Q.; Zhang, H.; Zhang, Q. Discovering and explaining the representation bottleneck of dnns. arXiv 2021, arXiv:2111.06236. [Google Scholar]
  44. Li, S.; Wang, Z.; Liu, Z.; Tan, C.; Lin, H.; Wu, D.; Chen, Z.; Zheng, J.; Li, S. Efficient multi-order gated aggregation network. arXiv 2022, arXiv:2211.03295. [Google Scholar]
  45. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 1833–1844. [Google Scholar]
  46. Gu, J.; Dong, C. Interpreting super-resolution networks with local attribution maps. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Montreal, BC, Canada, 11–17 October 2021; pp. 9199–9208. [Google Scholar]
  47. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-Complexity Single-Image Super-Resolution Based on Nonnegative Neighbor Embedding. In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012. [Google Scholar]
  48. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-recursive convolutional network for image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  49. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  50. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762. [Google Scholar]
  51. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
  52. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  53. Li, W.; Zhou, K.; Qi, L.; Jiang, N.; Lu, J.; Jia, J. Lapar: Linearly-assembled pixel-adaptive regression network for single image super-resolution and beyond. Adv. Neural Inf. Process. Syst. 2020, 33, 20343–20355. [Google Scholar]
  54. Luo, X.; Xie, Y.; Zhang, Y.; Qu, Y.; Li, C.; Fu, Y. Latticenet: Towards lightweight image super-resolution with lattice block. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XXII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 272–289. [Google Scholar]
  55. Lu, Z.; Liu, H.; Li, J.; Zhang, L. Efficient transformer for single image super-resolution. arXiv 2021, arXiv:2108.11084. [Google Scholar]
  56. Sun, L.; Pan, J.; Tang, J. ShuffleMixer: An Efficient ConvNet for Image Super-Resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 17314–17326. [Google Scholar]
  57. Sun, L.; Dong, J.; Tang, J.; Pan, J. Spatially-Adaptive Feature Modulation for Efficient Image Super-Resolution. arXiv 2023, arXiv:2302.13800. [Google Scholar]
  58. Wu, G.; Jiang, J.; Bai, Y.; Liu, X. Incorporating Transformer Designs into Convolutions for Lightweight Image Super-Resolution. arXiv 2023, arXiv:2303.14324. [Google Scholar]
  59. Hendrycks, D.; Gimpel, K. Gaussian error linear units (gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  60. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef]
  61. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  62. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  63. Wang, W.; Xie, E.; Li, X.; Fan, D.-P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pvt v2: Improved baselines with pyramid vision transformer. Comput. Vis. Media 2022, 8, 415–424. [Google Scholar] [CrossRef]
  64. Agustsson, E.; Timofte, R. NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  65. Zeyde, R.; Elad, M.; Protter, M. On Single Image Scale-Up Using Sparse-Representations, Curves and Surfaces. In Proceedings of the 7th International Conference, Avignon, France, 24–30 June 2010. Revised Selected Papers. [Google Scholar]
  66. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001. [Google Scholar]
  67. Huang, J.B.; Singh, A.; Ahuja, N. Single Image Super-Resolution from Transformed Self-Exemplars; IEEE: Piscataway, NJ, USA, 2015. [Google Scholar]
  68. Matsui, Y.; Ito, K.; Aramaki, Y.; Fujimoto, A.; Ogawa, T.; Yamasaki, T.; Aizawa, K. Sketch-based Manga Retrieval using Manga109 Dataset. Multimed. Tools Appl. 2017, 76, 21811–21838. [Google Scholar] [CrossRef]
  69. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  70. Athiwaratkun, B.; Finzi, M.; Izmailov, P.; Wilson, A.G. There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average. arXiv 2018, arXiv:1806.05594. [Google Scholar]
  71. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  72. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  73. Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  74. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1026–1034. [Google Scholar]
Figure 1. MetaFormer architecture.
Figure 1. MetaFormer architecture.
Sensors 23 08213 g001
Figure 2. Comparison of local attribution maps (LAMs) [46] between the proposed SCAN and other efficient lightweight SR models. The LAMs demonstrate the significance of every pixel in the input LR image with respect to the patch marked with a red box’s SR. Additionally, the contribution area is shown in the third row. It is evident that the proposed SCAN can aggregate more information.
Figure 2. Comparison of local attribution maps (LAMs) [46] between the proposed SCAN and other efficient lightweight SR models. The LAMs demonstrate the significance of every pixel in the input LR image with respect to the patch marked with a red box’s SR. Additionally, the contribution area is shown in the third row. It is evident that the proposed SCAN can aggregate more information.
Sensors 23 08213 g002
Figure 3. Model performance and complexity comparison between the proposed SCAN model and other lightweight SISR methods on Set5 [47] for ×4 SR. The circle sizes indicate the Floating-Point Operations (FLOPs). The proposed SCAN achieves a good trade-off between model performance and complexity.
Figure 3. Model performance and complexity comparison between the proposed SCAN model and other lightweight SISR methods on Set5 [47] for ×4 SR. The circle sizes indicate the Floating-Point Operations (FLOPs). The proposed SCAN achieves a good trade-off between model performance and complexity.
Sensors 23 08213 g003
Figure 4. Overview of our spatial and channel aggregation network (SCAN).
Figure 4. Overview of our spatial and channel aggregation network (SCAN).
Sensors 23 08213 g004
Figure 5. Structure of spatial and channel aggregation groups (SCAG).
Figure 5. Structure of spatial and channel aggregation groups (SCAG).
Sensors 23 08213 g005
Figure 6. Structure of Triple-Scale Spatial Aggregation (TSSA) module.
Figure 6. Structure of Triple-Scale Spatial Aggregation (TSSA) module.
Sensors 23 08213 g006
Figure 7. Structure of the Channel Aggregation (CA) module.
Figure 7. Structure of the Channel Aggregation (CA) module.
Sensors 23 08213 g007
Figure 8. Visual comparison of state-of-the-art methods in some challenging cases (×4SR).
Figure 8. Visual comparison of state-of-the-art methods in some challenging cases (×4SR).
Sensors 23 08213 g008
Figure 9. Visual comparison of state-of-the-art methods in some challenging cases(×4SR).
Figure 9. Visual comparison of state-of-the-art methods in some challenging cases(×4SR).
Sensors 23 08213 g009
Figure 10. Visual comparison of state-of-the-art methods in some challenging remote sensing cases(×4SR).
Figure 10. Visual comparison of state-of-the-art methods in some challenging remote sensing cases(×4SR).
Sensors 23 08213 g010
Figure 11. Feature map at the stage of TSSA.
Figure 11. Feature map at the stage of TSSA.
Sensors 23 08213 g011
Figure 12. Feature map at the stage of CA.
Figure 12. Feature map at the stage of CA.
Sensors 23 08213 g012
Table 1. Quantitative comparison with state-of-the-art methods for image SR on benchmark datasets. ’Multi-Adds’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
Table 1. Quantitative comparison with state-of-the-art methods for image SR on benchmark datasets. ’Multi-Adds’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
MethodScaleParams
(K)
Multi-Adds(G)Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
Bicubic×2--33.66/0.929930.24/0.868829.56/0.843126.88/0.840330.80/0.9339
SRCNN [4]575336.66/0.954232.42/0.906331.36/0.887929.50/0.894635.74/0.9661
FSRCNN [20]136.037.00/0.955832.63/0.908831.53/0.892029.88/0.902036.67/0.9694
ShuffleMixer-tiny [56]18025.037.85/0.960033.33/0.915331.99/0.897231.22/0.918338.25/0.9761
LAPAR-B [53]25085.037.87/0.960033.39/0.916232.10/0.898731.62/0.923538.27/0.9764
MAN-tiny [42]1347.737.91/0.960333.47/0.917032.12/0.899131.74/0.924738.62/0.9770
SAFMN [57]22852.038.00/0.960533.54/0.917732.16/0.899531.84/0.925638.71/0.9771
SCAN-tiny (ours)15030.437.96/0.960433.56/0.918432.16/0.899631.89/0.926638.72/0.9772
Bicubic×3--30.39/0.868227.55/0.774227.21/0.738524.46/0.734926.95/0.8556
SRCNN [4]5753.032.75/0.909029.28/0.829028.41/0.786326.24/0.798930.59/0.9107
FSRCNN [20]125.033.16/0.914029.43/0.824228.53/0.791026.43/0.808030.98/0.9212
ShuffleMixer-tiny [56]11412.034.07/0.925030.14/0.838228.94/0.800927.54/0.837333.03/0.9400
LAPAR-B [53]27661.034.20/0.925630.17/0.838729.03/0.803227.85/0.845933.15/0.9417
MAN-tiny [42]1418.034.23/0.925830.25/0.840429.03/0.803927.85/0.846333.32/0.9427
SAFMN [57]23323.034.34/0.926730.33/0.841829.08/0.804827.95/0.847433.52/0.9437
SCAN-tiny (ours)15615.834.34/0.926530.30/0.841429.07/0.804327.95/0.848333.48/0.9436
Bicubic×4--28.42/0.810426.00/0.702725.96/0.667523.14/0.657724.89/0.7866
SRCNN [4]5753.030.48/0.862827.49/0.750326.90/0.710124.52/0.722127.66/0.8505
FSRCNN [20]124.630.71/0.865727.59/0.753526.98/0.715024.62/0.728027.90/0.8517
ECBSR-M16C64 [25]60334.731.87/0.890128.39/0.776827.44/0.731625.63/0.771029.80/0.8986
ShuffleMixer-tiny [56]1138.031.88/0.891228.46/0.777927.45/0.731325.66/0.769029.96/0.9006
LAPAR-B [53]31353.031.94/0.891728.46/0.778427.52/0.733525.84/0.777230.03/0.9025
MAN-tiny [42]1508.432.07/0.893028.53/0.780127.51/0.734525.84/0.778630.18/0.9047
SAFMN [57]24014.032.18/0.894828.60/0.781327.58/0.735925.97/0.780930.43/0.9063
SCAN-tiny (ours)1659.532.18/0.894628.59/0.781627.55/0.735825.97/0.782930.34/0.9068
Table 2. Quantitative comparison with state-of-the-art methods for image SR on benchmark datasets. ‘FLOPs’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
Table 2. Quantitative comparison with state-of-the-art methods for image SR on benchmark datasets. ‘FLOPs’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
MethodScaleParams
(K)
FLOPs
(G)
Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
CARN [21]×21592222.837.76/0.959033.52/0.916632.09/0.897831.92/0.925638.36/0.9765
LatticeNet [54]756169.538.06/0.960733.70/0.918732.20/0.899932.25/0.928839.00/0.9774
LAPAR-A [53]548171.138.01/0.960533.62/0.918332.19/0.899932.10/0.928338.67/0.9772
IMDN [23]694158.838.00/0.960533.63/0.917732.19/0.899632.17/0.928338.88/0.9774
ESRT [55]677208.438.03/0.960033.75/0.918432.25/0.900132.58/0.931839.12/0.9774
SwinIR-light [45]878195.638.14/0.961133.86/0.920632.31/0.901232.76/0.934039.12/0.9783
MAN-light [42]820180.438.18/0.961233.93/0.921332.36/0.902232.92/0.936439.44/0.9786
SCAN (ours)889204.738.23/0.961434.22/0.923332.38/0.902233.05/0.936839.50/0.9788
CARN [21]×31592118.832.29/0.925530.29/0.840729.06/0.803428.06/0.849333.50/0.9440
LatticeNet [54]76576.334.40/0.927230.32/0.841629.10/0.804928.19/0.851333.66/0.9440
LAPAR-A [53]544114.134.36/0.926730.34/0.842129.11/0.805428.15/0.852333.51/0.9441
IMDN [23]70371.534.36/0.927030.32/0.841729.09/0.804628.17/0.851933.61/0.9445
ESRT [55]770102.334.42/0.926830.43/0.843329.15/0.806328.46/0.857433.95/0.9455
SwinIR-light [45]88687.234.62/0.928930.54/0.846329.20/0.808228.66/0.862433.89/0.9464
MAN-light [42]82982.534.65/0.929230.60/0.847629.29/0.810128.87/0.867134.40/0.9434
SCAN (ours)89791.134.81/0.930230.65/0.848629.31/0.810728.80/0.869134.62/0.9505
CARN [21]×4159290.932.13/0.893728.60/0.780627.58/0.734926.07/0.783730.42/0.9070
LatticeNet [54]77743.632.18/0.894328.61/0.781227.57/0.735526.14/0.784430.46/0.9601
LAPAR-A [53]65994.032.15/0.894428.61/0.781827.61/0.736626.14/0.787130.42/0.9074
IMDN [23]71540.932.21/0.894828.58/0.781127.56/0.735326.04/0.783830.45/0.9075
ESRT [55]77767.732.19/0.894728.69/0.783327.69/0.737926.39/0.796230.75/0.9100
SwinIR-light [45]89749.632.44/0.897628.77/0.785827.69/0.740626.47/0.798030.92/0.9151
TCSR-L [58]103093.032.55/0.899228.89/0.788627.75/0.742326.67/0.803931.17/0.9107
MAN-light [42]84047.132.50/0.898828.87/0.788527.77/0.742926.70/0.805231.25/0.9170
SCAN (ours)90951.832.63/0.900128.91/0.789027.80/0.742926.79/0.806531.38/0.9148
Table 3. Quantitative comparison with state-of-the-art methods for light image SR on remote sensing datasets. ‘FLOPs’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
Table 3. Quantitative comparison with state-of-the-art methods for light image SR on remote sensing datasets. ‘FLOPs’ is calculated with a 1280 × 720 GT image. The best and second-best performances are in red and blue, respectively.
MethodScaleParams
(K)
FLOPs
(G)
Dior
PSNR/SSIM
Dota
PSNR/SSIM
RFDN [24]×455523.925.95/0.643024.79/0.5975
SwinIR-light [45]84047.127.50/0.678626.00/0.7400
IMDN [23]70371.527.52/0.677826.04/0.7381
LAPAR-A [53]65994.027.55/0.679926.01/0.7402
MAN-light [42]84047.127.59/0.680526.19/0.7453
SCAN (ours)90951.828.29/0.703726.30/0.7486
Table 4. Ablation studies on components of TSSA. The impact of FD and TSGA modules are shown upon SCAN-tiny on the ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Table 4. Ablation studies on components of TSSA. The impact of FD and TSGA modules are shown upon SCAN-tiny on the ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
ModuleParams
(K)
FLOPs
(G)
Set5Set14BSD100Urban100Manga109
FDTSGAPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
××99.275.6330.79/0.865427.62/0.758426.91/0.717824.67/0.735928.12/0.8610
×153.438.7532.01/0.892228.48/0.778627.48/0.733225.75/0.775530.05/0.9032
165.439.4532.13/0.893728.53/0.780027.52/0.734425.87/0.779030.17/0.9046
Table 5. Ablation studies on design of TSGA module. The impact of multi-scale approach, and single-scale approach are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics is highlighted in bold for emphasis.
Table 5. Ablation studies on design of TSGA module. The impact of multi-scale approach, and single-scale approach are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics is highlighted in bold for emphasis.
Convolution TypeParams
(K)
FLOPs
(G)
Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
D W D C o n v 5 × 5 , d = 2 159.039.0832.08/0.893128.49/0.779227.52/0.733925.82/0.777430.15/0.9043
D W D C o n v 7 × 7 , d = 3 164.799.4132.11/0.893328.51/0.779527.53/0.734425.86/0.778630.21/0.9050
D W D C o n v 9 × 9 , d = 4 172.479.8532.10/0.893628.52/0.779827.52/0.734625.88/0.779730.18/0.9048
Triple-scale165.439.4532.13/0.893728.53/0.780027.52/0.734425.87/0.779030.17/0.9046
Table 6. Ablation studies on different type of activation function. The impact of SiLU, ReLU, PReLU and GELU are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Table 6. Ablation studies on different type of activation function. The impact of SiLU, ReLU, PReLU and GELU are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Activation Function TypeParams
(K)
FLOPs
(G)
Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
ReLU165.439.4732.06/0.779628.52/0.779627.51/0.733925.84/0.777930.14/0.9041
PReLU165.439.4732.09/0.893228.54/0.779827.51/0.734125.86/0.778830.15/0.9045
GELU165.439.4532.09/0.893328.53/0.779627.51/0.734125.87/0.778630.17/0.9048
SiLU 165.439.4532.13/0.893728.53/0.780027.52/0.734425.87/0.779030.17/0.9046
Table 7. Ablation studies on components of CA. The impact of D W C o n v 3 × 3 and MSFR mechanism, are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Table 7. Ablation studies on components of CA. The impact of D W C o n v 3 × 3 and MSFR mechanism, are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
ModuleParams
(K)
FLOPs
(G)
Set5Set14BSD100Urban100Manga109
D W C o n v 3 × 3 MSFRPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
××160.159.1432.03/0.892628.47/0.778527.49/0.733125.76/0.775630.03/0.9025
×160.639.1731.35/0.876727.89/0.759727.11/0.710924.97/0.740028.46/0.8563
×164.469.4232.08/0.893328.52/0.779727.51/0.734225.86/0.778930.17/0.9045
165.439.4532.13/0.893728.53/0.780027.52/0.734425.87/0.779030.17/0.9046
Table 8. Ablation studies on different type of tail. The impact of with no tail, D W D C o n v 5 × 5 , d = 2 , D W D C o n v 7 × 7 , d = 3 and D W D C o n v 9 × 9 , d = 4 are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Table 8. Ablation studies on different type of tail. The impact of with no tail, D W D C o n v 5 × 5 , d = 2 , D W D C o n v 7 × 7 , d = 3 and D W D C o n v 9 × 9 , d = 4 are shown upon SCAN-tiny on ×4 SR task. ’FLOPs’ is calculated with a 1280 × 720 GT image. The best metrics are highlighted in bold for emphasis.
Tail TypeParams
(K)
FLOPs
(G)
Set5
PSNR/SSIM
Set14
PSNR/SSIM
BSD100
PSNR/SSIM
Urban100
PSNR/SSIM
Manga109
PSNR/SSIM
None141.818.0931.96/0.891528.44/0.778327.48/0.732925.72/0.774129.99/0.9020
D W D C o n v 5 × 5 , d = 2 149.308.5232.09/0.893628.52/0.779627.52/0.734325.88/0.779230.20/0.9050
D W D C o n v 7 × 7 , d = 3 156.218.9132.08/0.779628.51/0.779627.52/0.734425.85/0.778630.18/0.9048
D W D C o n v 9 × 9 , d = 4 165.439.4532.13/0.893728.53/0.780027.52/0.734425.87/0.779030.17/0.9046
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, X.; Zuo, L.; Huang, F. Spatial and Channel Aggregation Network for Lightweight Image Super-Resolution. Sensors 2023, 23, 8213. https://doi.org/10.3390/s23198213

AMA Style

Wu X, Zuo L, Huang F. Spatial and Channel Aggregation Network for Lightweight Image Super-Resolution. Sensors. 2023; 23(19):8213. https://doi.org/10.3390/s23198213

Chicago/Turabian Style

Wu, Xianyu, Linze Zuo, and Feng Huang. 2023. "Spatial and Channel Aggregation Network for Lightweight Image Super-Resolution" Sensors 23, no. 19: 8213. https://doi.org/10.3390/s23198213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop