Next Article in Journal
A High Gain Circularly Polarized Slot Antenna Array for 5G Millimeter-Wave Applications
Previous Article in Journal
Impedance Spectroscopy Study of Charge Transfer in the Bulk and Across the Interface in Networked SnO2/Ga2O3 Core–Shell Nanobelts in Ambient Air
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dual-Path Large Kernel Learning and Its Applications in Single-Image Super-Resolution †

1
School of Information and Control Engineering, China University of Mining and Technology, Xuzhou 221116, China
2
Key Laboratory of System Control and Information Processing, Ministry of Education, Shanghai 200240, China
3
School of Computer Science, China University of Mining and Technology, Xuzhou 221116, China
*
Authors to whom correspondence should be addressed.
This is an expanded research article based on the conference paper “Single Image Super Resolution Based on Dual-Path Large Kernel Learning” that was presented at 2nd International Conference on Internet of Things, Copenhagen, Denmark, 19–22 August 2024.
Sensors 2024, 24(19), 6174; https://doi.org/10.3390/s24196174
Submission received: 21 July 2024 / Revised: 11 September 2024 / Accepted: 19 September 2024 / Published: 24 September 2024
(This article belongs to the Section Sensor Networks)

Abstract

:
To enhance the performance of super-resolution models, neural networks frequently employ module stacking. However, this approach inevitably results in an excessive proliferation of parameter counts and information redundancy, ultimately constraining the deployment of these models on mobile devices. To surmount this limitation, this study introduces the application of Dual-path Large Kernel Learning (DLKL) to the task of image super-resolution. Within the DLKL framework, we harness a multiscale large kernel decomposition technique to efficiently establish long-range dependencies among pixels. This network not only maintains excellent performance but also significantly mitigates the parameter burden, achieving an optimal balance between network performance and efficiency. When compared with other prevalent algorithms, DLKL exhibits remarkable proficiency in generating images with sharper textures and structures that are more akin to natural ones. It is particularly noteworthy that on the challenging texture dataset Urban100, the network proposed in this study achieved a significant improvement in Peak Signal-to-Noise Ratio (PSNR) for the ×4 upscaling task, with an increase of 0.32 dB and 0.19 dB compared with the state-of-the-art HAFRN and MICU networks, respectively. This remarkable result not only validates the effectiveness of the present model in complex image super-resolution tasks but also highlights its superior performance and unique advantages in the field.

1. Introduction

Single-Image Super-Resolution (SISR) technology aims to finely recover high-resolution (HR) images with rich details from low-resolution (LR) images. This technology plays an indispensable role in various fields, such as medical image retrieval [1,2], computational photography [3,4,5,6,7], and smart mines [8]. Currently, mainstream SISR algorithms can be categorized into three types: interpolation-based methods, reconstruction-based methods, and learning-based methods. Although interpolation-based methods are favored for their fast execution, the reconstructed images often suffer from artifacts such as blur artifact, leading to unsatisfactory image quality. On the other hand, reconstruction-based methods constrain the solution space by introducing prior knowledge, which can improve the reconstruction quality to a certain extent. However, they often come with complex optimization processes and high computational costs, limiting their practical application to some extent.
In recent years, convolutional neural networks have achieved remarkable breakthroughs and advancements in the field of super-resolution technology. Among these advancements, the introduction of the SRCNN [9] stands out for its remarkable reconstruction performance, despite its succinct structure comprising merely three convolutional layers. This approach has served as the foundation for the classical three-stage framework that subsequent super-resolution models have followed: shallow feature extraction, nonlinear feature mapping, and image reconstruction. To alleviate the computational complexity of SRCNN, researchers introduced the pixel shuffle strategy and ingeniously applied it to the image reconstruction stage, thereby significantly enhancing the computational efficiency of the neural network. Subsequently, to further improve the reconstruction quality, more extensive and deeper super-resolution networks such as VDSR [10], EDSR-BASE [11], and DRCN [12] have emerged. Although there is a positive correlation between network depth and performance, deep networks often come with high training costs. Therefore, attention-based SISR networks gradually emerge, which can intelligently focus on significant information while ignoring minor details, thus achieving more accurate reconstruction. Models like RCAN [13], PAN [14], and IPT [15] are typical representatives of this approach, integrating channel attention, pixel attention, and self-attention mechanisms, respectively. However, existing attention mechanisms still face challenges in integrating local and global dependencies, and a single attention mechanism often struggles to capture diverse information. To address this issue, researchers propose methods that combine multiple attention mechanisms, such as MFRN [16], combining channel attention and pixel attention, while CBAM [17] integrates channel attention and spatial attention. Although these methods improve performance to some extent, they also significantly increase computational demands, making practical applications on resource-constrained mobile devices difficult. Therefore, achieving excellent performance while maintaining lightweight characteristics becomes a hot research topic. In recent years, researchers actively explore lightweight architectures such as LMAN [18], MSFIN [19], and MADNet [20]. Although these architectures reduce model parameters to some extent, achieving further reductions in computational complexity while maintaining performance, especially in preserving complex details, remains a challenging task.
To tackle this challenge, VAN [21] creatively proposes a novel approach called large kernel attention (LKA). This method forms a unique and efficient structural system by ingeniously combining deep convolution, deep dilated convolution, and pointwise convolution. Although LKA performs well in terms of computational efficiency, it has limitations in capturing and utilizing multivariate information due to its reliance on fixed kernel sizes. To address these limitations, this study further combines the multiscale mechanism with LKA, innovatively proposing the Dual-path Large Kernel Learning (DLKL) method. DLKL not only captures long-range dependencies in images but also effectively preserves local detailed information, enabling fine adjustments of image textures. Additionally, this method significantly enhances the acquisition of spatial diversity information while effectively suppressing redundant data, thereby improving processing efficiency. Experimental results demonstrate that DLKL outperforms mainstream algorithms in both objective metrics and visual perception. Importantly, compared with other lightweight algorithms, DLKL achieves a significant reduction in parameter count while maintaining excellent performance, thus achieving a perfect balance between model performance and efficiency. In summary, the contributions of this study mainly lie in the following three aspects:
  • An architecture called DLKL has been proposed.
  • A dual-path large kernel attention mechanism is designed based on DLKL.
  • A lightweight SISR model employing the DLKL method is introduced.

2. Proposed Method

Figure 1 provides a detailed illustration of a SISR system based on Dual-path Large Kernel Learning (DLKL). The system consists of four components: shallow feature extraction (SFE), nonlinear feature mapping (NFM), feature reconstruction (FR), and coarse feature reconstruction (CFR). It should be clear that each module in the system plays an indispensable role. Specifically, the SFE module, as the feature coarse extraction module, provides the necessary foundation for the subsequent fine texture extraction. The NFM module, i.e., the nonlinear mapping module, is responsible for utilizing the linear and nonlinear features to achieve the accurate fitting of the fine texture. And the FR and CFR modules, as pixel shuffling modules, not only achieve image resolution enhancement but also undertake the recovery of fine-grained and coarse-grained image details.

2.1. Shallow Feature Extraction (SFE)

Assuming the input low-resolution image is I L R , this image will be processed through the shallow feature extraction function H S F E to obtain the shallow feature F S . This process can be described by Equation (1).
F S = H S F E ( I L R )
where I L R represents the low-resolution image, H c o n v 3 ( ) denotes the convolution operation with a 3 × 3 kernel, and F S represents the extracted shallow features.
Due to the limitations of 1 × 1 convolutions in capturing spatial dependencies, they often result in significant loss of low-frequency information in applications. On the other hand, while 5 × 5 convolutions have a larger receptive field, they may also introduce a significant amount of task-irrelevant information, thereby increasing the complexity of the network. In contrast, 3 × 3 convolutions maintain spatial correlations while reducing the number of parameters, thus achieving network lightweighting. Therefore, in the shallow feature extraction function H S F E , a solitary 3 × 3 convolutional layer is exclusively employed. For a given tensor X , the process can be described using Equation (2), and C 3 × 3 denotes a 3 × 3 convolutional kernel.
H S F E ( X ) = C 3 × 3 ( X )

2.2. Nonlinear Feature Mapping (NFM)

Subsequently, the extracted shallow features are passed to a deep feature extraction network formed by cascading multiple MAFE modules. This processing is described by Equation (3), where H M A F E 6 ( · ) represents the continuous propagation function of six MAFE modules, and F D refers to the deep features obtained from the deep feature extraction network.
F D = H M A F E 6 ( I L R ) I L R

2.2.1. MAFE

MAFE stands for Multiscale Attention Feature Extraction. In order to achieve good super-resolution performance with a low number of parameters, this paper designs a lightweight SISR network. In this network, the article innovatively combines large kernel decomposition, multiscale learning, and channel attention enhancement to address the challenges in super-resolution tasks. Specifically, MAFE includes the following three core aspects: firstly, it can establish interdependence between features in the spatial dimension, thereby enhancing the correlation between features; secondly, MAFE can capture feature information at different scales, improving the network’s ability to process the complex image content while maintaining efficiency; finally, by introducing channel attention mechanism, MAFE significantly elevates the expression capability of features in each channel, thereby allowing the network to concentrate more intently on the pivotal features that are indispensable for super-resolution reconstruction.
MAFE consists of multiple convolutional layers, two Gaussian Error Linear Unit (GELU) activation functions, the LKA mechanism, and the Enhanced Channel Attention (ECA) module [22]. Given the input tensor X , the processing flow of MAFE can be described by Equations (4) and (5). In these equations, Y and Z represent the generated features, δ ( · ) denotes the GELU activation function used to introduce nonlinearity to the network, c a t ( · ) is a versatile feature concatenation function that effectively integrates features from various sources to enhance model performance, and H L K A and H E C A are the propagation functions of the LKA and ECA mechanisms, respectively, used to capture long-range dependencies and enhance the discriminative power of channel features, while C 1 × 1 is a 1 × 1 convolutional kernel used to adjust the dimensionality of features.
Y = C 1 × 1 ( c a t [ H L K A 1 ( δ ( C 3 × 3 ( X ) ) ) , H L K A 2 ( δ ( C 3 × 3 ( X ) ) ) ] ) ) X
Z = C 3 × 3 ( H E C A ( δ ( C 3 × 3 ( Y ) ) ) ) Y

2.2.2. LKA

Given the input feature X , the LKA mechanism ingeniously decomposes the convolution into three key components, thus establishing long-range dependencies in the image. As shown in Figure 2, these components include a ( 2 d 1 ) × ( 2 d 1 ) depth convolution H D C ( · ) , a K / d × K / d depth-dilated convolution H D D C ( · ) , and a point-wise convolution C 1 × 1 ( · ) . The input feature first passes through the depth convolution layer, then through the depth dilated convolution layer, and finally through the point-wise convolution layer, resulting in the generation of a precise attention map. This attention map is subsequently element-wise multiplied with the original feature map, highlighting the crucial feature regions for super-resolution reconstruction, and generating a new enhanced feature map. The entire process is depicted by Equation (6), where ⊗ denotes the element-wise multiplication operation.
H L K A ( X ) = C 1 × 1 ( H D D C ( H D C ( X ) ) ) X
In this paper, two sets of LKA mechanisms are designed. Among them, the LKA with a larger receptive field focuses on capturing long-range information dependencies in the image, while the LKA with a smaller receptive field emphasizes preserving local texture details. Compared with common attention methods, the proposed LKA mechanism integrates the advantages of multiscale mechanisms, considering not only the dynamic fusion process of local semantic information but also reducing the channel dimensionality through pointwise convolution, thereby achieving adaptive adjustment of the spatial dimension.

2.3. Feature Reconstruction (FR)

The feature reconstruction stage follows a classical structural design, consisting of a series of cascaded 3 × 3 convolutional layers and pixel shuffle layers. This process can be represented by Equation (7), where H F R ( · ) and H P S ( · ) denote the propagation functions of feature reconstruction and pixel shuffle, respectively. They work together to achieve precise reconstruction of the input feature tensor X . The principle of subpixel convolution is shown in Figure 3.
H F R ( X ) = C 3 × 3 ( H P S ( C 3 × 3 ( X ) ) )

2.4. Coarse Feature Reconstruction (CFR)

The shallow features carry rich detail information, which is crucial for SISR tasks. In order to fully utilize these details, this paper carefully designs a dual-path reconstruction architecture aimed at integrating features from different paths to generate high-quality reconstructed images. In the coarse feature reconstruction path of the dual-path, we only employ a 3 × 3 convolutional layer and a pixel shuffle layer to ensure efficient and accurate extraction and utilization of shallow features. This process is precisely expressed by Equation (8), where H C F R ( · ) and H P S ( · ) represent the propagation functions of coarse feature reconstruction and pixel shuffle, respectively, and X is the input feature tensor. Finally, the H F R ( X ) obtained from feature reconstruction is added to the H C F R ( X ) obtained from coarse feature reconstruction to obtain the final reconstructed high-definition image I S R , which can be represented by Equation (9).
H C F R ( X ) = H P S ( C 3 × 3 ( X ) )
I S R = H F R ( X ) H C F R ( X )

2.5. Loss Function

In terms of optimization strategy, this paper adopts the L 1 loss function, which exhibits high sensitivity to small variations in data and helps the model converge faster. Specifically, this optimization process is explicitly demonstrated by Equation (10). In this equation, θ represents the set of parameters that need to be learned in the entire model, N denotes the total number of training samples, and I S R and I H R , respectively, represent the images reconstructed by the model and the corresponding high-resolution ground truth images.
L 1 ( θ ) = 1 N i = 1 N | | I S R i I H R i | | 1

3. Experimental Results and Analysis

3.1. Datasets

In this paper, the DIV2K dataset is selected as the training set to ensure that the model receives sufficient training and optimization. During the testing phase, standardized benchmark datasets are employed to comprehensively evaluate the performance of different models, including Set5 [23], Set14 [24], BSD100 [25], and Urban100 [26].

3.2. Experimental Details

This paper adopts PyTorch 1.6 as the deep learning framework and leverages the powerful computational capabilities of the GeForce RTX 3090 GPU (NVIDIA, Santa Clara, CA, USA) for model training and testing. To enhance the model’s generalization ability, rotation and mirroring operations are used for data augmentation, enriching the diversity of the dataset. During the training process, the specific parameter settings are as follows: the batch size is set to 16, and the size of input image patches is 96 × 96 to ensure that the model can fully learn image features. Adam optimizer is chosen as the optimizer, with the key parameters β 1 = 0.9 and β 2 = 0.999 . The initial learning rate is set to 10 4 to accelerate the convergence speed of the model in the early stage of training. When the number of model iterations reaches 2 × 10 5 , the learning rate is halved in a timely manner to ensure that the model can perform finer adjustments in the later stage. In order to comprehensively and objectively evaluate the performance of the model, peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) are selected as the two quantitative evaluation metrics.

3.3. Quantitative Analytics

This study primarily employs two metrics, PSNR and SSIM, to evaluate the quality of image reconstruction. As a widely recognized image quality assessment method, PSNR accurately measures the pixel differences between the reconstructed image and the original image. Meanwhile, SSIM focuses on assessing the structural similarity of the image. In our objective quality evaluation experiment, only the y-channel component of the image was tested. The specific calculation methods for PSNR and SSIM can be found in Equations (11) and (12). In these two equations, m × n represents the size of the image, where m and n represent the number of rows and columns of the image, respectively. I H R stands for the ground truth image, and I S R represents the reconstructed image. μ x and μ y represent the pixel means of I H R and I S R , respectively, while δ x 2 and δ y 2 represent their pixel variances, and δ x y represents the pixel covariance between I H R and I S R . Additionally, to avoid zero denominators, extremely small constants c 1 and c 2 are introduced, and their values are set to 0.001 in this study.
PSNR = 20 log 10 255 × m n i = 1 m j = 1 n I H R ( i , j ) I S R ( i , j ) 2
SSIM = 2 μ x μ y + c 1 2 δ x y + c 2 μ x 2 + μ y 2 + c 1 δ x 2 + δ y 2 + c 2
Table 1 compares quantitative metrics for lightweight SISR models across multiple datasets, with the best results in bold and the second-best underlined. The DLKL model consistently outperforms others across all four benchmark datasets, especially at a ×4 scale. This highlights the DLKL model’s superior performance, stability, and ability to handle complex textures and detailed recovery.

3.4. Visual Perception

To intuitively compare the reconstruction performance of various models and highlight their differences in different reconstruction effects, this study deliberately selected the Urban100 dataset with extremely high texture complexity, reconstructed the images at a ×4 magnification factor, and conducted in-depth analysis through visual effect comparisons, as shown in Figure 4, Figure 5, Figure 6, Figure 7, Figure 8 and Figure 9.
Figure 4 presents the results of reconstructing the Img004 image from the Urban100 dataset using different methods. From the figures, it can be seen that the reconstructed images of the SRCNN [9], DRCN [12], and LapSRN [27] models all exhibit significant blurring effects, showing considerable differences from the original image. The images reconstructed by the ARRFN [32] and IMDN [28] models suffer from unclear textures due to pixel loss, with adjacent black dots connected due to blurring. The LBNet [31] and LatticeNet [33] models exhibit excessive distortion in lines. In contrast, our proposed DLKL model performs better in retaining the overall shape and clarity of the black dots in the image. Figure 5 shows the reconstruction comparison of the Img012 image from the Urban100 dataset. Observing the reconstructed images produced by various algorithms, it is evident that the SRCNN [9], DRCN [12], LapSRN [27], and ARRFN [32] models lack detail processing, resulting in blurry textures. The IMDN [28], LAMRN [39], and LBNet [31] models exhibit a large number of water ripples, causing the reconstructed architectural lines to distort and be inconsistent in direction. In contrast, our DLKL model demonstrates superior subjective visual performance, with clear architectural details in the reconstruction and natural lines without distortion. Figure 6 compares the reconstruction results of the building Img019 in the Urban100 dataset. The reconstructed images of SRCNN [9], DRCN [12], IMDN [28], and LatticeNet [33] all show deficiencies in the clarity of window frame contours, background blurriness, and texture details. The ARRFN [32] and LBNet [31] models have improved in window frame details and line distortion, but the overall image is slightly dim and accompanied by slight distortion. In contrast, in our reconstruction, the DLKL model exhibits clear window frame contours, significantly improved overall clarity, and closer resemblance to the original image. Figure 7 further demonstrates the reconstruction effects of other images in the Urban100 dataset. Through comparison, we can see that our proposed DLKL algorithm significantly alleviates the overall blurring issue of the reconstruction images produced by SRCNN [9], DRCN [12], and LapSRN [27]. The images reconstructed by LatticeNet [33] and LBNet [31] exhibit line distortion, while those reconstructed by IMDN [28] and LAMRN [36], although relatively clear in railing lines, have obvious horizontal ripples in the upper part of the image. In contrast, the images reconstructed by the DLKL model not only retain the structural texture of the railing but also have clearer and more prominent lines. In summary, through subjective visual evaluation, our DLKL model demonstrates excellent performance in texture details, line authenticity, and structural clarity of reconstructed images, producing reconstructions with richer details, more natural lines, and clearer structures.
To verify the robustness of the model, we simulate not only regular texture blocks but also irregular ones, such as animal fur and grasses. Figure 8 and Figure 9 display the results for two images from the BSD100 dataset. Despite the geometric irregularity of these image blocks, our algorithm consistently achieves excellent results in quantitative metrics, namely PSNR/SSIM, demonstrating its strong performance and generalization ability.
Given that it is difficult for the benchmark dataset to fully reflect the diversity and complexity of real-world images, we further apply the DLKL algorithm to image testing in real-world scenarios, which are demonstrated in Figure 10, Figure 11 and Figure 12. Through careful observation, especially the hairs of the tiger in Figure 10, the epidermis of the tree in Figure 11, and the schools of fishes in Figure 12, we validate the applicability of the DLKL algorithm on real images. The algorithm not only effectively captures and recovers the complex and fine elements in the image, but also maintains a high degree of visual fidelity, which is a strong proof of the utility and effectiveness of the DLKL algorithm in the field of super-resolution technology.

3.5. Reconstruction Error

In order to clearly demonstrate the superiority of our method (DLKL) in terms of reconstruction accuracy, we systematically compare it with a series of frontier models (SRCNN [9], DRCN [12], IMDN [28], ARRFN [32], LatticeNet [33]). The evaluation is based on the key metric of reconstruction error, specifically using the mean square error to measure the closeness between the super-resolution images generated by each model and the real image. The results show that the reconstruction error of DLKL is significantly lower than that of all compared models, reaching the lowest level. In order to visualize this advantage, we innovatively apply a pseudocoloring technique to process the reconstruction error map, where the pixel regions with error values lower than 15 are marked in blue to indicate the low error part. As shown in Figure 13, Figure 14 and Figure 15, the red pixels (indicating high error regions) in the error maps of SRCNN [9] and DRCN [12] are more significant compared with DLKL, while the mean square errors of the other models are also higher than those of DLKL. This finding, both through visual intuitive comparison and quantitative data analysis, solidly verifies the superior performance of DLKL in reducing reconstruction errors.
To verify the advantages of the DLKL algorithm in accurately processing complex texture details, we planned and implemented four special tests, each targeting a different type of complex texture image: two groups of tests focused on complex texture features in the field of modern architecture, and the other two focused on the rendering of complex textures in natural environments. The results of the experiments are presented in Figure 16, Figure 17, Figure 18 and Figure 19. Quantitative analysis shows that the DLKL algorithm achieves the lowest level of mean square error (MSE) on all texture images involved in the tests, a key metric that provides strong evidence of the algorithm’s superior performance in maintaining image quality. At the visual evaluation level, compared with algorithms such as SRCNN [9] and DRCN [12], the DLKL algorithm achieves a significant reduction in error during the image reconstruction process, as evidenced by the drastic reduction in the originally significant red error highlighted areas in the image. This significant visual improvement again reinforces the extraordinary capability and efficiency of the DLKL algorithm in the field of complex natural texture and fine architectural texture restoration.

3.6. MOS Comparisons

The Mean Opinion Score (MOS) stands as a widely adopted subjective evaluation metric within the domain of visual tasks, enjoying broad international usage. This metric necessitates the selection of a representative cohort of individuals, encompassing both experts and nonexperts in suitable proportions, to solicit their assessments of the provided images.
Evaluation criteria predominantly hinge on the visual comfort experienced by human observers during the image observation process. To ensure the accuracy and fairness of the evaluation, extreme scores are first excluded, and then the remaining scores are averaged in descending order to derive the final MOS result. As depicted in Table 2, our algorithm consistently demonstrates outstanding performance across most scenarios, securing the top rank in MOS assessments. Furthermore, it attains second-place performance in select contexts. This outcome not only furnishes substantial evidence of our DLKL’s capability to generate visually pleasing and natural images but also underscores its robustness and proficiency in effectively restoring a diverse array of textures present in the dataset.

3.7. Comparisons with the Benchmark Model IMDN

IMDN [28] is widely regarded as the benchmark algorithm for SISR, with its effectiveness assessed using quantitative metrics such as PSNR and SSIM. The histogram in Figure 20 illustrates the ∆PSNR or ∆SSIM values between our DLKL and the IMDN benchmarks. The results demonstrate that DLKL consistently surpasses IMDN across all four internationally recognized texture datasets, particularly in more challenging ones such as BSD100 and Urban100. These findings firmly establish DLKL’s superior performance over IMDN in texture restoration.

3.8. Comparisons with the Recurrent-Based Methods

Since our DLKL is a recurrent-based method, presenting a concise overview of the recursive algorithm is essential. Recurrent algorithms are specifically designed to minimize training time by facilitating module reuse and effectively reducing feature loss during information transmission, thereby ensuring optimal information utilization across the entire network. To ensure a fair comparison, we benchmark our DLKL against eleven prominent SISR networks: IMDN [28], SMSR [29], ShuffleMixer [30], LatticeNet [33], LBNet [31], ARRFN [32], PILN [16], DLSR [35], MICU [36], HAFRN [37], and LBRN [38]. The evaluation results, depicted in Figure 21, clearly indicate that our DLKL achieves the highest reconstruction performance among all considered networks.

3.9. Comparison with Transformer-Based Methods

In recent years, researchers have started applying the transformer model to the field of image super-resolution reconstruction, achieving remarkable results. To comprehensively evaluate the performance of the DLKL model in this domain, this study specifically selects four transformer-based lightweight networks for comparison: LBNet [31], ESRT [40], NGSwin [40], and our self-developed DLKL. Detailed comparison results are presented in Table 3. To ensure fairness and accuracy, this study uses Set5, an internationally recognized and widely representative dataset, for rigorous testing. The test results indicate that the DLKL model successfully achieves optimal reconstruction results while maintaining a lightweight framework. This outcome demonstrates DLKL’s ability to efficiently recover image details during reconstruction and its stronger capacity to handle complex textures.

3.10. Comparisons of Topologies of Information Flow

In the design of network architectures, the connection patterns between network modules exhibit diverse characteristics. We can view these modules as hubs of information flow, whose connection layouts spatially form distinct topological structures that determine unique ways of information computation and propagation, further resulting in diverse feature expressions.
In Figure 22, five typical information flow topologies are clearly presented. Among them, the black circular nodes symbolize the key nodes of information flow, while the blue lines depict the paths of information flow. Each topology has its unique information reuse mechanism, thus giving them distinct characteristics. Although skip connection networks significantly enhance algorithm efficiency by reducing computational complexity and the number of parameters, they may also lead to issues such as information loss and insufficient feature fusion. Dense connection networks greatly facilitate gradient propagation and feature reuse by directly connecting each layer to all previous layers, which is highly beneficial for training deep networks. However, this connection pattern also poses challenges such as a large number of parameters, high memory consumption, and increased computational complexity. In contrast, our method, through the ingenious introduction of dual-path reconstruction strategies, can more effectively integrate feature information, reduce the number of parameters, lower information loss, and better fuse shallow features, thus improving the quality of image reconstruction.

3.11. Comparisons of the Inference Speed

The inference speed of neural network models plays a crucial role in enhancing user experience, cutting costs, enabling real-time decision making, boosting deployment flexibility, and sharpening business competitiveness. Hence, during the development and optimization of models, inference speed is a vital metric that cannot be overlooked. This paper compares the reasoning speed of various common algorithms, and as evident from the Table 4, the reasoning speed presented in this paper is notably fast, further underscoring the significance of the large kernel technology employed. In this paper, we choose Urban100 as the test dataset, which contains multiple images with high-definition resolution. To ensure fairness, we adopt the average speed of processing these 100 images as the final inference speed evaluation index of the model.

3.12. Ablation Study

To validate the rationality of the LKA attention module proposed in this paper, we design seven sets of ablation experiments for verification. These experiments focus on different scales of LKA and multipath strategies and are evaluated using PSNR and SSIM metrics under various scale factors. Each experiment consists of four feature extraction basic blocks proposed in this paper, which are trained for a total of 400 epochs, as shown in Table 5. Initially, we investigate the influence of different sizes of LKA on the quality of reconstructed images. As the size of LKA increases, PSNR initially increases and then decreases on datasets Set5, Set14, BSD100, and Urban100. This trend occurs because both long-distance information correlation and local texture information are equally important and indispensable in image super-resolution reconstruction tasks. Subsequently, we explore the impact of multipath mechanisms on its performance. As the number of paths increases, the performance of the module improves. By considering the computational complexity, operation volume, and performance comprehensively, we ultimately select the dual-path LKA with sizes 3-5-1 and 5-7-1 as our multipath large kernel attention model.
The data presented in Table 6 provide evidence that augmenting the number of MAFEs or channels can effectively enhance the performance of the network model, albeit at the cost of increased parameter count. When holding the number of MAFEs constant, a substantial improvement in model reconstruction performance can be achieved by expanding the number of channels. However, this improvement comes with a significant surge in parameter count and computational requirements. Conversely, solely increasing the number of MAFEs results in a relatively modest rise in parameter count and computational demands. Taking into account the requirements of a lightweight super-resolution reconstruction network model, the proposed DLKL model strikes a fine balance by employing six MAFEs and setting the channel count to 64, effectively optimizing both model reconstruction performance and computational overhead.
To assess the influence of each component of MAFE on model performance, this paper conducts four sets of ablation experiments, with the results elaborated in Table 7. Initially, MAFE is solely composed of convolutional layers and GELU demonstrates subpar performance. However, upon incorporating LKA and ECA [22], notable improvements are observed in PSNR and SSIM across all four datasets. This indicates the effective capture of long-range information and intricate textures during image reconstruction, ultimately leading to the retention of both attention mechanisms in the final DLKL model.
To verify the effectiveness of the feature extraction module MAFE proposed in this paper, five sets of ablation experiments were designed. These experiments were conducted under the same experimental conditions and compared with the standard residual block (Resblock) [42], residual-in-residual dense block (RRDB) [46], Residual Channel Attention Block (RCAB) [13], and multiattention block (MAB) [47] implemented using the same LKA. The experimental results are shown in Table 8. The results indicate that the reconstruction performance of the MAFE proposed in this paper is significantly better than the other four modules while using fewer parameters and computational resources. Compared with the commonly used RRDB module in the super-resolution field, the MAFE module proposed in this paper achieved a PSNR improvement of 0.12 dB on the Set5 dataset and 0.43 dB on the challenging Urban100 dataset, which is rich in texture information. These experimental results demonstrate that the MAFE module proposed in this paper achieves a good balance between parameters, computation, and performance.
Finally, to assess the effectiveness of the dual-path reconstruction strategy, ablation experiments are conducted, and the results are presented in Table 9. The findings demonstrate that the introduction of the dual-path reconstruction strategy leads to notable improvements in evaluation metrics across the four datasets while incurring minimal increases in parameter count and computational cost. Particularly noteworthy are the significant enhancements observed on Set14 and Urban100 datasets, both achieving a notable improvement of 0.04 dB. This strategy effectively preserves the low-frequency information and enhances the overall performance of the model.

3.13. Trade-off between Performance and Efficiency

In the experimental study of DLKL model ablation, we fine-tuned several core parameters and mechanisms, including LKA convolutional kernel size, number of MAFE (Multi-Attention Feature Extractor), number of channels, and incorporated LKA and ECA (Enhanced Spatial Attention) mechanisms. This initiative effectively reduces the number of model parameters and computational complexity, while maintaining excellent performance metrics. Specifically, as shown in Table 5, by adjusting the LKA configurations to 7-9-1 and 5-7-1, the number of parameters is reduced from 904.38K to 864.25K, and the FLOPs are reduced from 133.47 G to 127.35 G. Despite the slight decrease in the PSNR on the Urban100 dataset, the model training and inference process is significantly accelerated, which demonstrates the highly efficient computational capability. Further analyzing the data in Table 6, it is found that increasing the number of MAFEs and the number of channels can improve the model performance, but it is accompanied by an increase in computational complexity. After comprehensive consideration, we choose the configuration of six MAFEs with 64 channels to achieve the optimal balance of performance and computational efficiency. Table 7 demonstrates the significant enhancement of model performance by the LKA and ECA mechanisms, especially the excellent performance on PSNR and SSIM, which fully verifies the key role of these mechanisms in enhancing the model to capture long-range dependence and detail information. In addition, the comparative study in Table 8 emphasizes that the MAFE module achieves higher reconstruction accuracy than other feature extraction modules while maintaining lower parameter cost. Finally, Table 9 reveals the positive role of the dual-path reconstruction strategy in preserving low-frequency information and improving the overall reconstruction quality. With these optimizations, the DLKL model successfully achieves an effective balance between performance and computational efficiency.

4. Conclusions

In this paper, a novel Dual-path Large Kernel Learning (DLKL) mechanism is innovatively introduced, leading to the construction of a lightweight SISR model. This model effectively addresses many issues faced by traditional deep learning models, such as excessive parameters, long computation time, and difficulties in real-world deployment. By ingeniously integrating multiscale and large kernel decomposition mechanisms, DLKL successfully establishes robust long-range correlations between different regions while finely preserving intricate local detail information. Additionally, the performance of the network is further enhanced by improving the channel attention mechanism, enabling selective focus on key information within different channels. This design not only ensures compact parameter counts and efficient computational capabilities but also significantly enhances modeling expressiveness and reconstruction performance. Extensive experimental validation demonstrates the outstanding performance of DLKL. Compared with the state-of-the-art lightweight SISR methods like DLSR, DLKL achieves improvements of 0.15 dB and 0.0045 in PSNR and SSIM on the Urban100 dataset, respectively. This achievement fully demonstrates the superior performance of DLKL in terms of fewer parameters and computational resources.
While the information flow topology based on Dual-path Large Kernel Learning has shown some achievements in the super-resolution task, we acknowledge that it may not represent the ultimate optimization solution. Hence, in our future research, we aim to explore more efficient information flow typologies and refine large kernel learning strategies to attain substantial enhancements in system performance and further optimization.

Author Contributions

Z.S., Conceptualization, formal analysis, methodology, software, validation, visualization, writing—original draft, writing—review and editing; H.J., conceptualization, data curation, funding acquisition, resources, supervision, writing—review and editing; M.S., writing—review and editing; X.M., writing—review and editing; R.Z., writing—review and editing; C.L., software, validation, writing—review and editing; Q.K., resources and validation; D.C., resources and validation. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant No. 52304182 and 52204177) and the Open Fund of Key Laboratory of System Control and Information Processing of the Ministry of Education of China (Grant No. SCIP20240105), and Key Laboratory of Pattern Recognition and Intelligent Information Processing, Institutions of Higher Education of Sichuan Province (Grant No. MSSB-2024-01).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, H.; Cheng, D.; Kou, Q.; Asad, M.; Jiang, H. Indicative Vision Transformer for end-to-end zero-shot sketch-based image retrieval. Adv. Eng. Inform. 2024, 60, 102398. [Google Scholar] [CrossRef]
  2. Zhang, H.; Cheng, D.; Jiang, H.; Liu, J.; Kou, Q. Task-like training paradigm in CLIP for zero-shot sketch-based image retrieval. Multimed. Tools Appl. 2024, 83, 57811–57828. [Google Scholar] [CrossRef]
  3. Jiang, H.; Asad, M.; Liu, J.; Zhang, H.; Cheng, D. Single image detail enhancement via metropolis theorem. Multimed. Tools Appl. 2024, 83, 36329–36353. [Google Scholar] [CrossRef]
  4. Park, S.; Choi, G.; Park, S.; Kwak, N. Dual-stage Super-resolution for edge devices. IEEE Access 2023, 11, 123798–123806. [Google Scholar] [CrossRef]
  5. Nguyen, D.P.; Vu, K.H.; Nguyen, D.D.; Pham, H.A. F2SRGAN: A Lightweight Approach Boosting Perceptual Quality in Single Image Super-Resolution via a Revised Fast Fourier Convolution. IEEE Access 2023, 11, 29062–29073. [Google Scholar] [CrossRef]
  6. Cheng, D.; Wang, Y.; Zhang, H.; Li, L.; Kou, Q.; Jiang, H. Intermediate-term memory mechanism inspired lightweight single image super resolution. Multimed. Tools Appl. 2024, 83, 76905–76934. [Google Scholar] [CrossRef]
  7. Cheng, D.; Yuan, H.; Qian, J.; Kou, Q.; Jiang, H. Image Super-Resolution Algorithms Based on Deep Feature Differentiation Network. J. Electron. Inf. Technol. 2024, 46, 1–10. [Google Scholar]
  8. Cheng, D.; Chen, J.; Kou, Q.; Nie, S.; Zhang, L. Lightweight Super-resolution Reconstruction Method Based on Hierarchical Features Fusion and Attention Mechanism for Mine Image. Chin. J. Sci. Instrum. 2022, 43, 73–84. [Google Scholar]
  9. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 38, 295–307. [Google Scholar] [CrossRef]
  10. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1646–1654. [Google Scholar]
  11. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honoluu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar]
  12. Kim, J.; Lee, J.K.; Lee, K.M. Deeply-Recursive Convolutional Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1637–1645. [Google Scholar]
  13. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  14. Zhao, H.; Kong, X.; He, J.; Qiao, Y.; Dong, C. Efficient Image Super-Resolution Using Pixel Attention. In Proceedings of the Computer Vision—ECCV 2020 Workshops, Glasgow, UK, 23–28 August 2020; pp. 56–72. [Google Scholar]
  15. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-Trained Image Processing Transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12294–12305. [Google Scholar]
  16. Kou, Q.; Cheng, D.; Zhang, H.; Liu, J.; Guo, X.; Jiang, H. Single Image Super Resolution via Multi-Attention Fusion Recurrent Network. IEEE Access 2023, 11, 98653–98665. [Google Scholar] [CrossRef]
  17. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  18. Wan, J.; Yin, H.; Liu, Z.; Chong, A.; Liu, Y. Lightweight Image Super-Resolution by Multi-Scale Aggregation. IEEE Trans. Broadcast. 2021, 67, 372–382. [Google Scholar] [CrossRef]
  19. Wang, Z.; Gao, G.; Li, J.; Yu, Y.; Lu, H. Lightweight Image Super-Resolution with Multi-Scale Feature Interaction Network. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  20. Lan, R.; Sun, L.; Liu, Z.; Lu, H.; Pang, C.; Luo, X. MADNet: A Fast and Lightweight Network for Single-Image Super Resolution. IEEE Trans. Cybern. 2020, 51, 1443–1453. [Google Scholar] [CrossRef] [PubMed]
  21. Guo, M.H.; Lu, C.Z.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Visual attention network. Comput. Vis. Media 2022, 9, 733–752. [Google Scholar] [CrossRef]
  22. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 11531–11539. [Google Scholar]
  23. Bevilacqua, M.; Roumy, A.; Guillemot, C.; Alberi-Morel, M.L. Low-complexity single-image super-resolution based on nonnegative neighbor embedding. In Proceedings of the British Machine Vision Conference, Surrey, UK, 3–7 September 2012; pp. 135.1–135.10. [Google Scholar]
  24. Zeyde, R.; Elad, M.; Protter, M. On Single Image Scale-Up Using Sparse-Representations. In Curves and Surfaces; Springer: Berlin/Heidelberg, Germany, 2010; pp. 711–730. [Google Scholar]
  25. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision, Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  26. Huang, J.B.; Singh, A.; Ahuja, N. Single image super-resolution from transformed self-exemplars. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 5197–5206. [Google Scholar]
  27. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep Laplacian Pyramid Networks for Fast and Accurate Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5835–5843. [Google Scholar]
  28. Hui, Z.; Gao, X.; Yang, Y.; Wang, X. Lightweight Image Super-Resolution with Information Multi-distillation Network. In Proceedings of the ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 2024–2032. [Google Scholar]
  29. Wang, L.; Dong, X.; Wang, Y.; Ying, X.; Lin, Z.; An, W.; Guo, Y. Exploring Sparsity in Image Super-Resolution for Efficient Inference. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4915–4924. [Google Scholar]
  30. Sun, L.; Pan, J.; Tang, J. Shufflemixer: An efficient convnet for image super-resolution. Adv. Neural Inf. Process. Syst. 2022, 35, 17314–17326. [Google Scholar]
  31. Gao, G.; Wang, Z.; Li, J.; Li, W.; Yu, Y.; Zeng, T. Lightweight bimodal network for single-image super-resolution via symmetric CNN and recursive transformer. In Proceedings of the International Joint Conferences on Artificial Intelligence Organization, Vienna, Austria, 23–29 July 2022; pp. 913–919. [Google Scholar]
  32. Qin, J.; Zhang, R. Lightweight single image super-resolution with attentive residual refinement network. Neurocomputing 2022, 500, 846–855. [Google Scholar] [CrossRef]
  33. Luo, X.; Qu, Y.; Xie, Y.; Zhang, Y.; Li, C.; Fu, Y. Lattice Network for Lightweight Image Restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 4826–4842. [Google Scholar] [CrossRef]
  34. Qin, J.; Chen, L.; Jeon, S.; Yang, X. Progressive interaction-learning network for lightweight single-image super-resolution in industrial applications. IEEE Trans. Ind. Inform. 2023, 19, 2183–2191. [Google Scholar] [CrossRef]
  35. Huang, H.; Shen, L.; He, C.; Dong, W.; Liu, W. Differentiable Neural Architecture Search for Extremely Lightweight Image Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 2672–2682. [Google Scholar] [CrossRef]
  36. Chen, Y.; Xia, R.; Yang, K.; Zou, K. MICU: Image super-resolution via multi-level information compensation and U-net. Expert Syst. Appl. 2024, 245, 123111. [Google Scholar] [CrossRef]
  37. Wanga, K.; Yanga, X.; Jeon, G. Hybrid attention feature refinement network for lightweight image super-resolution in metaverse immersive display. IEEE Trans. Consum. Electron. 2024, 70, 3232–3244. [Google Scholar] [CrossRef]
  38. Hao, F.; Wu, J.; Liang, W.; Xu, J.; Li, P. Lightweight blueprint residual network for single image super-resolution. Expert Syst. Appl. 2024, 250, 123954. [Google Scholar] [CrossRef]
  39. Yan, Y.; Xu, X.; Chen, W.; Peng, X. Lightweight Attended Multi-Scale Residual Network for Single Image Super-Resolution. IEEE Access 2021, 9, 52202–52212. [Google Scholar] [CrossRef]
  40. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. Transformer for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 457–466. [Google Scholar]
  41. Choi, H.; Lee, J.; Yang, J. N-gram in swin transformers for efficient lightweight image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 2071–2081. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  43. Mao, X.; Shen, C.; Yang, Y.B. Image restoration using very deep convolutional encoder-decoder networks with symmetric skip connections. Adv. Neural Inf. Process. Syst. 2016, 29, 2810–2818. [Google Scholar]
  44. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4539–4547. [Google Scholar]
  45. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  46. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Change Loy, C.; Qiao, Y.; Tang, X. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018; pp. 63–79. [Google Scholar]
  47. Muqeet, A.; Hwang, J.; Yang, S.; Kang, J.; Kim, Y.; Bae, S.H. Multi-attention based ultra lightweight image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Glasgow, UK, 23–28 August 2020; pp. 103–118. [Google Scholar]
Figure 1. Flowchart of the DLKL system, where the symbol ⊕ means element-wise sum.
Figure 1. Flowchart of the DLKL system, where the symbol ⊕ means element-wise sum.
Sensors 24 06174 g001
Figure 2. Structure of LKA, where DC means depth convolution and DDC means depth-dilated convolution.
Figure 2. Structure of LKA, where DC means depth convolution and DDC means depth-dilated convolution.
Sensors 24 06174 g002
Figure 3. Schematic diagram of pixel shuffle.
Figure 3. Schematic diagram of pixel shuffle.
Sensors 24 06174 g003
Figure 4. Comparison of × 4 scale reconstruction results for different algorithms on Img004 in Urban100.
Figure 4. Comparison of × 4 scale reconstruction results for different algorithms on Img004 in Urban100.
Sensors 24 06174 g004
Figure 5. Comparison of × 4 scale reconstruction results for different algorithms on Img012 in Urban100.
Figure 5. Comparison of × 4 scale reconstruction results for different algorithms on Img012 in Urban100.
Sensors 24 06174 g005
Figure 6. Comparison of × 4 scale reconstruction results for different algorithms on Img019 in Urban100.
Figure 6. Comparison of × 4 scale reconstruction results for different algorithms on Img019 in Urban100.
Sensors 24 06174 g006
Figure 7. Comparison of × 4 scale reconstruction results for different algorithms on Img024 in Urban100.
Figure 7. Comparison of × 4 scale reconstruction results for different algorithms on Img024 in Urban100.
Sensors 24 06174 g007
Figure 8. Comparison of × 4 scale reconstruction results for different algorithms on 38082 in BSD100.
Figure 8. Comparison of × 4 scale reconstruction results for different algorithms on 38082 in BSD100.
Sensors 24 06174 g008
Figure 9. Comparison of × 4 scale reconstruction results for different algorithms on 69015 in BSD100.
Figure 9. Comparison of × 4 scale reconstruction results for different algorithms on 69015 in BSD100.
Sensors 24 06174 g009
Figure 10. Comparison of × 4 -scale reconstruction results for different algorithms on tiger.
Figure 10. Comparison of × 4 -scale reconstruction results for different algorithms on tiger.
Sensors 24 06174 g010
Figure 11. Comparison of × 4 -scale reconstruction results for different algorithms on tree.
Figure 11. Comparison of × 4 -scale reconstruction results for different algorithms on tree.
Sensors 24 06174 g011
Figure 12. Comparison of × 4 -cale reconstruction results for different algorithms on fish.
Figure 12. Comparison of × 4 -cale reconstruction results for different algorithms on fish.
Sensors 24 06174 g012
Figure 13. Comparison of × 4 -scale reconstruction errors for different algorithms on flowers in Set14, with the name of the algorithm and its mean square error value below the image.
Figure 13. Comparison of × 4 -scale reconstruction errors for different algorithms on flowers in Set14, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g013
Figure 14. Comparison of × 4 -scale reconstruction errors for different algorithms on comic in Set14, with the name of the algorithm and its mean square error value below the image.
Figure 14. Comparison of × 4 -scale reconstruction errors for different algorithms on comic in Set14, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g014
Figure 15. Comparison of × 4 -scale reconstruction errors for different algorithms on zebra in Set14, with the name of the algorithm and its mean square error value below the image.
Figure 15. Comparison of × 4 -scale reconstruction errors for different algorithms on zebra in Set14, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g015
Figure 16. Comparison of × 4 -scale reconstruction errors for different algorithms on img008 in Urban100, with the name of the algorithm and its mean square error value below the image.
Figure 16. Comparison of × 4 -scale reconstruction errors for different algorithms on img008 in Urban100, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g016
Figure 17. Comparison of × 4 -scale reconstruction errors for different algorithms on img090 in Urban100, with the name of the algorithm and its mean square error value below the image.
Figure 17. Comparison of × 4 -scale reconstruction errors for different algorithms on img090 in Urban100, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g017
Figure 18. Comparison of × 4 -scale reconstruction errors for different algorithms on 42049 in BSD100, with the name of the algorithm and its mean square error value below the image.
Figure 18. Comparison of × 4 -scale reconstruction errors for different algorithms on 42049 in BSD100, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g018
Figure 19. Comparison of × 4 -scale reconstruction errors for different algorithms on monarch in Set14, with the name of the algorithm and its mean square error value below the image.
Figure 19. Comparison of × 4 -scale reconstruction errors for different algorithms on monarch in Set14, with the name of the algorithm and its mean square error value below the image.
Sensors 24 06174 g019
Figure 20. Comparison with the IMDN model. × 2 , × 3 and × 4 tests in Set5, Set14, BSD100, and Urban100. Objective evaluation metrics are ∆PSNR or ∆SSIM. The test results are displayed as histograms, with the dataset names directly below.
Figure 20. Comparison with the IMDN model. × 2 , × 3 and × 4 tests in Set5, Set14, BSD100, and Urban100. Objective evaluation metrics are ∆PSNR or ∆SSIM. The test results are displayed as histograms, with the dataset names directly below.
Sensors 24 06174 g020
Figure 21. × 4 SISR performance comparisons of recurrent-based algorithms. The objective evaluation metric is PSNR, and Set5, Set14, BSD100, and Urban100 are test datasets.
Figure 21. × 4 SISR performance comparisons of recurrent-based algorithms. The objective evaluation metric is PSNR, and Set5, Set14, BSD100, and Urban100 are test datasets.
Sensors 24 06174 g021
Figure 22. Topologies of information flow, and they are plain-based, residual-based [42], skip connection [43], memory-based [44], dense connection [45], and our approach.
Figure 22. Topologies of information flow, and they are plain-based, residual-based [42], skip connection [43], memory-based [44], dense connection [45], and our approach.
Sensors 24 06174 g022
Table 1. Quantitative tests: The average PSNR and SSIM values for SISR results at a scale of × 4 on the datasets: Set5, Set14, BSD100, and Urban100. To assess the DLKL model’s performance, it is compared against other state-of-the-art methods, emphasizing the best results in bold and the second-best results with underline, respectively. The notation “-” indicates that the pertinent data are absent in instances where the authors have not released their source code publicly or in officially published papers.
Table 1. Quantitative tests: The average PSNR and SSIM values for SISR results at a scale of × 4 on the datasets: Set5, Set14, BSD100, and Urban100. To assess the DLKL model’s performance, it is compared against other state-of-the-art methods, emphasizing the best results in bold and the second-best results with underline, respectively. The notation “-” indicates that the pertinent data are absent in instances where the authors have not released their source code publicly or in officially published papers.
ModelScaleFLOPSet5 [23]Set14 [24]BSD100 [25]Urban100 [26]
PSNR/SSIMPSNR/SSIMPSNR/SSIMPSNR/SSIM
SRCNN (TPAMI 2014) [9] × 2 52.7 G36.66/0.954232.43/0.907331.34/0.887929.50/0.8953
LapSRN (TPAMI 2017) [27]29.9 G37.52/0.959033.08/0.913031.80/0.895030.41/0.9100
IMDN (ACM MM 2019) [28]186.7 G38.00/0.960433.63/0.917332.19/0.899432.17/0.9283
SMSR (CVPR 2021) [29]131.6 G38.00/0.960133.64/0.917332.17/0.899032.19/0.9284
ShuffleMixer (NeurIPS 2022) [30]91.0 G38.01/0.960633.63/0.918032.17/0.899531.89/0.9257
LBNet (IJCAI 2022)  [31]-38.05/0.960733.65/0.917732.15/0.899432.30/0.9291
ARRFN (NC 2022) [32]-38.01/0.960333.66/0.917432.20/0.899932.37/0.9295
LatticeNet (TPAMI 2023) [33]169.0 G38.06/0.960733.70/0.918732.20/0.899932.25/0.9302
PILN (TII 2023) [34]-38.08/0.960733.72/0.918132.23/0.900332.38/0.9306
DLSR (TCSVT 2023) [35]-38.04/0.960633.67/0.918332.21/0.900232.26/0.9297
MICU (ESWA 2024) [36]-37.93/0.960133.63/0.917032.17/0.898732.09/0.9271
HAFRN (TCE 2024) [37]-38.05/0.960633.66/0.918732.21/0.899932.20/0.9289
LBRN (ESWA 2024) [38]-38.08/0.960833.57/0.917332.23/0.900532.35/0.9303
Our DLKL174.1 G38.12/0.961433.71/0.918732.23/0.900432.47/0.9315
SRCNN (TPAMI 2014) [9] × 3 52.7 G32.75/0.909029.30/0.821928.41/0.786326.25/0.7989
LapSRN (TPAMI 2017) [27]-33.82/0.923229.87/0.832428.82/0.798027.07/0.8281
IMDN (ACM MM 2019) [28]84.0 G34.36/0.927030.32/0.841729.09/0.804628.17/0.8519
SMSR (CVPR 2021) [29]67.8 G34.40/0.927030.33/0.841229.10/0.805028.25/0.8536
ShuffleMixer (NeurIPS 2022)  [30]43.0 G34.40/0.927230.37/0.842329.12/0.805128.08/0.8498
LBNet (IJCAI 2022) [31]-34.47/0.927730.38/0.841729.13/0.806128.42/0.8559
ARRFN (NC 2022) [32]-34.38/0.927230.36/0.842229.09/0.805028.22/0.8533
LatticeNet (TPAMI 2023) [33]76.0 G34.40/0.927230.32/0.841629.10/0.804928.19/0.8513
PILN (TII 2023) [34]-34.39/0.926930.34/0.841529.08/0.804828.09/0.8500
DLSR (TCSVT 2023) [35]-34.49/0.927930.39/0.842829.13/0.806128.26/0.8548
MICU (ESWA 2024) [36]-34.38/0.927430.35/0.841929.10/0.804828.14/0.8518
HAFRN (TCE 2024) [37]-34.45/0.927630.40/0.843329.12/0.805828.16/0.8528
LBRN (ESWA 2024) [38]-34.43/0.927630.39/0.842929.13/0.805928.29/0.8545
Our DLKL101.3 G34.57/0.928930.45/0.843829.16/0.806528.37/0.8569
SRCNN (TPAMI 2014) [9] × 4 52.7 G30.48/0.862827.50/0.750326.90/0.711024.53/0.7212
LapSRN (TPAMI 2017) [27]149.4 G31.54/0.885028.19/0.772027.32/0.728025.21/0.7560
IMDN (ACM MM 2019) [28]48.0 G32.21/0.894828.58/0.781127.56/0.735326.04/0.7838
SMSR (CVPR 2021) [29]41.6 G32.12/0.893228.55/0.780827.55/0.735126.11/0.7868
ShuffleMixer (NeurIPS 2022) [30]28.0 G32.21/0.895328.66/0.782727.61/0.736626.08/0.7835
LBNet (IJCAI 2022)  [31]-32.29/0.896028.68/0.783227.62/0.738226.27/0.7906
ARRFN (NC 2022) [32]-32.22/0.895228.60/0.781727.57/0.735526.09/0.7858
LatticeNet (TPAMI 2023)  [33]43.0 G32.18/0.894328.61/0.781227.57/0.735526.14/0.7844
PILN (TII 2023) [34]-32.22/0.894928.62/0.781327.59/0.736526.19/0.7878
DLSR (TCSVT 2023)  [35]20.41 G32.33/0.896328.68/0.783227.61/0.737426.19/0.7892
MICU (ESWA 2024) [36]-32.21/0.894528.65/0.782027.57/0.735926.15/0.7872
HAFRN (TCE 2024) [37]-32.24/0.895328.60/0.781627.58/0.736526.02/0.7849
LBRN (ESWA 2024) [38]-32.33/0.896428.62/0.782627.60/0.737726.17/0.7882
Our DLKL66.1 G32.39/0.897628.71/0.785327.65/0.738626.34/0.7937
Table 2. MOS comparisons. The top 5 algorithms for scale factors × 4 on datasets Set5, Set14, BSD100, and Urban100. Our DLKL is highlighted in black bold.
Table 2. MOS comparisons. The top 5 algorithms for scale factors × 4 on datasets Set5, Set14, BSD100, and Urban100. Our DLKL is highlighted in black bold.
DatasetScaleTop 5 Algorithms
Set5  [23] × 4 DLKL > LBNet > LatticeNet> LAMRN > ARRFN
Set14  [24] × 4 LatticeNet > DLKL > LAMRN > IMDN > LBNet
BSD100  [25] × 4 DLKL > LAMRN > ARRFN > LatticeNet> LBNet
Urban100  [26] × 4 DLKL > LBNet > ARRFN> LAMRN > LatticeNet
Table 3. The average PSNR and SSIM values for SISR results at a scale of × 4 on the datasets: Set5. To assess the DLKL model’s performance, it is compared against other state-of-the-art transformer methods, emphasizing the best results in bold.
Table 3. The average PSNR and SSIM values for SISR results at a scale of × 4 on the datasets: Set5. To assess the DLKL model’s performance, it is compared against other state-of-the-art transformer methods, emphasizing the best results in bold.
Transformer-Based MethodsScaleParaSet5 [23]
LBNet (IJCAI 2022)  [31] × 4 742 K32.29/0.8960
ESRT (CVPR 2022)  [40] × 4 751 K32.19/0.8947
NGswin (CVPR 2023) [41] × 4 1019 K32.33/0.8963
DLKL × 4 864 K32.39/0.8976
Table 4. Comparisons on the model’s inference speed (ms).
Table 4. Comparisons on the model’s inference speed (ms).
MethodSpeedMethodSpeedMethodSpeedMethodSpeed
SRCNN [9]297LapSRN [27]189EDSR-BASE [11]315IMDN [28]38
SMSR [29]309ARRFN [32]234LBNet [31]679DLKL142
Table 5. Effects of the LKA scale and multibranch mechanism on the performance of DLKA. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance. Specially, ✓ means the method is used, × means the method is not used.
Table 5. Effects of the LKA scale and multibranch mechanism on the performance of DLKA. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance. Specially, ✓ means the method is used, × means the method is not used.
LKA
3-5-1
LKA
5-7-1
LKA
7-9-1
ParaFLOPSet5 [23]
PSNR/SSIM
Set14 [24]
PSNR/SSIM
BSD100 [25]
PSNR/SSIM
Urban100 [26]
PSNR/SSIM
××767.96 K113.69 G37.93/0.958733.43/0.915232.08/0.897531.74/0.9237
××778.20 K115.20 G37.94/0.959233.46/0.915932.10/0.897831.74/0.9238
××791.01 K117.31 G37.93/0.958433.44/0.915432.09/0.897731.73/0.9234
×837.08 K123.73 G37.97/0.959833.50/0.916832.12/0.898331.80/0.9243
×864.25 K127.35 G37.98/0.960133.51/0.917132.11/0.898131.78/0.9238
×853.47 K125.84 G37.96/0.960133.50/0.916632.10/0.897631.77/0.9235
904.38 K133.47 G37.98/0.960433.52/0.917632.11/0.898431.80/0.9246
Table 6. Comparison of model complexity and performance metrics: Average PSNR/SSIM for × 2 SISR results on benchmark datasets, where MAFEn and Chan indicate the number of MAFE and channels, respectively. Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance.
Table 6. Comparison of model complexity and performance metrics: Average PSNR/SSIM for × 2 SISR results on benchmark datasets, where MAFEn and Chan indicate the number of MAFE and channels, respectively. Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance.
MAFEnChanParaFLOPSet5 [23]
PSNR/SSIM
Set14 [24]
PSNR/SSIM
BSD100 [25]
PSNR/SSIM
Urban100 [26]
PSNR/SSIM
4640.84 M123.78 G38.06/0.960333.61/0.916032.18/0.899432.15/0.9281
4961.86 M274.45 G38.13/0.961733.71/0.918632.24/0.900632.40/0.9304
6641.18 M174.1 G38.12/0.961433.71/0.918732.23/0.900432.47/0.9315
6962.62 M386.26 G38.16/0.962133.83/0.919732.29/0.901132.60/0.9327
Table 7. Impact of LKA and ECA on MAFE Performance: Average PSNR/SSIM for × 2 SISR on Benchmarks, where Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance. Specially, ✓ means the method is used, × means the method is not used.
Table 7. Impact of LKA and ECA on MAFE Performance: Average PSNR/SSIM for × 2 SISR on Benchmarks, where Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance. Specially, ✓ means the method is used, × means the method is not used.
LKAECAParaFLOPSet5 [23]
PSNR/SSIM
Set14 [24]
PSNR/SSIM
BSD100 [25]
PSNR/SSIM
Urban100 [26]
PSNR/SSIM
××743.92 K109.99 G37.81/0.957333.34/0.914232.03/0.896931.62/0.9213
×837.08 K123.73 G37.97/0.959833.50/0.916832.12/0.898331.80/0.9243
×743.93 K110.03 G37.84/0.957633.40/0.914932.06/0.897231.72/0.9225
837.09 K123.77 G37.99/0.960233.51/0.916932.15/0.898731.91/0.9255
Table 8. Comparison of feature extraction blocks at scale factor 2, where Number represents the number of base blocks, Chan represents the number of channels, Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance.
Table 8. Comparison of feature extraction blocks at scale factor 2, where Number represents the number of base blocks, Chan represents the number of channels, Para denotes the number of model parameters, and FLOP refers to the quantity of floating-point operations. The strategy used in this article is highlighted in bold, which is a comprehensive consideration of efficiency and performance.
BlockNumberChanParaFLOPSet5 [23]
PSNR/SSIM
Set14 [24]
PSNR/SSIM
BSD100 [25]
PSNR/SSIM
Urban100 [26]
PSNR/SSIM
Resblock  [42]12641.04 M153.48 G37.74/0.958733.25/0.913931.95/0.896331.07/0.9162
RRDB [46]4640.89 M131.73 G37.87/0.959833.38/0.915532.05/0.897231.48/0.9193
RCAB [13]4640.74 M66.54 G37.77/0.959133.21/0.913231.92/0.896031.09/0.9165
MAB [47]14901.08 M158.50 G37.94/0.960133.48/0.916232.15/0.898631.79/0.9232
DLKL4640.83 M123.77 G37.99/0.960433.51/0.916532.15/0.898731.91/0.9254
Table 9. Impact of dual-path reconstruction on ×2 SISR performance: Average PSNR/SSIM on Benchmarks. The best results are highlighted in bold. Specially, ✓ means the method is used, × means the method is not used.
Table 9. Impact of dual-path reconstruction on ×2 SISR performance: Average PSNR/SSIM on Benchmarks. The best results are highlighted in bold. Specially, ✓ means the method is used, × means the method is not used.
Dual
Path
Set5 [23]
PSNR/SSIM
Set14 [24]
PSNR/SSIM
BSD100 [25]
PSNR/SSIM
Urban100 [26]
PSNR/SSIM
×37.99/0.960233.51/0.916932.15/0.898731.91/0.9255
38.00/0.960533.55/0.917532.16/0.899331.95/0.9264
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Su, Z.; Sun, M.; Jiang, H.; Ma, X.; Zhang, R.; Lv, C.; Kou, Q.; Cheng, D. Dual-Path Large Kernel Learning and Its Applications in Single-Image Super-Resolution. Sensors 2024, 24, 6174. https://doi.org/10.3390/s24196174

AMA Style

Su Z, Sun M, Jiang H, Ma X, Zhang R, Lv C, Kou Q, Cheng D. Dual-Path Large Kernel Learning and Its Applications in Single-Image Super-Resolution. Sensors. 2024; 24(19):6174. https://doi.org/10.3390/s24196174

Chicago/Turabian Style

Su, Zhen, Mang Sun, He Jiang, Xiang Ma, Rui Zhang, Chen Lv, Qiqi Kou, and Deqiang Cheng. 2024. "Dual-Path Large Kernel Learning and Its Applications in Single-Image Super-Resolution" Sensors 24, no. 19: 6174. https://doi.org/10.3390/s24196174

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop