Next Article in Journal
The Born Rule—100 Years Ago and Today
Previous Article in Journal
Analysis of the Relationship Between Enguri Large Dam Monitoring Entropic Features
Previous Article in Special Issue
Action Recognition with 3D Residual Attention and Cross Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution

1
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, Kunming 650500, China
2
Yunnan Province Key Laboratory of Computer, Kunming University of Science and Technology, Kunming 650500, China
*
Author to whom correspondence should be addressed.
Entropy 2025, 27(4), 414; https://doi.org/10.3390/e27040414
Submission received: 15 February 2025 / Revised: 31 March 2025 / Accepted: 9 April 2025 / Published: 11 April 2025

Abstract

:
Single-image super-resolution (SISR) based on GANs has achieved significant progress. However, these methods still face challenges when reconstructing locally consistent textures due to a lack of semantic understanding of image categories. This highlights the necessity of focusing on contextual information comprehension and the acquisition of high-frequency details in model design. To address this issue, we propose the Semantic Prior-Improved GAN (SP-IGAN) framework, which incorporates additional contextual semantic information into the Real-ESRGAN model. The framework consists of two branches. The main branch introduces a Graph Convolutional Channel Attention (GCCA) module to transform channel dependencies into adjacency relationships between feature vertices, thereby enhancing pixel associations. The auxiliary branch strengthens the correlation between semantic category information and regional textures in the Residual-in-Residual Dense Block (RRDB) module. The auxiliary branch employs a pretrained segmentation model to accurately extract regional semantic information from the input low-resolution image. This information is injected into the RRDB module through Spatial Feature Transform (SFT) layers, generating more accurate and semantically consistent texture details. Additionally, a wavelet loss is incorporated into the loss function to capture high-frequency details that are often overlooked. The experimental results demonstrate that the proposed SP-IGAN outperforms state-of-the-art (SOTA) super-resolution models across multiple public datasets. For the X4 super-resolution task, SP-IGAN achieves a 0.55 dB improvement in Peak Signal-to-Noise Ratio (PSNR) and a 0.0363 increase in Structural Similarity Index (SSIM) compared to the baseline model Real-ESRGAN.

1. Introduction

The goal of SISR is to recover high-frequency information that is missing from low-resolution images, a task that poses significant challenges. SISR has a wide range of practical applications requiring high-resolution images, such as surveillance image enhancement [1], satellite remote sensing data processing [2], and identity recognition. Traditional super-resolution methods [3] are often constrained by algorithmic resolution limits and unrealistic detail recovery, making them inadequate for practical applications.
In recent years, SISR methods have been developed based on convolutional neural networks (CNNs) [4] and generative adversarial networks (GANs) [5], among others. For example, SRCNN [6] was the first to introduce deep convolutional neural networks to learn the mapping from low-resolution to high-resolution images, incorporating pixel loss (L1 loss and Mean Squared Error (MSE) loss). Models such as VDSR [7], EDSR [8], and RCAN [9] increased network depth and introduced mechanisms like residual learning to mitigate the issue of gradient explosion while enhancing the network’s depth. Subsequently, various CNN design frameworks [6,10] have been proposed to enhance the representational capacity of super-resolution networks through pixel-level loss constraints. Although these methods improve reconstruction quality, the generated textures are overly smooth and lack sufficient high-frequency details. To better meet subjective visual quality requirements [11], more detailed and visually appealing textures are needed. Inspired by generative adversarial networks (GANs) [12], Christian Ledig et al. proposed SRGAN [13], which was the first model to incorporate perceptual loss into the original loss function, placing greater emphasis on visual texture details. ESRGAN [14], proposed by Wang et al, is an improved version of SRGAN. It incorporates Residual-in-Residual Dense Blocks (RRDBs) and a Relativistic Discriminator, which not only enhance image quality but also preserve important details. A series of GAN-based models [15,16,17] have advanced SR networks in their generation of fine-grained semantic textures. However, GAN-based models often produce unrealistic textures and unwanted artifacts during adversarial training when generating high-resolution images.
With the rapid advancement of deep learning, the application scope of single-image super-resolution (SISR) has gradually expanded to real-world scenarios. However, traditional degradation processes are often simplified into known standard models, such as those in [6,13], to generate low-resolution input images. Nonetheless, super-resolution methods based on fixed degradation models [3,4,6,12] face significant challenges when dealing with the more complex degradation processes of real-world images. To address this issue, Fritsche et al. proposed a real-world degradation model [18], which improves the traditional degradation process in SRGAN by incorporating real-world factors such as noise, blur, and JPEG compression, significantly enhancing the model’s generalization ability. Additionally, the Real-ESRGAN model proposed by Wang et al. [19] introduced a high-order degradation model that simulates more realistic image degradation processes. By incorporating real-world image data during training, it further improves the model’s generalization ability, enabling it to learn and reconstruct richer, more detailed image textures. The DASR model [20] employs an adaptive degradation mechanism, excelling at handling unpredictable degradation in real-world images and effectively recovering more realistic image details. However, these methods still face two major issues: (1) in complex scenarios, models are more prone to inconsistent local texture reconstruction due to a lack of semantic understanding; (2) it has yet to be determined how we can better leverage contextual information to improve texture generation and reduce artifacts.
In summary, this paper extends the GCCA module to the Real-ESRGAN model with semantic information. The introduction of semantic priors enhances the model’s understanding of image structure, while abstracting convolutionally extracted texture information into feature representations. The texture data from previous layers are treated as feature vertices, with their relationships represented by an adjacency matrix, enabling contextual dependencies. This allows the model to capture long-range contextual relationships, effectively utilizing semantic information to reduce semantic inconsistencies in reconstructed images. Additionally, incorporating wavelet loss into the total loss function further enhances the model’s ability to capture high-frequency details and reduce artifacts. The proposed branch structure, SP-IGAN, enables the model to fully understand the generated image structure and type, preventing information confusion and producing realistic detailed information. The contributions of this paper are summarized as follows:
  • The Spatial Feature Transform (SFT) layer is introduced into the Residual-in-Residual Dense Block (RRDB) module. The pre-trained segmentation network in the auxiliary branch accurately extracts regional semantic information, which is then conditioned and input into the RRDB module through the SFT layer via affine transformations. This provides the model with additional contextual semantic understanding, enhancing the correlation between texture information and category information, which leads to more accurate and semantically consistent texture reconstruction. Additionally, a Star-shaped Residual-in-Residual Dense Block (StarRRDB) [21] is employed, which offers higher network capacity compared to the RRDB module used in Real-ESRGAN. The design of StarRRDB is inspired by ESRGAN+ [22], providing greater flexibility and potential for the integration of the SFT layer and the subsequent GCCA module.
  • The Graph Convolutional Channel Attention (GCCA) module enhances the RRDB module by transforming channel dependencies into adjacency relationships of feature vertices, thereby improving pixel correlation. The GCCA module draws on the theory of graph neural networks, focusing on the attention mechanism for feature interaction between channels. It effectively highlights key features, suppresses secondary features, reduces computational redundancy, and enhances the model’s ability to capture long-range dependencies.
  • Add the wavelet loss [23] to the overall loss function. Wavelet loss effectively captures high-frequency details that are often overlooked by conventional pixel-based loss functions. By combining wavelet loss with the standard GAN loss, the model can more accurately recover high-frequency components, thereby generating images with richer textures and finer details.

2. Related Work

2.1. Single-Image Super-Resolution Method Incorporating the Attention Mechanism

The attention mechanism [24] simulates human cognitive processes, effectively focusing on key information and re-weighting the features in the network. However, when processing outdoor scenes and real-world images with complex textures, the generated textures often lack realism, and the utilization of long-range contextual information is insufficient. Therefore, it is crucial to guide the model in understanding the relevance of long-range texture information through the attention mechanism. In recent years, many attention-based networks have been proposed to address various visual tasks and have been widely applied in CNNs. For example, Hu et al. first introduced SENet [25], which introduced a more modular attention mechanism compared to other networks. Zhang et al. introduced the attention mechanism into the image super-resolution task, incorporating the Channel Attention (CA) mechanism in their proposed RCAN [9] to focus attention on more important channels. Buades A et al. employed the Non-Local (NL) [26] module, calculating the response at a specific location as the weighted sum of features from all locations. Woo and Park used the CBAM [27] to enhance the expressive power of channel and spatial features. Furthermore, Dai et al. proposed the Second-order Attention Network (SAN) [28], which incorporates long-range dependencies and global information from the entire network structure. This method enables the model to effectively leverage information across different scales and levels when capturing image details, resulting in significant performance improvements.

2.2. Single-Image Super-Resolution Method Incorporating Prior Information

Traditional super-resolution methods typically rely on a single low-resolution image for reconstruction. However, introducing auxiliary conditional information provides the generative network with a richer understanding of the contextual information, enabling more detailed image reconstruction. Wang et al. first introduced the SFT layer in the SFTGAN [29], integrating semantic probability maps into the network. This allows the model to generate corresponding texture features during image generation, thereby enhancing the realism of textures. The SROOE [30] proposed by Park et al. predicts the optimal target image as conditional information to be input into the generative model, which, through training on the target trajectory, generates more credible results. The FxSR [31] incorporates multi-task learning, training the model with combinations of various loss functions as conditional information. By introducing different weightings for perceptual loss, GAN loss, and L1 loss for different features, it significantly reduces artifact generation. Ma et al. proposed the SPSR model [32], which utilizes gradient information of images to guide the super-resolution process. The model employs a gradient branch to restore high-resolution gradient maps, thereby providing structural priors for the super-resolution process. Additionally, a gradient loss is introduced to impose a second-order constraint, which reduces geometric distortions while preserving image structure and details, resulting in more natural and realistic generated images. Li et al. proposed SeD [33], which incorporates semantic information of the image as a condition into the discriminator, encouraging the super-resolution network to learn a more refined distribution. These studies clearly demonstrate that incorporating additional conditional information significantly enhances the model’s understanding of image content. Overall, the introduction of conditional information provides new insights and methods for super-resolution techniques, driving further improvements in image reconstruction quality.

3. Methods

3.1. Network Architecture

To address the issue of inconsistent local texture reconstruction in images due to the lack of categorical semantic understanding, we propose a GAN model branch structure, SP-IGAN, that effectively utilizes semantic priors. Using Real-ESRGAN as the baseline model, we introduce additional semantic information as auxiliary input. The overall model structure is shown in Figure 1, and its core components include the Spatial Feature Transform (SFT) Layer, Graph Convolutional Channel Attention (GCCA), and Residual-in-Residual Dense Block (RRDB). The entire framework consists of two branches: the main branch and the auxiliary branch.
The main branch introduces the GCCA module, which converts channel dependencies into adjacency relationships between feature vertices, thus enhancing pixel correlations. Additionally, the auxiliary branch strengthens the dependence between class semantic information and regional texture within the RRDB module. Furthermore, we employ StarRRDB [21] to expand the network capacity of RRDB.
The auxiliary branch employs the pre-trained segmentation model OutdoorSceneSeg, which is used in the SFTGAN [29] model, to generate semantic maps. This model accurately extracts region-specific semantic information from the input low-resolution images and injects it into the RRDB module through the SFT layer. During this process, the Non-Local (NL) operation [26] is used to treat the feature maps of previous and subsequent layers as an adjacency matrix, establishing contextual dependencies. This allows the model to capture long-range contextual information, while the SFT layer injects additional semantic information, enhancing its utilization efficiency. The following sections will detail the specific design and implementation of the related modules.

3.2. The RRDB Structure with the SFT Layer

The RRDB module adopts a residual structure, using dense residual connections between layers to enhance feature reuse efficiency and model stability, while simultaneously reducing the number of parameters.
F R R D B = x + F ( x , θ )
where F ( x , θ ) represents the nonlinear transformation of low-resolution image x R C × H × W by the RRDB module, where θ denotes network parameters. Inspired by [30,31], the introduction of additional conditional information significantly enhances the model’s understanding of image content. We integrate the SFT layer into the RRDB module, where the SFT layer performs spatial transformations on features by introducing a pretrained semantic map, thereby improving the network’s ability to comprehend image content. The specific network architecture is shown in Figure 2. The SFT layer learns a mapping function f M ( · ) , which includes a set of parameters ( γ , β ) and semantic prior information Φ . Spatial affine transformations are applied to the intermediate features at each layer.
y = G θ ( x γ , β ) , ( γ , β ) = f M ( Φ )
Here, the prior information Φ is first used to construct a pair of affine transformation parameters M : Φ ( γ , β ) through the mapping function ( γ , β ) , where ( γ , β ) represents the scaling and translation factors, and the input variable x R C × H × W denotes the low-resolution image. The mapping function f M ( · ) serves to convert the prior information Φ into affine transformation parameters ( γ , β ) , which are then used for further affine transformation processing of the input image.
F S F T ( F R R D B γ , β ) = γ F R R D B + β
where ⊙ represents element-wise multiplication, where F R R D B is the intermediate feature to be modulated, γ R C × H × W and β R C × H × W are the two modulation parameters predicted by the auxiliary network. In the main branch, the dimensions of the feature map are aligned with parameters γ and β , and fusion is performed via element-wise multiplication. The global semantic features are then extracted through the feature extraction module ε .
F G l o b a l = ε ( x ; γ , β )
where ε ( · ) is a function that fuses the input image features with the modulation parameters, achieving both spatial transformation and feature fusion. Additionally, multiple SFT layers share parameters, which enhances the efficiency of semantic information transfer. This design allows for fine-grained control of features, enabling the network to better capture the structural and texture information of the image.
Figure 2. The SFT layer transforms semantic information through a mapping function. The auxiliary network generates auxiliary semantic information and shares it with all SFT layers to produce modulation parameters.
Figure 2. The SFT layer transforms semantic information through a mapping function. The auxiliary network generates auxiliary semantic information and shares it with all SFT layers to produce modulation parameters.
Entropy 27 00414 g002

3.3. The RRDB Module Integrated with GCCA Attention

Although star-shaped dense residual connections can more effectively learn features and stabilize the training of deep networks, the interdependencies between layers have not been fully exploited. After introducing semantic information, the network needs to better understand the relationships between image features, improving its representational and perceptual abilities. Inspired by the design in [34], we propose extending GCCA attention to the RRDB module, as shown in Figure 3. The GCCA module performs non-local (NL) operations [26] directly on the features, enabling better capture of long-range dependencies in the image, improving the understanding of global context, and assigning different weights to the feature map.

3.3.1. NL Operations

First, the definition of non-local (NL) [26] operations in CNNs is as follows:
f o u t ( a , b ) = ( a ˜ , b ˜ ) L 1 N ( a ˜ , b ˜ ) w ( a ˜ , b ˜ ) f i n ( a ˜ , b ˜ )
where f i n ( · ) and f o u t ( · ) represent the input and output feature maps, respectively, while ( a , b ) and ( a ˜ , b ˜ ) denote the horizontal and vertical coordinates of the pixels on the feature map. w ( a ˜ , b ˜ ) is the weight corresponding to the convolutional layer, representing the relationship between input and output. L is the 3 × 3 convolutional region for the NL operation, which includes ( a , b ) and eight adjacent pixels. Therefore, the output at each position aggregates all information related to ( a , b ) and its neighboring pixels. N ( · ) denotes the normalization operation applied to ( a ˜ , b ˜ ) .
Considering the intentional representation of spatial encoding in NL operations, the channel attention module utilizes the channel weights computed based on spatial information to adjust the response intensity of each channel. However, focusing on each corresponding pixel significantly increases computational complexity and leads to redundant calculations, which negatively affects the reconstruction results. To address this, we use Graph Convolutional Networks (GCNs) [35] to directly perform NL operations on the features. Therefore, the GCNs operation on the feature vertex i θ can be expressed as follows:
f o u t ( i a ) = i b C a 1 N ( i b ) W a b f i n ( i b )
where a and b represent the indices of the vertices in the feature map, while i denotes a vertex in the feature map. C a represents the convolutional sampling set of the output vertex i a and i b refers to the set of vertices adjacent to vertex i a . w a b is the corresponding weight vector, which represents the relationship between each output vertex i a and input vertex i b . N ( · ) is used to balance and normalize each subset.
This allows the model to focus more on the corresponding features rather than the pixels, effectively reducing feature redundancy. We employ a Graph Convolutional Module (GCCM) to capture the adjacency relationships of different features across layers, enabling NL operations on the features. This allows the construction of interdependencies between layers, enhancing the representational power of the features.

3.3.2. Graph Convolution Channel Attention (GCCA)

We integrate the GCCA module into each Basic Block, performing NL operations on the features directly after semantic information is extracted through the SFT layer in the SP-IGAN model. The GCCA module consists of two parts, as shown in Figure 3. The first part is a feature mapping layer formed by two linear embedding functions f ( · ) , while the second part is the Graph Convolutional Module (GCCM), which extracts feature weights W. The design is inspired by the bottleneck structure [36,37]. On one hand, this structure effectively enables channel interaction, generating interdependencies. On the other hand, it has been shown to reduce feature redundancy, thereby enhancing model performance. Consequently, the bottleneck structure replaces the convolutional layers, reducing the number of convolutional layers from C × C to C × ( C ω ) , compressing the adjacency matrix size to ( C ω ) × ( C ω ) , with a compression factor of ω , to reduce the parameter count.
The workflow of the GCCA module can be divided into several steps: First, global features from each channel of the input feature map f i n are extracted using Global Average Pooling (GAP) [38], compressing spatial dimension F C × H × W into the channel dimension F C × 1 × 1 , where C represents the number of channels, which is also the number of feature vertices i in the GCCA module. Next, the global features are used to generate channel weights W. The feature mapping layer consists of a linear embedding function f o u t = W f i n , where W is a learnable weight matrix. The mapping is performed using 1 × 1 convolution. The weight of each vertex is generated by the Graph Convolutional Module (GCCM), and these weights are then mapped back to the initial feature map f i n , as the weights for the corresponding channel C. Finally, the weight for each channel is calculated using the Sigmoid activation function. These channel weights, W, are element-wise multiplied with the original feature map f i n , to adjust each channel.
The introduction of the GCCA module enables the model to more effectively capture key features while suppressing irrelevant ones, thereby improving overall performance. The correlation and importance between channels have a significant impact on the final outcome. By adaptively adjusting the channel weights, the GCCA module significantly enhances the model’s expressive power and generalization ability. The formula for the GCCA module is as follows:
y = x · s i g m o i d F r R e L U ( G ( x , M ) ) x = F r ( G A P ( x ) )
The two embedding functions, F r ( · ) and F r ( · ) , form a bottleneck structure, where F r ( G A P ( x ) ) serves as the input to the GCCM, and M represents an adjacency matrix that captures the relationships between features. x and y denote the input and output feature maps of the GCCA module, respectively. Since each feature node in a GCN should be independent [39], the feature nodes in the GCCA module are particularly sensitive to feature redundancy. When two feature nodes are similar or identical, their relationships are often difficult to accurately express in the adjacency matrix. Therefore, the bottleneck structure plays a crucial role in the GCCA module. Meanwhile, the GCCM module can be viewed as guiding attention through the GCNs, enabling the module to adaptively learn the interdependencies between feature vertices, thereby more effectively capturing the correlations between nodes.

3.3.3. Graph Convolutional Channel Module (GCCM)

In GCNs, the feature map is mapped to F × N , where F represents the number of features per vertex and N denotes the number of vertices. Therefore, Equation (7) is rewritten as follows:
f o u t = W f i n A ˜
The matrix A ˜ of N × N represents the normalization operation on the adjacency matrix, which can be defined as follows: A ˜ = D 1 2 A D 1 2 , A ˜ is the adjacency matrix of N × N , and D denotes the degree matrix, defined as D a a = b A a b + ϵ , where ϵ is the small value used to prevent division by zero. A a b represents the connection between vertex a and vertex b. Due to the varying feature distributions across layers, the relationships between features are difficult to capture directly. Therefore, we adopt this adaptive, learnable graph convolution structure, which automatically learns the dependencies between feature vertices a b based on the data. Therefore, Equation (8) is rewritten as follows:
f o u t ( x ) = W f i n ( A 1 × A 2 + A 3 )
The adjacency matrix is composed of three parts: A A d j = { A 1 , A 2 , A 3 } ; A 1 represents the identity matrix of N × N , capturing the information of the feature vertex itself. From Equation (10), it follows that A ˜ 1 = D 1 2 A 1 D 1 2 = A 1 remains the identity matrix;
A 2 represents a diagonal matrix of N × N , where the weights for each feature vertex a b are calculated through the self-attention mechanism. The self-attention map enables the network to adaptively emphasize or suppress features. The calculation process of the self-attention map is as follows:
A 2 = T ( s o f t m a x ( W f in ) )
where T represents arranging the results into a diagonal matrix, and W denotes the weights of the one-dimensional convolutional layer;
A 3 represents the adjacency matrix of N × N , generating the relationships between any two feature vertices during training, and customizing the dependencies for various feature vertices contained in different layers.
The adjacency matrix A A d j = { A 1 , A 2 , A 3 } is used to describe the relationships between different feature groups in the convolutional neural network structure. These relationships are progressively constructed through adaptive learning. Subsequently, the adjacency matrix A A d j is used to update the feature vertices, and the updated feature vertices serve as channel weights, enabling precise calibration of channel-level responses. In the channel attention mechanism, this design precisely adjusts the importance of features, effectively capturing key features and significantly enhancing the model’s expressive power and performance.

3.4. Loss Functions

Most GAN-based super-resolution models typically adopt a strategy that combines L1 loss [6], perceptual loss [13], and GAN loss [12]. L1 loss improves the overall image fidelity by reducing the mean absolute error between the predicted and true image pixel values, as given by the following formula: L L 1 = 1 N i = 1 N y ^ i y i . Perceptual loss strengthens feature representations related to human perception by aligning features extracted from a pre-trained network, as given by the following formula: L p e r c e p t u a l = 1 M i = 1 M φ ( y ^ i ) φ ( y i ) 2 2 . GAN loss focuses on generating results that approximate the real image, significantly enhancing the realism of image textures. The formulas are as follows: Generator loss: L G = E x p z log D ( G ( z ) ) , Discriminator loss: L D = E x p d a t a log D ( x ) E x p z log ( 1 D ( G ( z ) ) ) . In real-world super-resolution, capturing more visually realistic images requires focusing on high-frequency details. Inspired by [40], we introduced wavelet loss into the loss function of the Real-ESRGAN model. Wavelet loss is evaluated by calculating the L1 fidelity difference between the generated and true images in the SWT subbands. The principle of wavelet loss is illustrated in Figure 4.
The specific steps for wavelet loss are as follows: Apply wavelet transform S W T ( · ) to the generated image G and the true image X, converting from the RGB space to the YCbCr space. Then, perform wavelet transform S W T ( · ) to decompose the Y channel into different frequency subbands, as expressed below:
S W T ( G ) = { G L , G H } S W T ( X ) = { X L , X H }
Here, S W T ( · ) decomposes the Y channel of the image into one low-frequency (LF) subband, denoted as G L = { G L L } , X L = { X L L } , and multiple high-frequency (HF) subbands, named G H = { G L H , G H L , G H H } , X H = { X L H , X H L , X H H } , collectively forming the output. The specific implementation of the wavelet loss function is as follows:
L S W T = E j λ j S W T ( G ( x ) ) j S W T ( X ) j 1
where G ( · ) represents the generator model of the proposed SP-IGAN structure, and λ j is the scaling factor that controls the high-frequency details generated. Therefore, our total loss function incorporates wavelet loss in addition to the original L 1 loss, perceptual loss, and GAN loss, and can be expressed as follows:
L = λ 1 · L 1 + λ p e r · L p e r + λ G A N · L G A N + λ S W T · L S W T
Traditional pixel-level loss functions are ineffective at capturing high-frequency details in images. By incorporating wavelet loss, the model’s ability to reconstruct high-frequency details is enhanced, addressing this limitation.

4. Experiments

This chapter first introduces the experimental setup, including the datasets and test sets used, implementation details, and the computation methods for evaluation metrics. It then analyzes the impact of the proposed SP-IGAN structure on the final reconstruction results under different classification priors. Subsequently, the SP-IGAN structure is compared with SOTA methods on the dataset, and finally, an ablation study is conducted on the model.

4.1. Experimental Setup

4.1.1. Datasets

To begin with, it is important to define high-resolution (HR) and low-resolution (LR) images. HR and LR are relative concepts, and the specific resolution values depend on the application scenario. High-resolution images typically refer to the original, unprocessed images, characterized by rich details, sharp edges, and a high pixel density. These images are usually the ones in the dataset and serve as the ground truth in training data, paired with low-resolution images. In contrast, low-resolution images are derived from high-resolution images through a degradation model, which may include downsampling, noise, blurring, or JPEG compression. The purpose of this process is to reduce the image details and resolution, simulating real-world degradation scenarios, and these low-resolution images are used as input for super-resolution models. They generally have fewer pixels and exhibit blurred details. Super-resolution tasks are often defined by a scaling factor (e.g., ×2, ×4, ×8), indicating the target magnification, and thus there is a certain multiplicative relationship between HR and LR images. We trained the model using the DF2K and OST datasets, where DF2K is a combined dataset consisting of DIV2K and Flickr2K. DIV2K contains 800 images, Flickr2K provides 2650 images, and the OST (Outdoor Scene Train/Test) dataset includes 300 high-resolution outdoor scene images. The pre-trained segmentation model used is OutdoorSceneSeg, which is employed in the SFTGAN [29] model. It is pre-trained on the COCO dataset and further fine-tuned on the ADE20K dataset. This segmentation model is capable of distinguishing 8 categories (7 object categories, including building, grass, plant, water, mountain, sky, and animal, along with 1 background category). For undefined categories, the model classifies them as “background”. The number of category labels can be flexibly adjusted according to the specific application scenario and task requirements. The test set includes three public benchmark datasets for single-image super-resolution tasks: DF2K, Set5, BSD100, and the outdoor dataset OST.

4.1.2. Implementation Details

The training process is similar to Real-ESRGAN, with high-resolution images set to 256 × 256 and a total batch size of 48. The SP-IGANNet model is first trained using L1 loss, with the L1 loss derived from the pre-trained ESRGAN model, for 100K iterations. The SP-IGANNet model is then initialized as the generator, and SP-IGAN is trained for 50K iterations using a combination of L1 loss, perceptual loss, GAN loss, and wavelet loss. We applied exponential moving average (EMA) to stabilize the training process and achieve better performance. Additionally, we visually enhanced the sharpness of the images using the Unsharp Masking (USM) method. During training, the real images are sharpened, which helps enhance image clarity while effectively suppressing artifacts caused by over-enhancement. The model is trained using the sharpened real images. The entire process was conducted on an RTX 3060 GPU with a batch size of 2, a convolution kernel size of 3 × 3, and a stride of 1. The Adaptive Moment Estimation (Adam) optimizer was used, with an initial learning rate of 0.0002 and an exponential moving average decay factor of 0.999. The model was trained under the PyTorch (https://pytorch.org/) framework.

4.1.3. Evaluation Index

In image super-resolution, Peak Signal-to-Noise Ratio (PSNR) is commonly used as a metric to evaluate the quality of image reconstruction. A higher PSNR value indicates better image quality. PSNR is calculated based on the Mean Squared Error (MSE) of the image, as shown in the following formula:
P S N R = 10 · log 10 R 2 M S E
where MSE represents the difference between the predicted and ground truth values of an image, and is calculated using the following formula:
M S E = 1 N i = 1 N y ^ i y i 2
The Structural Similarity Index (SSIM) is a structure-based image quality metric designed to simulate structural similarity in human visual perception. SSIM measures visual information, including pixel differences, luminance, contrast, and structure, and better aligns with human visual perception compared to PSNR. The formula is as follows:
S S I M ( x , y ) = ( 2 μ x μ y + C 1 ) ( 2 σ x y + C 2 ) ( μ x 2 + μ y 2 + C 1 ) ( σ x 2 + σ y 2 + C 2 )
The value of SSIM ranges from 0 to 1, where a value closer to 1 indicates greater similarity between two images and higher image quality.

4.2. The Effect of Low-Resolution Images on Segmentation Models

In super-resolution research, low-resolution (LR) images are obtained by downsampling high-resolution (HR) images. Therefore, in our experiments, we need to consider whether there is a significant difference between the semantic maps derived from HR and LR images through the segmentation model, or if low-resolution images are too affected to provide accurate results, as shown in Figure 5. We found that even at this low resolution, the semantic maps obtained through segmentation are not significantly affected. The segmentation results from LR and HR images are very similar, but this is mainly true for larger and more distinct objects; smaller objects were not tested.

4.3. Effect of Different Priors on the Reconstruction Results

In the experiment, we need to analyze whether prior information influences the output reconstruction results. As shown in Figure 6, under the influence of different categories of prior information, a given LR image can be reconstructed to match the texture corresponding to the guided prior. We can observe that for the water category in the lake, when guided by prior information from buildings and grass, the reconstructed texture matches that of the corresponding category. Similarly, guidance from prior information of grass and buildings also leads to the corresponding textures being generated. Regardless of the input category, the prior for the building category always results in the generation of a regular geometric texture. This phenomenon demonstrates the importance of semantic priors in reconstructing real textures in single-image super-resolution tasks. In the absence of semantic priors, the model may produce reconstructions that are inconsistent with the real texture.

4.4. Comparison with State-of-the-Art Methods

4.4.1. Quantitative Results

We compared the SSIM and PSNR metrics to evaluate the model’s performance, as shown in Table 1. Although the PSNR metric measures image reconstruction quality to some extent, it cannot fully reflect human perceptual preferences. Therefore, we also computed the SSIM metric, which better reflects the subjective quality of human visual perception. Compared to recent GAN-based super-resolution models, our model demonstrates significant superiority. For example, on the ×4 super-resolution of the DF2K test set, our model achieved a 0.55 dB improvement in PSNR over the baseline Real-ESRGAN model, and a 0.0363 improvement in SSIM, validating the effectiveness of our proposed solution. Although Real-ESRGAN [19] and IG-CFAT [40] excel in restoring high-frequency details, they tend to generate unnatural textures, making it difficult to ensure the accuracy of the generated textures. In contrast, our approach introduces a semantic prior conditional network, providing features for each category to the generation network, making the generated textures more natural and realistic. Compared to the SFTGAN [29] model, which also uses semantic priors, our model shows a 2.22 dB improvement in PSNR and a significant advantage in SSIM, with an increase of 0.1308.

4.4.2. Qualitative Results

To demonstrate the performance of our model, Figure 7 presents a comparative analysis with several state-of-the-art (SOTA) models in real-world image super-resolution, including Real-ESRGAN [19], SFTGAN [29], BSRGAN [17], IG-CFAT [40], and ESRGAN [14]. A comprehensive comparison was conducted on the performance of these models under ×2, ×4, and ×8 super-resolution scenarios. All experiments were conducted on the same dataset, DIV2K, and the qualitative results are presented in Figure 7. As observed, compared to other models, our model reconstructs details more naturally and realistically, preserving texture characteristics closer to those of real images. The first three images depict scenes with architectural texture details, where the model accurately restores design details and corresponding textures, producing structures highly similar to the original images. In the last image, the model also clearly reconstructs the soldier’s features, further demonstrating its superiority in fine-detail reconstruction.

4.5. Ablation Study

To better assess the effectiveness of SP-IGAN, we conducted ablation experiments on the StarRRDB connection, GCCA module, SFT layer, and wavelet loss L S W T . The evaluation was performed using the DIV2K test set in the ×4 super-resolution scenario. As shown in the results of Table 2, when the GCCA module was used, the PSNR, which focuses on objective quality, improved by 0.33, while the SSIM, which reflects subjective quality, increased by 0.021. This indicates that the introduction of the module contributes effectively to the restoration of image texture details in the reconstruction. With the gradual incorporation of StarRRDB, the GCCA module, the SFT layer, and wavelet loss L S W T into the model, both PSNR and SSIM values steadily improved. After introducing the GCCA module, the model can reallocate weights to semantically relevant parts, thereby focusing more on semantic texture details. The introduction of the SFT layer transfers semantically relevant conditional information to the super-resolution network, enhancing the realism of the generated textures. As shown in Figure 7, despite the blurry boundary between the buildings and grass, the model is still able to capture semantic information to some extent. Furthermore, after incorporating wavelet loss L S W T , the model focuses more on high-frequency details, compensating for the traditional loss function’s inadequacy in capturing these details. Overall, the introduction of these modules significantly enhances the model’s ability to capture details and improves the quality of the generated images, as validated by the experimental results in the table.

5. Conclusions

We propose a GAN-based single-image super-resolution method, SP-IGAN. By integrating the SFT layer and GCCA module into the dense residual network (RRDB) module, SP-IGAN effectively captures and extracts image features. Additionally, by incorporating extra semantic information, the model further aggregates contextual data, enhancing its ability to comprehend image structure. The GCCA module redistributes pixel weights, reducing computational redundancy, while the SFT layer strengthens the model’s dependence on the relationship between texture and category information. Moreover, the inclusion of wavelet loss L S W T in the total loss calculation enhances the recovery and reconstruction of high-frequency details in the image, further strengthening the network’s representation capability. Extensive experiments demonstrate that SP-IGAN outperforms state-of-the-art methods across multiple public datasets, achieving superior PSNR and SSIM scores. It significantly improves texture consistency in region-specific categories and enhances subjective visual quality in single-image super-resolution reconstruction.
Future work can further explore single-image super-resolution methods that incorporate prior information, such as utilizing multimodal priors by integrating text descriptions into the super-resolution model. Considering the varying demands for super-resolution images in different scenarios, we can also train the model using domain-specific datasets, which would significantly enhance the performance of single-image super-resolution in complex scenarios. This study optimizes the model by improving the generator network, and it also explores an alternative approach by working with the discriminator. By introducing a semantic information extraction module, the authenticity of super-resolution images can be better assessed, and distributed learning can help improve the understanding of semantic structures. This approach not only reduces the computational cost of the generative model but also makes the model more lightweight and easier to integrate. This direction is worth further investigation.

Author Contributions

Conceptualization, Z.L., Z.C. and K.C.; methodology, Z.L. and Z.C.; software, Z.L., Z.C. and K.C.; validation, Z.L., M.W. and K.C.; formal analysis, H.L.; investigation, M.W.; resources, M.W.; data curation, M.W.; writing—original draft preparation, Z.L.; writing—review and editing, Z.L., M.W. and K.C.; visualization, Z.L.; supervision, M.W. and H.L.; project administration, M.W.; funding acquisition, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (62062048, 62263017), and the Yunnan Department of Science and Technology Project (202201AT070113).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

We thank all the editors and reviewers in advance for their valuable comments that will improve the presentation of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Al-Mekhlafi, H.; Liu, S. Single image super-resolution: A comprehensive review and recent insight. Front. Comput. Sci. 2024, 18, 181702. [Google Scholar] [CrossRef]
  2. Wang, P.; Bayram, B.; Sertel, E. A comprehensive review on deep learning based remote sensing image super-resolution methods. Earth Sci. Rev. 2022, 232, 104110. [Google Scholar] [CrossRef]
  3. Azam, N.; Yazid, H.; Rahim, S.A. Super Resolution with Interpolation-based Method: A Review. IJRAR Int. J. Res. Anal. Rev. (IJRAR) 2022, 9, 168–174. [Google Scholar]
  4. Ooi, Y.K.; Ibrahim, H. Deep Learning Algorithms for Single Image Super-Resolution: A Systematic Review. Electronics 2021, 10, 867. [Google Scholar] [CrossRef]
  5. Fu, K.; Peng, J.; Zhang, H.; Wang, X.; Jiang, F. Image super-resolution based on generative adversarial networks: A brief review. Comput. Mater. Contin. 2020, 64, 1977–1997. [Google Scholar] [CrossRef]
  6. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef] [PubMed]
  7. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  8. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  9. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  10. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  11. Li, X.; Ren, Y.; Jin, X.; Lan, C.; Wang, X.; Zeng, W.; Wang, X.; Chen, Z. Diffusion Models for Image Restoration and Enhancement—A Comprehensive Survey. arXiv 2023, arXiv:2308.09388. [Google Scholar]
  12. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Advances in Neural Information Processing Systems 27 (NIPS 2014); MIT Press: Cambridge, MA, USA, 2014. [Google Scholar]
  13. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  14. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Loy, C.C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the 2018 European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  15. Zhang, W.; Liu, Y.; Dong, C.; Qiao, Y. RANKSRGAN: Generative Adversarial Networks with Ranker for Image Super-Resolution. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 3096–3105. [Google Scholar]
  16. Schonfeld, E.; Schiele, B.; Khoreva, A. A U-Net Based Discriminator for Generative Adversarial Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8207–8216. [Google Scholar]
  17. Zhang, K.; Liang, J.; Van Gool, L.; Timofte, R. Designing a Practical Degradation Model for Deep Blind Image Super-Resolution. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4791–4800. [Google Scholar]
  18. Fritsche, M.; Gu, S.; Timofte, R. Frequency Separation for Real-World Super-Resolution. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 3599–3608. [Google Scholar]
  19. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  20. Wang, L.; Wang, Y.; Dong, X.; Xu, Q.; Yang, J.; An, W.; Guo, Y. Unsupervised Degradation Representation Learning for Blind Super-Resolution. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10581–10590. [Google Scholar]
  21. Vo, K.D.; Bui, L.T. StarSRGAN: Improving Real-World Blind Super-Resolution. arXiv 2023, arXiv:2307.16169. [Google Scholar]
  22. Rakotonirina, N.C.; Rasoanaivo, A. ESRGAN+: Further Improving Enhanced Super-Resolution Generative Adversarial Network. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 3637–3641. [Google Scholar]
  23. Korkmaz, C.; Tekalp, A.M. Training Transformer Models by Wavelet Losses Improves Quantitative and Visual Performance in Single Image Super-Resolution. arXiv 2024, arXiv:2404.11273. [Google Scholar]
  24. Niu, Z.; Zhong, G.; Yu, H. A Review on the Attention Mechanism of Deep Learning. Neurocomputing 2021, 452, 48–62. [Google Scholar]
  25. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  26. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  27. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the 2018 European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  28. Dai, T.; Cai, J.; Zhang, Y.; Xia, S.T.; Zhang, L. Second-Order Attention Network for Single Image Super-Resolution. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11065–11074. [Google Scholar]
  29. Wang, X.; Yu, K.; Dong, C.; Loy, C.C. Recovering Realistic Texture in Image Super-Resolution by Deep Spatial Feature Transform. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 606–615. [Google Scholar]
  30. Park, S.H.; Moon, Y.S.; Cho, N.I. Perception-Oriented Single Image Super-Resolution Using Optimal Objective Estimation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1725–1735. [Google Scholar]
  31. Park, S.H.; Moon, Y.S.; Cho, N.I. Flexible Style Image Super-Resolution Using Conditional Objective. IEEE Access 2022, 10, 9774–9792. [Google Scholar] [CrossRef]
  32. Ma, C.; Rao, Y.; Cheng, Y.; Chen, C.; Lu, J.; Zhou, J. Structure-preserving super resolution with gradient guidance. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7769–7778. [Google Scholar]
  33. Li, B.; Li, X.; Zhu, H.; Jin, Y.; Feng, R.; Zhang, Z.; Chen, Z. SeD: Semantic-Aware Discriminator for Image Super-Resolution. arXiv 2024, arXiv:2402.19387. [Google Scholar]
  34. Xiang, X.; Wang, Z.; Zhang, J.; Xia, Y.; Chen, P.; Wang, B. AGCA: An Adaptive Graph Channel Attention Module for Steel Surface Defect Detection. IEEE Trans. Instrum. Meas. 2023, 72, 5008812. [Google Scholar] [CrossRef]
  35. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating More Pixels in Image Super-Resolution Transformer. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22367–22377. [Google Scholar]
  36. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2015), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  37. Zhou, D.; Hou, Q.; Chen, Y.; Feng, J.; Yan, S. Rethinking Bottleneck Structure for Efficient Mobile Network Design. In Computer Vision—ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part III 16; Springer International Publishing: Cham, Switzerland, 2020; pp. 680–697. [Google Scholar]
  38. Lin, M. Network in Network. arXiv 2013, arXiv:1312.4400. [Google Scholar]
  39. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  40. Aghelan, A.; Amiryan, A.; Zarghani, A.; Rouhani, M. IG-CFAT: An Improved GAN-Based Framework for Effectively Exploiting Transformers in Real-World Image Super-Resolution. arXiv 2024, arXiv:2406.13815. [Google Scholar]
Figure 1. The overall structure of SP-IGAN consists of three core modules: Spatial Feature Transform (SFT) Layer, Graph Convolutional Channel Attention (GCCA), and Residual-in-Residual Dense Block (RRDB). Additionally, a star-shaped residual connection is used to provide greater flexibility for future extensions.
Figure 1. The overall structure of SP-IGAN consists of three core modules: Spatial Feature Transform (SFT) Layer, Graph Convolutional Channel Attention (GCCA), and Residual-in-Residual Dense Block (RRDB). Additionally, a star-shaped residual connection is used to provide greater flexibility for future extensions.
Entropy 27 00414 g001
Figure 3. The framework of GCCA extended to the RRDB module consists of two parts. The first part involves the mapping operations represented by F r and F r , where ⊙ denotes element-wise multiplication. The second part is the GCCM (Graph Convolutional Channel Module), where T represents the diagonal arrangement result.
Figure 3. The framework of GCCA extended to the RRDB module consists of two parts. The first part involves the mapping operations represented by F r and F r , where ⊙ denotes element-wise multiplication. The second part is the GCCM (Graph Convolutional Channel Module), where T represents the diagonal arrangement result.
Entropy 27 00414 g003
Figure 4. The structure diagram of the discriminator, which includes wavelet loss L S W T .
Figure 4. The structure diagram of the discriminator, which includes wavelet loss L S W T .
Entropy 27 00414 g004
Figure 5. Examples of segmentation are shown as follows: Row 1: Ground Truth (GT) image; Row 2: Segmentation result of GT; Row 3: low-resolution (LR) image; Row 4: segmentation result of LR.
Figure 5. Examples of segmentation are shown as follows: Row 1: Ground Truth (GT) image; Row 2: Segmentation result of GT; Row 3: low-resolution (LR) image; Row 4: segmentation result of LR.
Entropy 27 00414 g005
Figure 6. Reconstruction results under priors from different categories.
Figure 6. Reconstruction results under priors from different categories.
Entropy 27 00414 g006
Figure 7. Visual comparison of SP-IGAN ×4 and other super-resolution methods.
Figure 7. Visual comparison of SP-IGAN ×4 and other super-resolution methods.
Entropy 27 00414 g007
Table 1. A quantitative comparison is conducted between the SP-IGAN model and recent SOTA super-resolution methods, most of which are based on GANs. The model with the best performance is highlighted in red, while the second-best is marked in blue.
Table 1. A quantitative comparison is conducted between the SP-IGAN model and recent SOTA super-resolution methods, most of which are based on GANs. The model with the best performance is highlighted in red, while the second-best is marked in blue.
Model DF2KOST300Set5BSD100
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
ESRGAN [14] 25.450.757024.870.722126.100.778123.220.7064
SFTGAN [29] 25.530.798125.800.788126.820.797024.530.7153
Real-ESRGAN [19] 26.880.856926.630.855327.550.822524.400.7521
BSRGAN [17]×226.210.859126.010.862126.120.837924.390.7560
FxSR-PD [31] 27.500.881127.200.875527.640.853025.360.7890
IG-CFAT [40] 26.020.852025.450.781326.910.802424.110.7559
ours 27.220.895127.310.860027.740.856126.010.7611
ESRGAN [14] 22.560.655322.220.684423.650.684920.890.6281
SFTGAN [29] 22.980.725522.820.712524.710.705121.740.6520
Real-ESRGAN [19] 24.650.820025.100.825125.470.809123.510.7222
BSRGAN [17]×424.050.801224.350.841224.990.753822.970.6588
FxSR-PD [31] 25.150.832025.220.850525.660.829523.690.7583
IG-CFAT [40] 24.190.799823.950.776424.220.787322.590.7023
ours 25.200.856325.180.852025.510.846023.900.7619
ESRGAN [14] 22.160.646321.350.597921.850.625120.660.5830
SFTGAN [29] 22.750.700723.350.689624.010.686821.120.6258
Real-ESRGAN [19] 24.560.789023.990.803625.030.773523.050.7272
BSRGAN [17]×823.850.778124.030.789024.470.753822.970.6588
FxSR-PD [31] 24.950.809725.120.812125.060.805523.280.7351
IG-CFAT [40] 23.910.776923.850.753424.100.762422.310.6821
ours 24.820.821124.910.809225.350.835023.580.7400
Table 2. Ablation experiments of SP-IGAN in the ×4 super-resolution model. The model with the best performance is highlighted in red, while the second-best is marked in blue.
Table 2. Ablation experiments of SP-IGAN in the ×4 super-resolution model. The model with the best performance is highlighted in red, while the second-best is marked in blue.
StarRRDBGCCASFT L SWT PSNR ↑SSIM ↑
1××××24.650.8200
2×××24.680.8211
3××25.010.8421
4×25.100.8489
525.200.8563
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, M.; Li, Z.; Liu, H.; Chen, Z.; Cai, K. SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution. Entropy 2025, 27, 414. https://doi.org/10.3390/e27040414

AMA Style

Wang M, Li Z, Liu H, Chen Z, Cai K. SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution. Entropy. 2025; 27(4):414. https://doi.org/10.3390/e27040414

Chicago/Turabian Style

Wang, Meng, Zhengnan Li, Haipeng Liu, Zhaoyu Chen, and Kewei Cai. 2025. "SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution" Entropy 27, no. 4: 414. https://doi.org/10.3390/e27040414

APA Style

Wang, M., Li, Z., Liu, H., Chen, Z., & Cai, K. (2025). SP-IGAN: An Improved GAN Framework for Effective Utilization of Semantic Priors in Real-World Image Super-Resolution. Entropy, 27(4), 414. https://doi.org/10.3390/e27040414

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop