Next Article in Journal
An Investigation of ECAPA-TDNN Audio Type Recognition Method Based on Mel Acoustic Spectrograms
Next Article in Special Issue
Micro-Expression Spotting Based on VoVNet, Driven by Multi-Scale Features
Previous Article in Journal
S2AC: Self-Supervised Attention Correlation Alignment Based on Mahalanobis Distance for Image Recognition
Previous Article in Special Issue
Optimal Scheduling of Emergency Materials Based on Gray Prediction Model under Uncertain Demand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PSRGAN: Perception-Design-Oriented Image Super Resolution Generative Adversarial Network

1
National Model Software Institute, Huazhong University of Science and Technology, Wuhan 430074, China
2
PSS Lab of Big Data and National Communication Strategy, MOE, Huazhong University of Science and Technology, Wuhan 430074, China
3
Hytera Communication Co., Ltd., Shenzhen 518057, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2023, 12(21), 4420; https://doi.org/10.3390/electronics12214420
Submission received: 14 September 2023 / Revised: 13 October 2023 / Accepted: 17 October 2023 / Published: 27 October 2023
(This article belongs to the Special Issue Applications of Computational Intelligence, Volume 2)

Abstract

:
Among recent state-of-the-art realistic image super-resolution (SR) intelligent algorithms, generative adversarial networks (GANs) have achieved impressive visual performance. However, there has been the problem of unsatisfactory perception of super-scored pictures with unpleasant artifacts. To address this issue and further improve visual quality, we proposed a perception-design-oriented PSRGAN with double perception turbos for real-world SR. The first-perception turbo in the generator network has a three-level perception structure with different convolution kernel sizes, which can extract multi-scale features from four  1 4  size sub-images sliced by original LR image. The slice operation expands adversarial samples to four and could alleviate artifacts during GAN training. The extracted features will be eventually concatenated in later 3 × 2 upsampling processes through pixel shuffle to restore SR image with diversified delicate textures. The second-perception turbo in discriminators has cascaded perception turbo blocks (PTBs), which could further perceive multi-scale features at various spatial relationships and promote the generator to restore subtle textures driven by GAN. Compared with recent SR methods (BSRGAN, real-ESRGAN, PDM_SR, SwinIR, LDL, etc.), we conducted an extensive test with a  × 4  upscaling factor on various datasets (OST300, 2020track1, RealSR-Canon, RealSR-Nikon, etc.). We conducted a series of experiments that show that our proposed PSRGAN based on generative adversarial networks outperforms current state-of-the-art intelligent algorithms on several evaluation metrics, including NIQE, NRQM and PI. In terms of visualization, PSRGAN generates finer and more natural textures while suppressing unpleasant artifacts and achieves significant improvements in perceptual quality.

1. Introduction

Single-image super-resolution (SISR) aims to reconstruct a high-resolution (HR) image from a low-resolution (LR) one. The traditional methods for solving the SR problems are mainly interpolation-based methods [1,2,3,4] and reconstruction-based methods [5,6,7]. Intelligent computing has also been applied in the field of image super-resolution. Super-resolution methods based on genetic algorithms, guided by imaging models, utilize optimization techniques to seek the optimal estimation of the original image. At its core, this approach transforms the problem of reconstructing multiple super-resolved images into a linear system of equations. The convolutional neural network (CNN) has greatly promoted the vigorous development of SR field and demonstrates vast superiority over traditional methods. The main reason it achieves good results is due to its strong capability of learning rich features from big data in an end-to-end manner [8]. CNN-based SR methods often use PSNR as the evaluation metric; although some SR methods achieve good results for PSNR, it is still not completely satisfactory in terms of perception.
The generative adversarial network (GAN) [9] has achieved impressive visual performance in the field of super-resolution (SR) since the pioneering work of SRGAN [10]. GANs have proven their capability to generate more realistic images with high perceptual quality. In pursuit of further enhancing visual quality, Wang et al. proposed ESRGAN [11]. Given the challenges of collecting well-paired datasets in real-world scenarios, unsupervised GANs have been introduced [12,13]. BSRGAN [14] and real-ESRGAN [15] are dedicated to simulating the practical degradation process to obtain better visual results on real datasets.
However, perceptual dissatisfaction accompanied by unpleasant artifacts still exists in GAN-based SR models because of insufficient design in either generators or discriminators. In GAN-based SR methods, it is obvious that the decisive capability to recover naturally finer textures in generators is dependent largely on the guidance of discriminators through GAN training, but discriminators are usually cloned from well-known networks (U-net [16], VGG [17], etc.) suitable for image segmentation or classification, which might not fully lead generators to restore subtle textures in SR. Moreover, the design of generators should be perceptive enough to extract multi-scale image features from low-resolution (LR) images and mitigate artifacts.
Research hypotheses and questions: Perceived quality improvement: How can we design a network structure of PSRGAN to suppress artifact generation in images, and how can we achieve the effect of suppressing artifacts? Generative adversarial network image quality assessment: Which evaluation metrics are used to assess the generated images to ensure their perceived quality is enhanced? Adversarial training stability: How can we ensure the stability and convergence of our PSRGAN training? To address these issues and further improve the visual quality of the restored SR images, we redesigned both generators and discriminators; the contributions of this paper are mainly in four aspects:
  • We present a novel perception-design-oriented PSRGAN with double perception turbos, which can generate real-world SR images with naturally finer textures while suppressing unpleasant artifacts by  × 4  upscaling factors (see Figure 1).
  • We design the first-perception turbo in the generator network, characterized by slice operation and a three-level perception structure, which can extract multi-scale features from sliced sub-images and mitigate artifacts.
  • We propose the second-perception turbo in the discriminator network with cascaded perception turbo blocks, which can further promote the generator to restore subtle textures.
  • We demonstrate that the proposed PSRGAN has achieved state-of-the-art perceptual capabilities calculated by NIQE, NRQM, and PI.

2. Related Work

Single-image super-resolution: SRCNN [18] is the first method to apply deep learning to SR reconstruction, and a series of learning-based works are subsequently proposed [19,20,21,22,23]. ESPCN [24] introduces an efficient sub-pixel convolution layer to perform the feature extraction stages in the LR space instead of HR space. VDSR [19] uses a very deep convolutional network. EDSR [25] removes the batch normalization layers from the network. SRGAN [10] first uses the GAN network for the SR problem and proposes perceptual loss, including adversarial loss and content loss. Based on human perceptual characteristics, the residual in the residual dense block strategy (RRDB) is exploited to implement various depths in network architectures [11,26]. ESRGAN [11] introduces the residual-in-residual dense block (RRDB) into the generator. RealSR [27] estimates various blur kernels and real noise distributions to synthesize different LR images. CDC [28] proposes a divide-and-conquer SR network. Luo et al., in [29], propose a probabilistic degradation model (PDM). Shao et al., in [30], propose a sub-pixel convolutional neural network (SPCNN) for image SR reconstruction.
Perceptual-driven approaches: The PSNR-oriented approaches lead to overly smooth results and a lack of high-frequency details, and the results sometimes do not agree with the subjective human perception. In order to improve the perceptual quality of SR results, the perceptual-driven approach is proposed. Based on the idea of perceptual similarity [31], Li Feifei et al. propose perceptual loss in [32]. Then, textures matching loss [33] and contextual loss [34] are introduced. ESRGAN [11] improves the perceptual loss by using the features before activation and wins the PIRM perceptual super-resolution challenge [35]. Christian Szegedy et al. propose inception [36], which can extract more features with the same amount of computation, thus improving the training results. For the purpose of extracting multi-scale information and enhance the feature discriminability, RFB-ESRGAN [8] applies the receptive field block (RFB) [37] to super resolution and wins the NTIRE 2020 perceptual extreme super-resolution challenge. There is still plenty of room for perceptual quality improvement [38].
The design of discriminator networks: The discriminator in SRGAN is VGG-style, which is trained to distinguish between SR images and GT images [10]. ESRGAN borrows ideas from relativistic GAN to improve the discriminator in SRGAN [11]. Real-ESRGAN improves the VGG-style discriminator in ESRGAN to an U-Net design [15]. In [39], Alejandro et al. propose a novel convolutional network architecture named “stacked hourglass”, which captures and consolidates information across all scales of the image. Inspired by [39], we propose a new discriminator structure, which can guide the generator to recover finer textures. All the related work as Table 1 shows.
Artifact suppression: The instability of the training of GANs often leads to the introduction of many perceptually unpleasant artifacts while generating details in the GAN-based SR networks [40]. There have been several SR models focusing on solving the problem. Zhang et al. propose a supervised pixel-wise generative adversarial network (SPGAN) to obtain higher-quality face images [41]. Gong et al., in [42], overcome the effect of artifacts in the super-resolution of remote sensing images using self-supervised hierarchical perceptual loss. Real-ESRGAN uses spectral normalization (SN) regularization to stabilize the training dynamics [15]. We propose a algorithm named “image slice and multi-scale feature extraction”, which can generate more delicate textures and suppress artifacts.
The evaluation metrics: The DCNN-based SR approaches have two main optimization objectives: the distortion metric (e.g., PSNR, SSIM, IFC, and VIF [43,44,45]) and perceptual quality (e.g., the human opinion score; no-reference quality measures such as Ma’s score [46], NIQE [47], BRISQUE [48], and PI [49]) [50]. Yochai et al. in [49] have revealed that distortion and perceptual quality are contradictory and there is always a trade-off between the two. Algorithms that are superior in terms of perceptual quality tend to be poorer in terms of, e.g., PSNR and SSIM. However, sometimes there is also inconsistency between the results observed by human eyes and these perceptual quality metrics. Because the no-reference metrics do not always match perceptual visual quality [51], some SR models such as SRGAN perform mean-opinion-score (MOS) tests to quantify the perceptual ability of different methods [10]. We use NIQE, NRQM, and PI as our image quality metrics, which do not depend on the GT image to measure the perceptual quality of the reconstructed image [52]. The related work on evaluation metrics as Table 2 shows.
The transformer: Vaswani et al. in [36] propose a new simple network architecture, transformer, based solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Transformer continues to show amazing capabilities in the NLP domain. Many researches have started to try to apply the powerful modeling ability of transformer to the field of computer vision [53]. In [54], Yang et al. propose TTSR, in which LR and HR images are formulated as queries and keys in transformer, respectively, to encourage joint feature learning across LR and HR images. Swin transformer [55] combines the advantages of convolution and transformer. Liang et al. in [56] propose SwinIR based on Swin transformer. Vision transformer is computationally expensive and consumes high GPU memory, so Lu et al. in [57] propose ESRT, which uses efficient transformers (ET), a lightweight version of the transformer structure.

3. Proposed Methods

To further improve perceptual quality as well as mitigate artifacts in SISR, we proposed a novel perception-design-oriented super resolution generative adversarial network (PSRGAN) with double perception turbos. In this section, we first introduce the generator network-containing first-perception turbo (GPT) and then describe the construction of the discriminator network with the second-perception turbo (DPT). At last, we discuss the perceptual loss function used.

3.1. Generator Network

The generator network consists of two components: first-perception turbo, and the feature blending and upsampling component (FBUC) as shown in Figure 2.
The first perception turbo has two major blocks: the image slice block (ISB) and the multi-scale feature-extraction block (MFEB). The image slice block (ISB) produces four  1 4  size sub-images ( I s u b 1 I s u b 2 I s u b 3 , and  I s u b 4 ) from the low-resolution image  I L R  via pixel reassembly. Specifically, suppose  I L R  has the resolution of  2 m · 2 n  pixels or padding to  2 m · 2 n  pixels; the sliced sub-images are  m · n  pixels. If the upper left pixel is denoted as  0 , 0 , and the lower right pixel is denoted as  2 m 1 , 2 n 1 , the relationship of the pixels between  I L R  and the sub-images can be formulated as below.
I L R = k , t 0 k < 2 m , 0 t < 2 n , k , t N 0 I s u b 1 = 2 p , 2 q 0 p < m , 0 q < n , p , q N 0 I s u b 2 = 2 p + 1 , 2 q 0 p < m , 0 q < n I s u b 3 = 2 p , 2 q + 1 0 p < m , 0 q < n I s u b 4 = 2 p + 1 , 2 q + 1 | 0 p < m , 0 q < n
The slice method above has the following characteristics:
  • The slice splits the LR image to multiple detail adversarial sub-images while preserving the pixel integrity of the LR image.
  • The subsequent MFEB could extract multi-scale features from smaller adversarial samples; thus, the generator is capable of generating diverse and delicate textures.
  • The slice weakens the correlations among noisy pixels in  I L R , which can effectively reduce noises and further alleviate artifacts in the restored SR image. Although the correlations among adjacent pixels might be also impaired, the meaningful semantic features will be eventually recovered in the SR image through GAN training.
The multi-scale feature extraction block (MFEB, Figure 3): It has been proven that each learned filter has its specific functionality and that a reasonably larger filter size could grasp richer structural information, which in turn could lead to better results [18]. The MFEB is perceptually designed to extract diverse image features from the LR image by three groups of convolutional layers inspired by inception networks [36], as depicted in Figure 3. Please refer to Appendix B for more detail.
The first convolution group has a tiny receptive field, used to retain micro subtle features, denoted as k1-n64-s1.
The second convolution group has a medium receptive field, used to capture moderate features, denoted as k1-n32-s1, k3-n64-s1.
The third convolution group has a large receptive field, used to seize macro features, denoted as k1-n32-s1, k3-n48-s1, k3-n64-s1.
The outputs of the three convolution groups are activated using the Sigmoid weighted liner unit (SiLU) and then  × 2  upsampled via pixel-shuffle to obtain multi-scale features  F 1 , F 2 , F 3 . The process can be formulated as:
F i = S i L U C o n v s i x s u b s , i { 1 , 2 , 3 } .
where  C o n v s i x s u b , i { 1 , 2 , 3 }  denote the three convolution groups, SiLU is the activation function, ↑ denotes upsampling, s denotes the scale factor and  s = 2  in this block, and  F i , i { 1 , 2 , 3 }  indicate the 3-scale feature maps extracted. Subsequently, the obtained feature maps  F 1 , F 2 , F 3  are added in the channel dimension as input, residual in residual dense block (RRDB) [11] is adopted to further capture semantic information and improve the recovered textures, and the output is denoted as F. The formal processing in the first-perception turbo is described in Algorithm 1.
Algorithm 1 Image slice and multi-scale feature extraction
  • Input: LR images set  X .
  • Output: Multi-scale features  F 1 , F 2 , F 3 , deeper features F.
  •  1: for all  I L R  such that  I L R X  do
  •  2:     generate  I s u b 1 , I s u b 2 , I s u b 3 , I s u b 4  through slice operation from  I L R .
  •  3:     Get  x s u b  by merging the four sub-images  I s u b 1 , I s u b 2 , I s u b 3 , I s u b 4  in color channel dimension.
  •  4:     for all i such that  1 i 3  do
  •  5:         input  x s u b  to  C o n v s i , S i L U , 2 U P  obtain  F i ,
  •  6:     end for
  •  7:     generate  F = R R D B ( F 1 + F 2 + F 3 ) ,
  •  8: end forreturn  F 1 , F 2 , F 3 , F .
Feature blending and upsampling component (FBUC, Figure 2): The FBUC reassembles the obtained multi-scale features to generate the corresponding  I S R  counterpart of  I L R . In the upsampling phase, the FBUC upsamples  I L R  with diversfied features F as the input via pixel shuffle and gradually blends the features extracted by the MFEB. The upsampling process can be formulated as follows:
F f i n a l = f C o n v S i L U ( f C o n v S i L U ( f C o n v S i L U ( ( F + F 3 ) s + ( F 2 ) s ) + ( F 1 ) s ) s )
where ‘ + ’ denotes concatenation operation, ↑ denotes upsampling, s denotes the scale factor, and  s = 2 f C o n v S i L U  denotes one convolutional kernel, SiLU is the activation function, and  F f i n a l  denotes the final features obtained from the FBUC.  F f i n a l  is passed through a triple convolutional layer with the kernel size of  3 × 3  and finally outputs  I S R , which is  × 4  upscaling according to the original  I L R .

3.2. Discriminator Network

We proposed a novel discriminator containing the pre-processing block, cascaded perception turbo blocks (PTBs), and the post-processing block. The structure of the discriminator is depicted in Figure 4.
The pre-processing block is utilized for the initial feature perception of  I S R  and  I H R . As shown in Figure 4, it includes a CSR block, two residual blocks, and a downsampling layer. The CSR block consists of a convolution layer, an SN layer, and a ReLU activation function. The specific structure of the two residual blocks Res1 and Res2 is shown in the Figure 5.
The second-perception turbo is the core structure of this discriminator, which consists of cascaded PTBs. In order to further promote the generator to restore subtle textures, we proposed the PTB structure and made the following four improvements on the basis of hour-glass module [39]:
  • As shown in the Figure 5, we adopt the CSR structure instead of BRC, which consists of the BN layer, the ReLU activation function, and the convolutional layer. It has been proven that removing the BN layers can prevent BN artifacts of SR images, improve the performance, and reduce the computational complexity in the SR task [25]. In addition, we improve the perceptual loss by using the features before activation, which could provide stronger supervision for brightness consistency and texture recovery [11].
  • In the upsampling procedure, we use pixel-shuffle instead of nearest neighbor interpolation, which may lose pixel information.
  • In the downsampling layer, we use convolution instead of Maxpool2d operation, which may lose the integrity of feature map.
  • We enlarge the input channels of PTB to 128, which improves the perceptive capabilities of the discriminator.
The post-processing block consists of three convolutional layers to further learn features and output a feature map that benefits the computation of adversarial loss.
Based on the above improvements, the discriminator could further perceive multi-scale features at various spatial relationships and promote the generator to restore subtle textures driven by GAN.

3.3. Perception Loss

We introduced the loss function similar to ESRGAN, which is a hybrid weighted loss function that takes into account pixel-level recovery and visual perception effects and is able to achieve better super-resolution quality. Therefore, the total loss function of the generator  L G  is a weighted combination of several losses: the adversarial loss  L G A N , pixel loss  L P i x e l , and perceptual loss  L P e r c e p . The loss function of the discriminator  L D  is the adversarial loss  L G A N . The  L G  is described as follows:
L G = α L G A N + β L P i x e l + γ L P e r c e p
where  L P i x e l = E x i G x i y 1  is the one-norm distance between the recovered image  G x i  and HR image y; it thus evaluates the average degree of approximation of  I S R  and  I H R  over pixels.  α , β , γ  are coefficients to balance different loss terms. Moreover,  L P e r c e p  is gained by introducing a fine-tuned VGG19 network to calculate the one-norm distance between the recovered image  G x i  and high-level features of y. It is used to evaluate the approximation of  I S R  and  I H R  in human perception. The perceptual loss is calculated as follows:
L P e r c e p = E x i V G G ( G x i ) V G G ( y ) 1
L G A N  aims to distinguish the SR image from the HR image by the superior perceptive capability of the discriminator, which could help to learn sharper edges and more detailed textures; it can be formulated as follows:
L G A N = E x h r l o g 1 D x h r , x s r E x s r l o g D x s r , x h r

4. Experiments

In this section, we will discuss our PSRGAN model trained in RGB three channels.

4.1. Training Details

The experiments are performed with a scaling factor of  × 4  between LR and HR images; we obtain corresponding four-times smaller LR images by degrading the HR pictures, which are cropped to size  400 × 400  using the high-order [15] algorithm. Meanwhile, the patch size of cropped HR is  256 × 256 , and the patch size of LR is  64 × 64 . When training, the batch size is set to  12 × 2 , which means that we use two GPUs and the batch size per GPU is 12.
The training process is divided into two stages. One is the pre-training generator, and the other is conducting GAN training combined with the generator and discriminator. First, in the pre-training process, we purely train the generator with the L1 loss. The learning rate is  2 × 10 4 , and the sum of the iteration is 0.4 million. Then, we employ the pre-training generator model as an initialization for the generator. The GAN is trained with a combination of L1 loss, perception loss, and GAN loss, with weights of 1, 1, and 0.1, respectively. The learning rate is set to  1 × 10 4  for both the generator and discriminator, and the sum of iteration is 0.28 million. Pre-training with L1 loss is beneficial to obtain more visually pleasing results by avoiding undesired local optima for the generator. Moreover, it can help the discriminator to distinguish more on the textures part so that the discriminator can receive relatively better super-resolved images during GAN training.
For optimization, we use Adam [58] with  β 1 = 0.9 β 2 = 0.99 . We alternately update the generator and discriminator network until the model converges. We implement our models with the PyTorch framework and train them using NVIDIA GeForce RTX 3090 GPUs.

4.2. Data

For training, we use the DIV2K dataset [59], the Flickr2K dataset [21], and the OutdoorSceneTraining(OST) dataset [60] as training datasets. We employ these large datasets with rich textures, which help to generate SR pictures with more natural and subtle textures [11].
We evaluate our models on widely used benchmark datasets, including OST300 [60], PIRM Self val [35], 2020track1 [51], RealSR-Canon [61], DRealSR Test x4 [28], and RealSR-Nikon [61]. In particular, the images from RealSR-Canon and RealSR-Nikon are the center subimages of original images, and those larger than  1 K × 1 K  are cropped to  1 K × 1 K .

4.3. Qualitative Results

Due to the accessibility of SR methods, we compare our PSRGAN with several state-of-the-art methods, including BSRGAN, PDM SR, SwinIR [56], LDL [40], ESRGAN, and real-ESRGAN+. We have shown some representative qualitative results with NIQE in Figure 6 and Table 3. More detailed results calculated by NRQM and PI are presented in Table 4, Table 5 and Table 6. It can be observed from the figure that the results of our proposed PSRGAN outperforms previous approaches in both details and clearness, with fewer artifacts. For instance, PSRGAN can produce clearer, more natural lion fur (see 0901) and more detailed wall structures (see OST 278) than BSRGAN and LDL, whose textures are unnatural, skewed, and contain unpleasing noise. Compared with PSRGAN, ESRGAN and real-ESRGAN+ fail to produce enough details. Moreover, PSRGAN is capable of boosting visual sharpness (see DSC 1454 x1), while other methods either produce blurry structures (ESRGAN, PDM SR, and SwinIR) or do not generate enough details (BSRGAN). In addition, previous GAN-based methods sometimes introduced unpleasant artifacts such as BSRGAN and real-ESRGAN+. Our PSRGAN eliminates these artifacts and obtains cleaner results (see Canon 40 x1).
Although the NIQE score of PSRGAN is not always best, we still believe that exploring the effect of focusing on the human visual perception of real pictures is crucial for SR; after all, the existing perception indexes do not reflect all the problems. Please refer to Appendix C for more qualitative results.

4.4. Ablation Study

In order to study the effects of each component in the proposed PSRGAN, we gradually modify the discriminators of PSRGAN and compare their differences. The overall visual comparison is illustrated in Figure 7. Each column represents a model with its configurations shown at the top. The red sign indicates the best performance. A detailed discussion is provided as Table 7 follows.
Number of PTBs: The discriminator with the optimal number of cascaded PTBs has a strong representation capacity to capture semantic information, which can further improve the recovered textures, especially for regular structures like the wall of image OST 278 in Figure 6. We set the order of the number to 2, 3, 4, 5, 6, and 7 for experimentation, respectively. For simplisity, we only demonstrate the results of 3, 5, and 7 numbers; the experimental results are depicted in Figure 7. As shown, when the number is 5, the results are relatively sharper with richer textures than others. For some cases, a prominent difference can be observed from the second, third and fifth column in Figure 7.
Channel size of PTB: The different channel sizes of PTB influence the perceptive capabilities of the discriminator. We have tested on 3, 128, and 256 channels. For simplisity, we only demonstrate the results of 128 and 256 channels, as shown in Figure 7. When the channel size is 128, the results are clearer and have fewer artifacts.
Cross verification between PTBs and U-net: Please refer to Appendix A for details.

4.5. Running Times

Our method achieves moderate GPU run times for both training and testing, thanks to its design characteristics. Our model achieves outstanding super-resolution performance, reaching a superior level of quality after a rigorous training regimen of 490 k iterations. Our model exhibits test times on multiple datasets that are comparable to existing state-of-the-art models. Notably, when compared to SwinIR and LDL, our model demonstrates a significant advantage in test time efficiency. The algorithms were trained and tested on a server with NVIDIA GeForce RTX 3090 GPUs. Table 8 and Table 9 compare the running times of different state-of-the-art models.

5. Discussion

In this study, we present the perception-design-oriented image super resolution generative adversarial network (PSRGAN), an innovative approach that fuses generative adversarial networks (GANs) and human perceptual insights. Through extensive experiments and analysis of the model, we have achieved the following major achievement.
Perceptually guided super-resolution enhancement: We successfully combined human perceptual insights and used them to guide super-resolution processes. This resulted in sharper, more realistic, and more human-perceivable high-resolution image generation, as illustrated by Figure 6, where our PSRGAN generates more detailed textures of animal hairs, fewer artifacts, and a sharper edge in text-related images.
The experimental results: Our extensive experiments show that PSRGAN achieves significant performance gains on multiple datasets and tasks. Quantitative evaluations show that PSRGAN outperforms traditional super-resolution methods (real-ESRGAN+, ESRGAN, and BSRGAN) on multiple standard image quality metrics such as NIQE, NRQM, and PI. More encouragingly, the images generated by PSRGAN are closer to the high-resolution original images in terms of human perception.
Limitations: Despite our satisfactory achievements, we have to recognize some limitations of PSRGAN. Computational requirements: the training and inference of PSRGAN requires a large number of computational resources, which may be a challenge for some applications. Data diversity: while our model performs well on multiple datasets, performance may be degraded in specific domains or with uneven data distribution.
In my opinion, the SR network will definitely develop in the direction of breaking through its current limitations in the future, and the trend of super-resolution application is to reduce the computational burden and to apply it to diversified datasets.

6. Conclusions

We have presented a PSRGAN model that achieves superior perceptual quality both in terms of evaluation metrics and visual effects. According to the experimental results, our proposed PSRGAN based on generative adversarial networks outperforms current state-of-the-art intelligent algorithms (BSRGAN, real-ESRGAN, PDM_SR, SwinIR, LDL, etc.) on several evaluation metrics (NIQE, NRQM and PI), with a  × 4  upscaling factor on various datasets (OST300, DRealSR_Test_x4, RealSR-Canon, etc.). The PSRGAN model mainly consists of two kinds of perception turbo (PT), GPT in the generator network, and DPT in the discriminator network. In terms of visual effects, the proposed image slice block mitigates the artifacts and noise in the reconstructed image, the three-level perception structure in GPT which could extract diversified textures. The cascaded PTBs in DPT could further promote the generator to restore subtle textures.

Author Contributions

All authors contributed to the study conception and design. T.W. and Y.Z. participated in the design of the network structure, S.X. is responsible for the budget and paper quality and revision, and the network implementation was carried out by Y.L. and H.T. Material preparation, data collection, and analysis were performed by H.T., H.L. (Huaizheng Liu) and J.Z. The model training and experimental results collection was carried out by Y.L., H.T. and H.L. (Huaizheng Liu). Y.Z. participated in the system integration and testing. The first draft of the manuscript was written by H.L. (Hui Liu) and J.Z. All of the authors commented on previous versions of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Tencent Technology (Shenzhen) Co., Ltd.

Data Availability Statement

The calculation method of NIQE is derived from the following resources available in the public domain: https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023). The verification program is available at https://drive.google.com/file/d/1AZmkgvgfcBicTP9tM5X-gahnirskm6-1/view?usp=share_link (accessed on 1 June 2023). The datasets analyzed during the current study are derived from the following resources available in the public domain: https://github.com/XPixelGroup/BasicSR/blob/master/docs/DatasetPreparation.md (accessed on 1 June 2023).

Acknowledgments

The authors would like to thank Tencent Company for helpful budget and resource on topics related to this work, special the Qilong Kou, the department of Tencent Institute of Games.

Conflicts of Interest

The authors have no competing interests to declare that are relevant to the content of this article. This work did not involve human participants or animal research.

Appendix A. Cross Verification between PTBs and U-Net

In PSRGAN, we now call its generator network PSRNet and its discriminator network PTBs. In this section, we further compare the differences between two kinds of GANs, PSRNet+PTBs, and PSRNet+U-net on diverse testing datasets. From the results in Table A1, we can conclude that using PTBs can promote the generator to restore more perceptive SR images driven by GAN; more qualitative comparisons are shown in Figure A1.
Figure A1. Qualitative comparisons on representative real-world samples with  × 4  upscaling factors. PSRNet+PTBs outperforms PSRNet+U-net in terms of both restoring texture details (See OST 99) and producing clearer results (See OST 164).
Figure A1. Qualitative comparisons on representative real-world samples with  × 4  upscaling factors. PSRNet+PTBs outperforms PSRNet+U-net in terms of both restoring texture details (See OST 99) and producing clearer results (See OST 164).
Electronics 12 04420 g0a1
Table A1. NIQE scores on several diverse testing datasets. The lower, the better.
Table A1. NIQE scores on several diverse testing datasets. The lower, the better.
PIRM Self ValOST300RealSR-NikonRealSR-Canon
PSRNet+U-net3.5272.8305.6735.896
PSRGAN (PSRNet+PTBs)3.3302.7355.1644.499

Appendix B. Structure of Multi-Scale Feature Extraction Block

We conducted experiments on the number of convolutional groups for multi-scale feature-extraction block (MFEB) in the generator network. As the experimental results in Table A2 show, the SR results show better performance when the number of convolutional groups is three.
Table A2. NIQE scores of feature extraction block at different scales on diverse testing datasets; the lower, the better. The calculation method of NIQE is derived from the basic SR package of PyTorch 1.11.0+cu113.
Table A2. NIQE scores of feature extraction block at different scales on diverse testing datasets; the lower, the better. The calculation method of NIQE is derived from the basic SR package of PyTorch 1.11.0+cu113.
Groups12345
OST3006.6503626.6353826.5203036.6262116.61622
DRealSR Test x47.8925497.9327097.8187157.9937317.821493
RealSR-Canon10.17990410.31331410.20835110.3283110.3463
RealSR-Nikon10.44569110.49159810.40944310.58292510.378877
PIRM Self val6.7177166.7179836.6539056.6998136.694974
2020track16.5963916.5839756.462276.601036.526587

Appendix C. More Qualitative Results

Figure A2. More qualitative results of PSRGAN and NIQE are provided for reference [ × 4 upscaling].
Figure A2. More qualitative results of PSRGAN and NIQE are provided for reference [ × 4 upscaling].
Electronics 12 04420 g0a2

References

  1. Duchon, C.E. Lanczos filtering in one and two dimensions. J. Appl. Meteorol. Climatol. 1979, 18, 1016–1022. [Google Scholar] [CrossRef]
  2. Zhang, L.; Wu, X. An edge-guided image interpolation algorithm via directional filtering and data fusion. IEEE Trans. Image Process. 2006, 15, 2226–2238. [Google Scholar] [CrossRef] [PubMed]
  3. Wu, Y.; Ding, H.; Gong, M.; Qin, A.; Ma, W.; Miao, Q.; Tan, K.C. Evolutionary multiform optimization with two-stage bidirectional knowledge transfer strategy for point cloud registration. IEEE Trans. Evol. Comput. 2022. [Google Scholar] [CrossRef]
  4. Wu, Y.; Zhang, Y.; Ma, W.; Gong, M.; Fan, X.; Zhang, M.; Qin, A.; Miao, Q. Rornet: Partial-to-partial registration network with reliable overlapping representations. IEEE Trans. Neural Netw. Learn. Syst. 2023. [Google Scholar] [CrossRef] [PubMed]
  5. Dai, S.; Han, M.; Xu, W.; Wu, Y.; Gong, Y.; Katsaggelos, A.K. Softcuts: A soft edge smoothness prior for color image super-resolution. IEEE Trans. Image Process. 2009, 18, 969–981. [Google Scholar] [PubMed]
  6. Sun, J.; Xu, Z.; Shum, H.Y. Image super-resolution using gradient profile prior. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, Alaskapp, 23–28 June 2008; pp. 1–8. [Google Scholar]
  7. Yan, Q.; Xu, Y.; Yang, X.; Nguyen, T.Q. Single image superresolution based on gradient profile sharpness. IEEE Trans. Image Process. 2015, 24, 3187–3202. [Google Scholar] [PubMed]
  8. Shang, T.; Dai, Q.; Zhu, S.; Yang, T.; Guo, Y. Perceptual extreme super-resolution network with receptive field block. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 440–441. [Google Scholar]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Adv. Neural Inf. Process. Syst. 2020, 63, 139–144. [Google Scholar] [CrossRef]
  10. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single-image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  11. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  12. Yuan, Y.; Liu, S.; Zhang, J.; Zhang, Y.; Dong, C.; Lin, L. Unsupervised image super-resolution using cycle-in-cycle generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 701–710. [Google Scholar]
  13. Zhang, Y.; Liu, S.; Dong, C.; Zhang, X.; Yuan, Y. Multiple cycle-in-cycle generative adversarial networks for unsupervised image super-resolution. IEEE Trans. Image Process. 2019, 29, 1101–1112. [Google Scholar] [CrossRef]
  14. Zhang, K.; Liang, J.; Van Gool, L.; Timofte, R. Designing a practical degradation model for deep blind image super-resolution. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4791–4800. [Google Scholar]
  15. Wang, X.; Xie, L.; Dong, C.; Shan, Y. real-ESRGAN: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  16. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  17. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  18. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
  19. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  20. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep laplacian pyramid networks for fast and accurate super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 624–632. [Google Scholar]
  21. Timofte, R.; Agustsson, E.; Van Gool, L.; Yang, M.H.; Zhang, L. Ntire 2017 challenge on single-image super-resolution: Methods and results. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 114–125. [Google Scholar]
  22. Haris, M.; Shakhnarovich, G.; Ukita, N. Deep back-projection networks for super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1664–1673. [Google Scholar]
  23. Dong, C.; Loy, C.C.; Tang, X. Accelerating the super-resolution convolutional neural network. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 391–407. [Google Scholar]
  24. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1874–1883. [Google Scholar]
  25. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single-image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  26. Musunuri, Y.R.; Kwon, O.S. Deep residual dense network for single-image super-resolution. Electronics 2021, 10, 555. [Google Scholar] [CrossRef]
  27. Ji, X.; Cao, Y.; Tai, Y.; Wang, C.; Li, J.; Huang, F. Real-world super-resolution via kernel estimation and noise injection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 466–467. [Google Scholar]
  28. Wei, P.; Xie, Z.; Lu, H.; Zhan, Z.; Ye, Q.; Zuo, W.; Lin, L. Component divide-and-conquer for real-world image super-resolution. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part VIII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 101–117. [Google Scholar]
  29. Luo, Z.; Huang, Y.; Li, S.; Wang, L.; Tan, T. Learning the degradation distribution for blind image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 6063–6072. [Google Scholar]
  30. Wang, X.; Yu, K.; Dong, C.; Loy, C.C. Recovering realistic texture in image super-resolution by deep spatial feature transform. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 606–615. [Google Scholar]
  31. Shao, G.; Sun, Q.; Gao, Y.; Zhu, Q.; Gao, F.; Zhang, J. Sub-Pixel Convolutional Neural Network for Image super-resolution Reconstruction. Electronics 2023, 12, 3572. [Google Scholar] [CrossRef]
  32. Bruna, J.; Sprechmann, P.; LeCun, Y. super-resolution with deep convolutional sufficient statistics. arXiv 2015, arXiv:1511.05666. [Google Scholar]
  33. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar]
  34. Sajjadi, M.S.; Scholkopf, B.; Hirsch, M. Enhancenet: Single-image super-resolution through automated texture synthesis. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4491–4500. [Google Scholar]
  35. Mechrez, R.; Talmi, I.; Zelnik-Manor, L. The contextual loss for image transformation with non-aligned data. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 768–783. [Google Scholar]
  36. Blau, Y.; Mechrez, R.; Timofte, R.; Michaeli, T.; Zelnik-Manor, L. The 2018 PIRM challenge on perceptual image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  37. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  38. Liu, S.; Huang, D. Receptive field block net for accurate and fast object detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 385–400. [Google Scholar]
  39. Zhang, K.; Gu, S.; Timofte, R. Ntire 2020 challenge on perceptual extreme super-resolution: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 492–493. [Google Scholar]
  40. Newell, A.; Yang, K.; Deng, J. Stacked hourglass networks for human pose estimation. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part VIII 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 483–499. [Google Scholar]
  41. Liang, J.; Zeng, H.; Zhang, L. Details or artifacts: A locally discriminative learning approach to realistic image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5657–5666. [Google Scholar]
  42. Zhang, M.; Ling, Q. Supervised pixel-wise GAN for face super-resolution. IEEE Trans. Multimed. 2020, 23, 1938–1950. [Google Scholar] [CrossRef]
  43. Gong, Y.; Liao, P.; Zhang, X.; Zhang, L.; Chen, G.; Zhu, K.; Tan, X.; Lv, Z. Enlighten-GAN for super resolution reconstruction in mid-resolution remote sensing images. Remote Sens. 2021, 13, 1104. [Google Scholar] [CrossRef]
  44. Sheikh, H.R.; Bovik, A.C.; De Veciana, G. An information fidelity criterion for image quality assessment using natural scene statistics. IEEE Trans. Image Process. 2005, 14, 2117–2128. [Google Scholar] [CrossRef]
  45. Sheikh, H.R.; Bovik, A.C. Image information and visual quality. IEEE Trans. Image Process. 2006, 15, 430–444. [Google Scholar] [CrossRef]
  46. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  47. Ma, C.; Yang, C.Y.; Yang, X.; Yang, M.H. Learning a no-reference quality metric for single-image super-resolution. Comput. Vis. Image Underst. 2017, 158, 1–16. [Google Scholar] [CrossRef]
  48. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  49. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  50. Blau, Y.; Michaeli, T. The perception-distortion tradeoff. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6228–6237. [Google Scholar]
  51. Vasu, S.; Thekke Madam, N.; Rajagopalan, A. Analyzing perception-distortion tradeoff using enhanced perceptual super-resolution network. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  52. Lugmayr, A.; Danelljan, M.; Timofte, R. Ntire 2020 challenge on real-world image super-resolution: Methods and results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 494–495. [Google Scholar]
  53. Cai, J.; Zeng, H.; Yong, H.; Cao, Z.; Zhang, L. Toward real-world single-image super-resolution: A new benchmark and a new model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3086–3095. [Google Scholar]
  54. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  55. He, E.; Chen, Q.; Zhong, Q. SL-Swin: A transformer-Based Deep Learning Approach for Macro-and Micro-Expression Spotting on Small-Size Expression Datasets. Electronics 2023, 12, 2656. [Google Scholar] [CrossRef]
  56. Yang, F.; Yang, H.; Fu, J.; Lu, H.; Guo, B. Learning texture transformer network for image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5791–5800. [Google Scholar]
  57. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  58. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  59. Lu, Z.; Li, J.; Liu, H.; Huang, C.; Zhang, L.; Zeng, T. transformer for single-image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 457–466. [Google Scholar]
  60. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  61. Agustsson, E.; Timofte, R. Ntire 2017 challenge on single-image super-resolution: Dataset and study. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 126–135. [Google Scholar]
Figure 1. Comparisons of visual quality among BSRGAN [14], real-ESRGAN+ [1], and PSRGAN on real-life images by  × 4  upscaling. The PSRGAN can generate naturally finer textures and remove or alleviate annoying artifacts for real-world images. Zoom in for best view.
Figure 1. Comparisons of visual quality among BSRGAN [14], real-ESRGAN+ [1], and PSRGAN on real-life images by  × 4  upscaling. The PSRGAN can generate naturally finer textures and remove or alleviate annoying artifacts for real-world images. Zoom in for best view.
Electronics 12 04420 g001
Figure 2. Architecture of generator network with corresponding kernel size (k), number of feature maps (n), and stride (s) indicated for each convolutional layer, where  F 1 F 2 , and  F 3  are multi-scale features extracted by MFEB described in Figure 3.
Figure 2. Architecture of generator network with corresponding kernel size (k), number of feature maps (n), and stride (s) indicated for each convolutional layer, where  F 1 F 2 , and  F 3  are multi-scale features extracted by MFEB described in Figure 3.
Electronics 12 04420 g002
Figure 3. Design of MFEB in first-perception turbo.
Figure 3. Design of MFEB in first-perception turbo.
Electronics 12 04420 g003
Figure 4. Discriminator network structure with second-perception turbo. The structure of CSR, Res1, and Res2 are shown in Figure 5.
Figure 4. Discriminator network structure with second-perception turbo. The structure of CSR, Res1, and Res2 are shown in Figure 5.
Electronics 12 04420 g004
Figure 5. (Left): Differences between BRC and CSR; (right): structure of Res1 and Res2 in PTBs.
Figure 5. (Left): Differences between BRC and CSR; (right): structure of Res1 and Res2 in PTBs.
Electronics 12 04420 g005
Figure 6. Qualitative results of PSRGAN. PSRGAN produces more subtle textures and clearer structures, e.g., animal texture and building structure, as well as fewer unpleasant artifacts, e.g., artifacts in fonts. Zoom in for best view.
Figure 6. Qualitative results of PSRGAN. PSRGAN produces more subtle textures and clearer structures, e.g., animal texture and building structure, as well as fewer unpleasant artifacts, e.g., artifacts in fonts. Zoom in for best view.
Electronics 12 04420 g006
Figure 7. Visual comparisons of different configurations in PSRGAN. The red sign indicates the best performance.
Figure 7. Visual comparisons of different configurations in PSRGAN. The red sign indicates the best performance.
Electronics 12 04420 g007
Table 1. Related work on design of discriminator networks.
Table 1. Related work on design of discriminator networks.
Different MethodsDesign of Discriminator Networks
SRGANVGG-style, which is trained to distinguish between SR images
ESRGANborrows ideas from relativistic GAN to improve the discriminator in SRGAN
Real-ESRGANproposed an U-Net design
RFB-ESRGANproposed stacked hourglass network which captures and consolidates information across all scales of the image
Table 2. Related work on evaluation metrics.
Table 2. Related work on evaluation metrics.
Evaluation MetricsAdvantageDisadvantage
Distortion metricsSimple calculationGreater inconsistency with perceived quality
Human opinion scoreConsistent with visual perceptionHigh labor costs
No-reference quality measuresBalancing consistency with perceived quality and computational costThere is some inconsistency with visual perception
Table 3. NIQE scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third NIQE results among models on each dataset row. The calculation method of NIQE is derived from the basic SR package of PyTorch 1.11.0 + cu113.
Table 3. NIQE scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third NIQE results among models on each dataset row. The calculation method of NIQE is derived from the basic SR package of PyTorch 1.11.0 + cu113.
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
OST3007.6003.3094.3192.9212.8173.5012.8062.735
DRealSR Test x49.7724.8037.6674.6985.2508.6444.8464.533
RealSR-Canon13.4805.99810.0154.9565.63713.0965.3524.499
RealSR-Nikon13.0176.3779.5444.8195.71212.4435.1805.164
PIRM Self val7.7473.8085.1323.6833.5393.5163.3503.330
2020track17.5963.7834.1013.6183.9587.4403.8203.411
Table 4. NIQE scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third NIQE results among models on each dataset row. The calculation method of NIQE is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
Table 4. NIQE scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third NIQE results among models on each dataset row. The calculation method of NIQE is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
OST3007.6123.4144.3083.0344.563.5512.9292.826
DRealSR Test x49.7664.8187.6359.7658.3728.6324.8484.543
RealSR-Canon13.4426.04610.0084.98513.18713.1015.3464.512
RealSR-Nikon13.0066.4359.5374.83412.3912.4465.1765.169
PIRM Self val7.7463.8385.1953.7162.9863.5113.3633.311
2020track17.6063.8134.0967.6063.2497.2173.8353.423
Table 5. NRQM scores on diverse testing datasets—the higher, the better. Colors R, G, and B indicate the best first, second, and third NRQM results among models on each dataset row. The calculation method of NRQM is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
Table 5. NRQM scores on diverse testing datasets—the higher, the better. Colors R, G, and B indicate the best first, second, and third NRQM results among models on each dataset row. The calculation method of NRQM is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
OST3003.2666.3195.7376.585.6836.2366.5766.714
DRealSR Test x42.5765.2643.5362.5763.3173.2445.2955.551
RealSR-Canon2.3374.5712.4844.8612.5482.4765.7436.131
RealSR-Nikon2.3664.6352.5975.2492.8662.6815.695.839
PIRM Self val3.768.0916.0968.1918.3938.4018.3478.524
2020track13.3076.2195.993.3076.4936.5916.1336.504
Table 6. PI scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third PI results among models on each dataset row. The calculation method of PI is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
Table 6. PI scores on diverse testing datasets—the lower, the better. Colors R, G, and B indicate the best first, second, and third PI results among models on each dataset row. The calculation method of PI is in PIRM2018 derived from https://github.com/roimehrez/PIRM2018 (accessed on 1 June 2023).
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
OST3007.1733.5484.2863.2274.4383.6583.1763.056
DRealSR Test x48.5954.7777.058.5957.5277.6944.7774.495
RealSR-Canon10.5525.7388.7625.06210.31910.3134.8024.191
RealSR-Nikon10.325.98.474.7939.7629.8834.7434.665
PIRM Self val6.9942.8744.5492.7632.2972.5552.5092.394
2020track17.153.7974.0537.153.3785.3133.8513.459
Table 7. Model with different configurations.
Table 7. Model with different configurations.
SecondThirdFourthFifth
PTBs3557
Channels128128256128
Table 8. The GPU run times for training of different networks. The unit is the number of iterators, and k represents thousands. Since Bicubic is not an adversarial neural network, there is no number of iterators.
Table 8. The GPU run times for training of different networks. The unit is the number of iterators, and k represents thousands. Since Bicubic is not an adversarial neural network, there is no number of iterators.
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
GAN Training Times (iters)None1000 k200 k500 k400 k400 k400 k490 k
Table 9. The GPU run times of different networks on diverse datasets. The unit is time, where m stands for minutes and s stands for seconds.
Table 9. The GPU run times of different networks on diverse datasets. The unit is time, where m stands for minutes and s stands for seconds.
BicubicBSRGANPDM SRSwinIRLDLESRGANreal-ESRGAN+PSRGAN
OST30054 s5 m 4 s5 m 13 s23 m 54 s7 m 15 s5 m 36 s5 m 16 s5 m 29 s
DRealSR Test x41 m 2 s5 m 47 s5 m 46 s21 m 56 s8 m 34 s6 m 21 s5 m 58 s6 m 20 s
RealSR-Canon17 s2 m2 m 1 s9 m 6 s2 m 20 s2 m2 m 3 s2 m 9 s
RealSR-Nikon22 s2 m 31 s2 m 30 s9 m 26 s3 m 12 s2 m 31 s2 m 34 s2 m 42 s
PIRM Self val1 s6 s6 s17 s19 s5 s7 s7 s
2020track110 s53 s54 s2 m 56 s1 m 18 s55 s56 s58 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, T.; Xiong, S.; Liu, H.; Zhao, Y.; Tuo, H.; Li, Y.; Zhang, J.; Liu, H. PSRGAN: Perception-Design-Oriented Image Super Resolution Generative Adversarial Network. Electronics 2023, 12, 4420. https://doi.org/10.3390/electronics12214420

AMA Style

Wu T, Xiong S, Liu H, Zhao Y, Tuo H, Li Y, Zhang J, Liu H. PSRGAN: Perception-Design-Oriented Image Super Resolution Generative Adversarial Network. Electronics. 2023; 12(21):4420. https://doi.org/10.3390/electronics12214420

Chicago/Turabian Style

Wu, Tao, Shuo Xiong, Hui Liu, Yangyang Zhao, Haoran Tuo, Yi Li, Jiaxin Zhang, and Huaizheng Liu. 2023. "PSRGAN: Perception-Design-Oriented Image Super Resolution Generative Adversarial Network" Electronics 12, no. 21: 4420. https://doi.org/10.3390/electronics12214420

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop