Next Article in Journal
Smartphone-Based Methodology Applied to Electromagnetic Field Exposure Assessment
Previous Article in Journal
Fault-Tolerant Control Based on Current Space Vectors against Total Sensor Failures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

A Residual Dense Attention Generative Adversarial Network for Microscopic Image Super-Resolution

1
Xiamen Key Laboratory of Mobile Multimedia Communications, College of Information Science and Engineering, Huaqiao University, Xiamen 361021, China
2
School of Opto-Electronic and Communication Engineering, Xiamen University of Technology, Xiamen 361024, China
3
Institute of Microelectronics Chinese Academy of Sciences, Beijing 100029, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(11), 3560; https://doi.org/10.3390/s24113560
Submission received: 7 May 2024 / Revised: 27 May 2024 / Accepted: 29 May 2024 / Published: 31 May 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
With the development of deep learning, the Super-Resolution (SR) reconstruction of microscopic images has improved significantly. However, the scarcity of microscopic images for training, the underutilization of hierarchical features in original Low-Resolution (LR) images, and the high-frequency noise unrelated with the image structure generated during the reconstruction process are still challenges in the Single Image Super-Resolution (SISR) field. Faced with these issues, we first collected sufficient microscopic images through Motic, a company engaged in the design and production of optical and digital microscopes, to establish a dataset. Secondly, we proposed a Residual Dense Attention Generative Adversarial Network (RDAGAN). The network comprises a generator, an image discriminator, and a feature discriminator. The generator includes a Residual Dense Block (RDB) and a Convolutional Block Attention Module (CBAM), focusing on extracting the hierarchical features of the original LR image. Simultaneously, the added feature discriminator enables the network to generate high-frequency features pertinent to the image’s structure. Finally, we conducted experimental analysis and compared our model with six classic models. Compared with the best model, our model improved PSNR and SSIM by about 1.5 dB and 0.2, respectively.

1. Introduction

Regardless of the magnification and numerical aperture of the objective lens used, the imaging throughput of current microscopes is typically only in the order of ten megapixels [1]. This leads to a compromise between high resolution and Field of View (FOV) when imaging. However, in biomedical research such as histopathology, hematology, and neuroscience, there is a growing need for high-resolution imaging of large samples. In order to accurately resolve cellular-level life activities at the scale of the entire sample, a balance between the global structure and microscale local details is required, along with quantitative analysis. Faced with this challenge, Super-Resolution (SR) imaging techniques were developed. Single Image Super-Resolution (SISR) reconstruction [2] aims to reconstruct a High-Resolution (HR) image from an input Low-Resolution (LR) image.This technique is widely used in critical fields, including bright field micrographs [3], fluorescent imaging [4,5], remote sensing images [6,7], and surveillance videos [8]. Traditional SISR reconstruction algorithms can be divided into three categories: interpolation-based [9,10], reconstruction-based [11,12], and learning-based [13]. Interpolation-based algorithms have the advantages of simple principle and easy implementation. However, these methods have difficulty recovering the detailed information of the image, and the reconstructed image is seriously distorted. Reconstruction-based algorithms first generate the constraints for SR reconstruction according to the imaging process of LR images, establish the corresponding mathematical model, and finally reconstruct HR images. However, when the magnification is high, it is still difficult to recover enough high-frequency information. The learning-based reconstruction algorithm mainly establishes the mapping relationship between LR and HR hyperspectral images, and then reconstructs according to the mapping relationship. In recent years, with the rapid development of deep learning, the excellent learning ability of convolutional neural networks also brings new opportunities for the development of SR imaging technology.
Specifically, Dong et al. [14] used deep learning in image SR reconstruction by introducing Convolutional Neural Networks (CNNs). This groundbreaking work paved the way for subsequent innovations. Kim et al. [15] introduced the Very Deep Super Resolution (VDSR) model, which leverages a deeply constructed neural network combined with residual learning for image reconstruction. This approach effectively addresses the issue of image size reduction caused by successive convolutions. Lai et al. [16] proposed the Laplacian Pyramid Super-Resolution Network (LapSRN) model, which employs a Laplacian pyramid to progressively reconstruct the image and perform feature extraction through a residual structure. Ledig et al. [17] advanced the field by incorporating residual blocks within the structure of a Generative Adversarial Network (GAN) to create the Super-Resolution Generative Adversarial Network (SRGAN) model. The use of GANs significantly improves the visual perception quality of the generated images, making them more closely resemble real images. Lim et al. [18] developed the Enhanced Deep Super-Resolution (EDSR) model, which refines the residual structure found in the SRGAN model for improved performance. Zhang et al. [19] contributed by introducing the Residual Dense Block (RDB) structure and the Residual Dense Network (RDN) model, further advancing the capabilities of SR reconstruction.
In terms of microscopic image reconstruction, Zhang et al. [20] proposed Registration-Free GAN Microscopy (RFGANM) workflow by combining an SRGAN network with an optical microscope and degradation model to achieve deep learning SR in large FOV, and improve the resolution of wide-field microscopy and light-sheet fluorescence microscopy images. Wang et al. [21] used a GAN network to develop SR techniques for cross-modal fluorescence microimaging by simulating the mapping relationships between different imaging techniques, such as wide-field fluorescence imaging, confocal to STED, and TIRF to TIRF-SIM, through deep learning. Van Sloun [22] applied deep learning to ultrasound microscopy SR study and proposed the Deep-ULM model, which is based on the U-Net network to reduce the effect of diffraction and obtain HR images in real time. Li et al. [23] achieved high-quality reconstruction of ordinary wide-field fluorescence images to SIM SR imaging results using a Deep Fourier Channel Attention (DCFA) network, greatly contributing to the development of Super-Resolution in microscopic images.
It can be seen that, although great progress has been made in the SR of microscopic images, most of them are focused on fluorescence images in a special way. We know that the features of fluorescence microscopy images are significantly different from those of cell images. Additionally, despite the above SR methods having achieved good results, there are some shortcomings in any of them. These include the absence of specialized datasets for microscopic images, which hampers model training; the inadequate use of hierarchical features in LR images by GAN-based SR models; insufficient utilization of information across convolutional layers due to varying receptive fields; a lack of prioritization in reconstructing feature map information, leading to a suboptimal focus on critical details; and the reliance on pixel-level reconstruction errors as loss functions, which fails to capture high-frequency details and often results in overly smooth and potentially inaccurate images.
In order to respond to these challenges, this paper has carried out corresponding optimization work on cell microscopy images. The specific contributions are described as follows:
(1)
We produce a dataset of high- and low-resolution images of four cell types using microscope acquisition. You can find the dataset at https://github.com/wxsssss/RDAGAN/tree/main (accessed on 15 January 2024).
(2)
In order to fully utilize the hierarchical features of the original image, we propose a Residual Dense Attention Generative Adversarial Network (RDAGAN), whose generator uses RDABs with increased attention mechanisms.
(3)
To reconstruct the high-frequency features associated with HR images, we add a feature discriminator to the original discriminator and optimize the loss function.
(4)
Our proposed optimized model is compared with six classic models. Compared to the best performing model among them, PSNR and SSIM have improved by 1.5 dB and 0.2, respectively.

2. Proposed Methodology

Our model is based on the aforementioned RFGANM [20] that has been used for microscopic cell image SR, constructing a generator network architecture composed of four main sections: the feature extraction part, the residual dense attention part, the dense feature fusion part, and the reconstruction part. The input LR image first passes through the feature extraction section, where initial features are extracted. These features are then processed by the residual dense attention section. Here, the RDB extracts rich local features through densely connected convolutional layers, fully leveraging the hierarchical features of the layers. The local feature fusion within the RDB allows for an adaptive and more effective integration of previous and current local features, thus stabilizing the training of broader network. An attention mechanism is added after each RDB, first applying a channel attention module to obtain a weighted result, followed by a spatial attention module to refine the weighting. This enables the network to concentrate on information that is most beneficial for image reconstruction. The RDB and the attention module together constitute the Residual Dense Attention Block (RDAB). After extracting dense local features through a series of residual dense attention modules, dense feature fusion is proposed to mine multi-level features from a global perspective. Additionally, a feature discriminator [24] is introduced to differentiate the detailed features of SR and HR images, encouraging the reconstructed image to generate more high-frequency features instead of noise.

2.1. Microscopic Cell Image Dataset

We used a microscope to capture cell images using both 40× and 20× objective lenses simultaneously to establish our dataset. We use the EasyScan NFC 300 microscope from Motic, which is a high-performance fully automatic digital scanning microscope equipped with advanced optical and imaging technology, to provide high-quality microscopic images. The microscope features fast and precise autofocus capabilities to ensure image clarity throughout the scanning process. It employs a high-quality infinity-corrected optical system that delivers clear and distortion-free images. The optical system incorporates advanced correction design to minimize common optical aberrations such as spherical aberration, chromatic aberration, and coma, providing accurate imaging quality. Additionally, the system undergoes rigorous color correction and transmission curve optimization to ensure uniform light transmission across different wavelengths, reducing color distortion and light loss. We perform center cropping on the captured images. HR images were cropped to a resolution of 1024 × 1024 pixels, while LR images were cropped to 512 × 512 pixels, reflecting the 2× objective relationship. In constructing the dataset, we faced challenges in precisely matching HR and LR images. To address this, we employed Bicubic Downsampling (BD) to generate corresponding LR images from their HR counterparts. These LR images were then used for both reconstruction purposes and performance comparison. Finally, we created 800 pairs of images for training data and 200 pairs of image val data for reconstruction for each of four cell images. You can refer to Appendix A for relevant characteristics of microscopic cell images.

2.2. Residual Dense Blocks for Fusion CBAM

The structure of our generator network is shown in Figure 1, consisting of four parts: a shallow feature extraction network, a RDAB, a Dense Feature Fusion (DFF), and an upsampling network. The input and output of the generator network are represented by I L R and, respectively, I S R . A convolutional layer is used to extract shallow features F 0 of the LR image and perform global residual learning, while F 0 also serves as an input to the RDAB. F 0 denotes the convolution operation on the LR image. Assuming that there are D RDABs in the generator, the Dth RDAB can be represented as follows:
F d = H R D A B , d F d 1 = H R D A B , d H R D A B . d 1 H R D A B , 1 F 0
where H R D A B , d represents the operation of the dth RDAB, which is a composite function of convolution and ReLU operations. Meanwhile, F d is generated by the dth RDAB fully utilizing the features from each convolutional layer within the block and F D can be considered as a local feature. The generator RDABs in this paper are 16 in total.
After extracting the hierarchical features with a set of D RDABs, further DFF is performed, including Global Feature Fusion (GFF) and Global Residual Learning (GRL). DFF fully utilizes the features of all previous layers; F D F is the output feature mapping of the composite function H D F F of the DFF module, which can be expressed as follows:
F D F = H D F F F 0 , F 1 , F D
Local and global features are extracted and fused in the LR space to obtain F D F , which is used as an input to the upsampling network. The upsampling network used in this invention is the same as SRGAN. The entire network of the generator as a whole can be represented as follows:
I S R = H G e n e r a t o r I L R
The structure of the RDAB is shown below in Figure 2. Our RDAB contains a dense connectivity layer, local feature fusion (LFF), Local Residual Learning (LRL), and CBAM, thus forming a Continuous Memory (CM) mechanism.
Denote the input and output of the dth RDAB by F d 1 and F d , respectively. The outputs of the previous RDAB and each convolutional layer of that RDAB are directly connected to all subsequent layers, which not only preserves the feed-forward properties but also extracts the local dense features. F d , c denotes the output of the cth convolutional layer in that RDAB, and F d , C denotes the output of the last convolutional layer of the densely connected layers in that RDAB. The purpose of LRL is to further improve the information flow, and the final output of the dth RDAB is denoted as follows:
F d , L F = H L F F d F d 1 , F d , , F d , c , F d , C
F d = F d 1 + F d , L F
where H L F F d · denotes the fusion of all layer feature information in the dth RDAB. DFF is feature fusion and residual learning for each RDAB feature obtained to utilize hierarchical features in a global manner. DFF includes GFF and GRL. The global features are extracted by fusing the features of all RDABs.
F G F = H G F F F 1 , , F D
F D F = F 0 + F G F
The RDAB introduces a continuous memory mechanism that allows previous RDABs to directly access each layer of the current RDAB. By using LFF, it is possible to train the deep network stably, while LRL can further optimize the information flow and gradient. In addition the RDAB uses GFF and GRL to further extract global features.
Convolutional Block Attention Module (CBAM) [25] is a simple and efficient forward convolutional neural attention module. CBAM consists of two independent sub-modules: the channel attention module and the spatial attention module. The output of the convolutional layer first passes through the channel attention module to obtain a weighted result, which then goes through the spatial attention module for final weighting. This approach not only saves parameters and computational power but also ensures that it can be integrated as a plug-and-play module into existing network architectures. The diagram of the attention module is shown in Figure 3.

2.3. Discriminator and Loss Function

In order to make the reconstructed image generate real high-frequency information, we add a feature discriminator under the original image discriminator [21]. The image discriminator, similar to the discriminators in traditional GAN models, takes the SR image as input to determine whether it is a real HR image or a SR image. The feature discriminator inputs the SR image into a VGG network to extract intermediate feature maps. Since the extracted feature maps contain structural information, the feature discriminator distinguishes between SR images and real HR images based not only on high-frequency components but also on structural components. Through adversarial training with the generator and the two discriminators, our generator is trained to synthesize realistic structural features rather than random high-frequency noise. The structure of our discriminator is shown in Figure 4.
Image discriminator D i and feature discriminator D f are next described. The image discriminator D i judges real HR images and SR images based on pixel values, while the feature discriminator D f distinguishes between real HR images and SR images through the mapping of feature maps. The generator loss function is as follows:
C g = C p + λ C a i + C a f
C p is the perceptual loss [20]. C a i is the generator’s image GAN loss for synthesizing high-frequency details in the pixel domain. C a f is the generator’s feature GAN loss for synthesizing structural details in the feature domain. λ is the weights of the GAN loss terms. To train the discriminators D i and D f , we minimize the losses C d i and C d f , which correspond to C a i and C a f , respectively. The generator and the discriminator are trained by alternately minimizing C g , C d i , and C d f . The image GAN loss C a i and image discriminator loss C d i of the generator are defined as follows:
C a i = log D i I g
C d i = log D i I h log 1 D i I g
where D i I is the output of the image discriminator. The feature GAN loss C a f and feature discriminator loss L d f of the generator are defined as follows:
C a f = log D f ϖ k I g
C d f = log D f ϖ k I h log 1 D f ϖ k I g
where D f ϖ k is the output of the feature discriminator. During training, the generator can be made to produce realistic structural high-frequency details instead of noise artifacts.

3. Experimental Results and Discussion

3.1. Experimental Settings

We used cell images obtained from a 40× microscope as HR images and downsampled them using BD to obtain the corresponding LR images. Each of the four cells has 800 pairs of training datasets and 5 test datasets. The test datasets are each named Cell A, Cell B, Cell C, and Cell D. We use the Peak Signal-to-Noise Ratio (PSNR), Structure Similarity Index Measure (SSIM), and Learned Perceptual Image Patch Similarity (LPIPS) to verify the authenticity of the reconstructed images and detect image distortion in various models. Calculating PSNR and SSIM in the YCbCr color space reduces the influence of chroma on the results and aligns better with human visual perception. We also calculated the number of parameters for each model and the reconstruction time for a single cell image.
During model training, the input HR images are randomly cropped to a size of 192 × 192, and the quantity is increased by random horizontal flips, and 90 , 180 , and 270 rotations. The LR images are obtained by downsampling the corresponding HR images by ×2 using bicubic downsampling, resulting in a size of 96 × 96. We set the weight λ in Equation (8) as 10 3 . For Φ m in Equations (11) and (12), we use the Conv5 layer of the VGG-19. We train our model using the ADAM optimizer with β 1 = 0.9 and β 2 = 0.999 . The learning rate is initialized to 2 × 10 4 and multiplied by 0.5 after the 200-th epoch. Our model is implemented using the PyTorch framework and trained on an NVIDIA 3090 GPU, and training takes approximately 8 to 9 h to complete.

3.2. Comparison with Classic Models

We conducted ablation experiments on our proposed model. We trained it using the collected microscope image dataset and evaluated it on the Cell A dataset. As can be seen from the results in Table 1, the CBAM, RDB, and feature discriminator that we have added to the model have a better PSNR value and SSIM value.
Then, we compare our model with six classic image SR models, including EDSR, Enhanced Super-Resolution Generative Adversarial Network (ESRGAN) [26], very deep Residual Channel Attention Network (RCAN) [27], RDN, SRResnet [17] (less a discriminator compared to SRGAN), and RFGANM. And, from Table 2, it can be seen that our model achieves the highest PSNR and SSIM for reconstructing images. This is because we added RDAB in the generator of our model, allowing it to fully utilize the hierarchical features of the original LR image and focus the reconstruction on the main information of the feature maps. To avoid generating artifacts during reconstruction, we added a feature discriminator, which enables the generator to produce high-frequency features relevant to the HR image, and we made corresponding improvements to the model’s loss function. On the LPIPS metric, the best model is ESRGAN, with our model being a close second. Regarding model parameters, the best model is SRResnet, and our model differs by only 5.91 M. For single image reconstruction time, EDSR is the fastest, and our model is only 0.51 s slower. In summary, compared to other SR models, our model achieves the highest PSNR and SSIM, and good LPIPS, model parameters, and reconstruction times.
In addition, Figure 5 shows the reconstruction detail views of our proposed model and two classic models. It can be observed that the cell detail textures and edge parts reconstructed by our model are closer to the HR images. In summary, these results demonstrate that our network model has achieved excellent performance.

4. Conclusions

In this paper, we provide a microscopic images dataset. And the SRGAN model is improved by modifying the residual block of the generator to an RDB and combining the attention mechanism. Additionally, a feature discriminator is introduced to ensure that the reconstruction avoids generating artifacts and instead produces high-frequency features pertinent to the image. The loss function is also enhanced. Finally, the experimental results demonstrate that our model outperforms the classical models and is hopeful to meet the demand for high-resolution imaging of large volume samples.

Author Contributions

Conceptualization, S.L. and X.W.; methodology, S.L.; software, S.L.; validation, X.W. and S.L.; formal analysis, S.L.; investigation, S.L. and X.W.; data curation, X.W.; writing—original draft preparation, S.L. and X.W.; writing—review and editing, S.L., X.W., X.G. and X.X.; visualization, X.W.; supervision, S.L. and L.Z.; project administration, S.L.; funding acquisition, S.L. and L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Fundamental Research Funds for the Central Universities (Grant No. ZQN-1005) and the Scientific Research Funds of Huaqiao University (Grant No. 21BS118).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author/s.

Acknowledgments

Many thanks for Muwang Chen’s contribution to the analysis method proposed and strong support in this article.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
FOVField of View
SRSuper-Resolution
LRLow-Resolution
HRHigh-Resolution
SISRSingle Image Super-Resolution Reconstruction
RDAGANResidual Dense Attention Generative Adversarial Network
CNNConvolutional Neural Network
VDSRVery Deep Super Resolution
LapSRNLaplacian Pyramid Super-Resolution Network
GANGenerative Adversarial Network
SRGANSuper-Resolution Generative Adversarial Network
EDSREnhanced Deep Super Resolution
RDBResidual Dense Block
RDNResidual Dense Network
RF-GANMRegistration-Free GAN
DCFADeep Fourier Channel Attention
RDABResidual Dense Attention Block
BDBicubic Downsampling
DFFDense Feature Fusion
ReLURectified Linear Unit
GFFGlobal Feature Fusion
GRLGlobal Residual Learning
LFFLocal Feature Fusion
LRLLocal Residual Learning
CBAMConvolutional Block Attention Module
CMContinuous Memory
PSNRPeak Signal-to-Noise Ratio
SSIMStructural SIMilarity index
LPIPSLearned Perceptual Image Patch Similarity
ESRGANEnhanced Super Resolution Generative Adversarial Network
RCANResidual Channel Attention Network

Appendix A

Appendix A.1

Table A1 introduces the image characteristics of the four microscopic cell datasets we collected using the MoticEasyScan NFC 300 microscope. We selected representative cells with distinctive features. The table provides detailed descriptions of the colors and structures within the cells. We hope this will help you understand the features of the training datasets.
Table A1. The characteristics of the four types of cells: Cell A, Cell B, Cell C, and Cell D.
Table A1. The characteristics of the four types of cells: Cell A, Cell B, Cell C, and Cell D.
DatasetImageImage Feature
1Cell ASensors 24 03560 i001The dark purple circular or oval structures in the image represent cell nuclei. The areas surrounding the nuclei are light pink, indicating the cytoplasm and extracellular matrix. There are also some small red circular structures.
2Cell BSensors 24 03560 i002The small dark blue circular or oval structures in the image represent cell nuclei. The cytoplasmic regions appear lighter blue or unstained.
3Cell CSensors 24 03560 i003The dark purple circular structures in the image represent cell nuclei. The areas surrounding the nuclei are light pink, indicating the cytoplasm.
4Cell DSensors 24 03560 i004The small dark blue circular or oval particles in the image are arranged in chains or clusters.

References

  1. Sanderson, J. Multi-Photon Microscopy. Curr. Protoc. 2023, 3, e634. [Google Scholar] [CrossRef] [PubMed]
  2. Tsai, R.; Huang, T.S. Multiframe Image Restoration and Registration. Adv. Comput. Vis. Image Process. 1984, 1, 317–339. [Google Scholar]
  3. Rivenson, Y.; Koydemir, H.C.; Wang, H.; Wei, Z.; Ren, Z.; Günaydın, H.; Zhang, Y.; Göröcs, Z.; Liang, K.; Tseng, D.; et al. Deep Learning Enhanced Mobile-Phone Microscopy. ACS Photonics 2018, 5, 2354–2364. [Google Scholar] [CrossRef]
  4. Wei, O.; Andrey, A.; Mickaël, L.; Xian, H.; Christophe, Z. Deep Learning Massively Accelerates Super-Resolution Localization Microscopy. Nat. Biotechnol. 2018, 36, 460–468. [Google Scholar] [CrossRef] [PubMed]
  5. Nehme, E.; Weiss, L.E.; Michaeli, T.; Shechtman, Y. Deep-STORM: Super-Resolution Single-Molecule Microscopy by Deep Learning. Optics 2018, 5, 458–464. [Google Scholar] [CrossRef]
  6. Xiao, Y.; Yuan, Q.; Jiang, K.; He, J.; Wang, Y.; Zhang, L. From Degrade to Upgrade: Learning a Self-Supervised Degradation Guided Adaptive Network for Blind Remote Sensing Image Super-Resolution. Inf. Fusion 2023, 96, 297–311. [Google Scholar] [CrossRef]
  7. Xiao, Y.; Su, X.; Yuan, Q.; Liu, D.; Shen, H.; Zhang, L. Satellite Video Super-Resolution via Multiscale Deformable Convolution Alignment and Temporal Grouping Projection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5610819. [Google Scholar] [CrossRef]
  8. Xiao, Y.; Yuan, Q.; Jiang, K.; Jin, X.; He, J.; Zhang, L.; Lin, C. Local-Global Temporal Difference Learning for Satellite Video Super-Resolution. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 2789–2802. [Google Scholar] [CrossRef]
  9. Blu, T.; Thévenaz, P.; Unser, M. Linear Interpolation Revitalized. IEEE Trans. Image Process. 2004, 13, 710–719. [Google Scholar] [CrossRef] [PubMed]
  10. Meijering, E.; Unser, M. A Note on Cubic Convolution Interpolation. IEEE Trans. Image Process. 2003, 12, 477–479. [Google Scholar] [CrossRef] [PubMed]
  11. Fan, C.; Wu, C.; Li, G.; Ma, J. Projections onto Convex Sets Super-Resolution Reconstruction Based on Point Spread Function Estimation of Low-Resolution Remote Sensing Images. Sensors 2017, 17, 362. [Google Scholar] [CrossRef] [PubMed]
  12. Li, F.; He, X.; Wei, Z.; Mu, Z.; Li, M. Super-Resolution Image Reconstruction Based on an Improved Maximum a Posteriori Algorithm. J. Beijing Inst. Technol. 2018, 27, 237–240. [Google Scholar] [CrossRef]
  13. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image Super-Resolution Via Sparse Representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  14. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Zurich, Switzerland, 6–12 September 2014; pp. 184–199. [Google Scholar] [CrossRef]
  15. Kim, J.; Lee, J.K.; Lee, K.M. Accurate Image Super-Resolution Using Very Deep Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar] [CrossRef]
  16. Lai, W.S.; Huang, J.B.; Ahuja, N.; Yang, M.H. Deep Laplacian Pyramid Networks for Fast and Accurate Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5835–5843. [Google Scholar] [CrossRef]
  17. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar] [CrossRef]
  18. Lim, B.; Son, S.; Kim, H.; Nah, S.; Lee, K.M. Enhanced Deep Residual Networks for Single Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 1132–1140. [Google Scholar] [CrossRef]
  19. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y.R. Residual Dense Network for Image Super-Resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar] [CrossRef]
  20. Zhang, H.; Fang, C.; Xie, X.; Yang, Y.; Mei, W.; Jin, D.; Fei, P. High-Throughput, High-Resolution Deep Learning Microscopy Based on Registration-Free Generative Adversarial Network. Biomed. Opt. Express 2019, 10, 1044–1063. [Google Scholar] [CrossRef] [PubMed]
  21. Wang, H.; Rivenson, Y.; Jin, Y.; Wei, Z.; Gao, R.; Günaydın, H.; Bentolila, L.A.; Kural, C.; Ozcan, A. Deep Learning Enables Cross-Modality Super-Resolution in Fluorescence Microscopy. Nat. Methods 2019, 16, 103–110. [Google Scholar] [CrossRef] [PubMed]
  22. van Sloun, R.J.G.; Solomon, O.; Bruce, M.; Khaing, Z.Z.; Wijkstra, H.; Eldar, Y.C.; Mischi, M. Super-Resolution Ultrasound Localization Microscopy Through Deep Learning. IEEE Trans. Med. Imaging 2021, 40, 829–839. [Google Scholar] [CrossRef] [PubMed]
  23. Qiao, C.; Li, D.; Guo, Y.; Liu, C.; Jiang, T.; Dai, Q.; Li, D. Evaluation and Development of Deep Neural Networks for Image Super-Resolution in Optical Microscopy. Nat. Methods 2021, 18, 194–202. [Google Scholar] [CrossRef] [PubMed]
  24. Park, S.J.; Son, H.; Cho, S.; Hong, K.S.; Lee, S. SRFeat: Single Image Super-Resolution with Feature Discrimination. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 439–455. [Google Scholar] [CrossRef]
  25. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef]
  26. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Loy, C.C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y.R. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar] [CrossRef]
Figure 1. Residual Dense Attention Generative Adversarial Network (RDAGAN) framework.
Figure 1. Residual Dense Attention Generative Adversarial Network (RDAGAN) framework.
Sensors 24 03560 g001
Figure 2. Residual Dense Attention Blocks (RDABs).
Figure 2. Residual Dense Attention Blocks (RDABs).
Sensors 24 03560 g002
Figure 3. Convolutional Block Attention Module (CBAM).
Figure 3. Convolutional Block Attention Module (CBAM).
Sensors 24 03560 g003
Figure 4. Architecture of our discriminator network.
Figure 4. Architecture of our discriminator network.
Sensors 24 03560 g004
Figure 5. Comparison of reconstructed detail views for Cell A, Cell B, Cell C, and Cell D using our proposed model and two classical models.
Figure 5. Comparison of reconstructed detail views for Cell A, Cell B, Cell C, and Cell D using our proposed model and two classical models.
Sensors 24 03560 g005
Table 1. The ablation experiments results (“✓” denotes the corresponding operation).
Table 1. The ablation experiments results (“✓” denotes the corresponding operation).
CBAMRDBFeature DiscriminatorPSNR/dB ↑SSIM ↑
1 28.120.72
2 28.750.75
3 32.830.85
433.130.87
Table 2. Comparison of PSNR, SSIM, LPIPS, model parameters, and reconstruction times for Cells A, B, C, and D using our proposed model and six classic image SR models (the red indicates the best result).
Table 2. Comparison of PSNR, SSIM, LPIPS, model parameters, and reconstruction times for Cells A, B, C, and D using our proposed model and six classic image SR models (the red indicates the best result).
ModelDatasetPSNR/dB ↑SSIM ↑LPIPS ↑Parameters/MTime/s
RFGANM [20]Cell A28.120.720.20485.922.27
Cell B27.450.850.1108
Cell C28.860.770.1645
Cell D29.620.840.1513
ESRGAN [26]Cell A28.740.750.1089102.162.70
Cell B31.710.900.0292
Cell C30.490.790.0775
Cell D33.500.890.0414
SRResnet [17]Cell A30.980.820.16165.342.33
Cell B33.530.930.0385
Cell C32.620.860.1356
Cell D35.920.930.0572
RCAN [27]Cell A31.060.820.168958.922.17
Cell B33.540.930.0412
Cell C32.660.860.1466
Cell D35.680.930.0590
EDSR [18]Cell A31.630.840.1542155.302.05
Cell B34.070.940.0362
Cell C32.730.860.1414
Cell D36.450.930.0606
RDN [19]Cell A31.680.840.153150.143.23
Cell B34.130.940.0365
Cell C32.810.860.1401
Cell D36.890.940.0609
RDAGANCell A33.130.870.136411.252.56
Cell B36.090.960.0259
Cell C34.290.890.1272
Cell D38.310.960.0550
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, S.; Weng, X.; Gao, X.; Xu, X.; Zhou, L. A Residual Dense Attention Generative Adversarial Network for Microscopic Image Super-Resolution. Sensors 2024, 24, 3560. https://doi.org/10.3390/s24113560

AMA Style

Liu S, Weng X, Gao X, Xu X, Zhou L. A Residual Dense Attention Generative Adversarial Network for Microscopic Image Super-Resolution. Sensors. 2024; 24(11):3560. https://doi.org/10.3390/s24113560

Chicago/Turabian Style

Liu, Sanya, Xiao Weng, Xingen Gao, Xiaoxin Xu, and Lin Zhou. 2024. "A Residual Dense Attention Generative Adversarial Network for Microscopic Image Super-Resolution" Sensors 24, no. 11: 3560. https://doi.org/10.3390/s24113560

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop