Next Article in Journal
Event-Triggered State Estimation for Uncertain Systems with Binary Encoding Transmission Scheme
Next Article in Special Issue
Correction: Chen et al. Online Trajectory Optimization Method for Large Attitude Flip Vertical Landing of the Starship-like Vehicle. Mathematics 2023, 11, 288
Previous Article in Journal
SSDStacked-BLS with Extended Depth and Width: Infrared Fault Diagnosis of Rolling Bearings under Dual Feature Selection
Previous Article in Special Issue
A Hybrid Medium and Long-Term Relative Humidity Point and Interval Prediction Method for Intensive Poultry Farming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Image Denoising Network for Multimedia Teaching System

1
School of Software, Northwestern Polytechnical University, Xi’an 710129, China
2
Research & Development Institute, Northwestern Polytechnical University, Shenzhen 518057, China
3
School of Economics and Management, Harbin Institute of Technology at Weihai, Weihai 264209, China
4
Faculty of Entrepreneurship and Business, Universiti Malaysia Kelantan, Kota Bharu 16100, Malaysia
5
School of AI and Advanced Computing, XJTLU Entrepreneurship College (Taicang), Xi’an Jiaotong-Liverpool University, Suzhou 215400, China
6
School of Power and Energy, Northwestern Polytechnical University, Xi’an 710129, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(17), 3678; https://doi.org/10.3390/math11173678
Submission received: 31 July 2023 / Revised: 22 August 2023 / Accepted: 24 August 2023 / Published: 25 August 2023
(This article belongs to the Special Issue Computational Methods and Application in Machine Learning)

Abstract

:
Due to COVID-19, online education has become an important tool for teachers to teach students. Also, teachers depend on a multimedia teaching system (platform) to finish online education. However, interacted images from a multimedia teaching system may suffer from noise. To address this issue, we propose a lightweight image denoising network (LIDNet) for multimedia teaching systems. A parallel network can be used to mine complementary information. To achieve an adaptive CNN, an omni-dimensional dynamic convolution fused into an upper network can automatically adjust parameters to achieve a robust CNN, according to different input noisy images. That also enlarges the difference in network architecture, which can improve the denoising effect. To refine obtained structural information, a serial network is set behind a parallel network. To extract more salient information, an adaptively parametric rectifier linear unit composed of an attention mechanism and a ReLU is used into LIDNet. Experiments show that our proposed method is effective in image denoising, which can also provide assistance for multimedia teaching systems.

1. Introduction

Traditional teaching requires students to learn knowledge through face-to-face methods. Although it has good effects, it has higher requirements for students in terms of time and space. To break these limitations, online education has been developed. It mainly depends on a multimedia system (platform) to complete teaching tasks. Also, obtained images in the multimedia system constitute important media for human-to-human interaction. However, these images often suffer from some challenges from noise caused by camera shake, hardware quality, weather [1], etc. After analyzing the process of collecting and disseminating relevant teaching resources, teaching images often suffer from challenges, i.e., noise from collection equipment. To address these mentioned drawbacks, image denoising techniques are also applied.
An image denoising technique is a classical low-level technique and has been applied in various fields, i.e., activities recognition [2] and remote sensing [3]. For instance, an expected patch log likelihood (EPLL) [4] used a mixed Gaussian model to learn prior knowledge from many natural image blocks for image denoising. Also, block matching and three-dimensional filtering (BM3D) [5] utilized collaborative filtering on similar two-dimensional image blocks to remove noise. A weighted nuclear norm minimization (WNNM) algorithm can exploit an image’s non-local self-similarity to extract more information for image denoising [6]. Although these methods can restore images, they face some challenges. That is, they excessively rely on manual adjustment of parameters and complex parameters. Due to strong expressive ability, convolutional neural networks (CNNs) have obtained abilities of feature extraction. Thus, CNNs have been applied in the field of image denoising. For instance, a denoising convolutional neural network (DnCNN) first utilized convolution and residual learning operations to complete denoising work [7]. To suppress the influence of the background on noise, an attention mechanism is fused in a CNN to separate background and foreground to suppress noise [8]. To address image denoising under complex scenes, a dynamic convolution is used in a CNN to achieve an adaptive denoiser, according to different noisy images [9]. To obtain a better denoising effect, a combination of an omni-dimensional dynamic convolution and attention mechanisms is integrated into a CNN to enhance the expressive ability of a denoising network, which can enhance interaction quality of the multimedia teaching system between students and teachers. Le et al. uses two phases, i.e., a feature augmentation stage and a feature refinement stage, to design a CNN to extract more accurate structural information for image denoising [10]. To reduce the complexity of a denoiser, Lin et al. simplified the residual spatial–spectral module and knowledge distillation to achieve a lightweight method to accelerate noise removal [11] Alternatively, a combination of a non-local algorithm and a residual CNN achieves a lightweight CNN to suppress noise [12]. That is, we present a lightweight image denoising network as well as LIDNet for multimedia teaching systems. LIDNet uses a parallel sub-network to mine complementary information for image denoising. To achieve an adaptive CNN, a dynamic convolution based on kernel information and input channel number and output channel number fused into an upper network can automatically adjust parameters to achieve a robust CNN, according to different input noisy images. That also enlarges differences in network architecture, which can improve the denoising effect. To refine the obtained structural information, a serial network is set behind a parallel network. To extract more salient information, an adaptively parametric rectifier linear unit composed of an attention mechanism and a ReLU is used in LIDNet. Experiments show that our proposed method is effective in image denoising, which may also provide assistance for multimedia teaching systems.
The contributions of the proposed method can be summarized as follows:
  • A dynamic convolution based on kernel information and input channel number and output channel number is used to adaptively mine more useful information, according to different input images.
  • A combination of attention mechanism and ReLU is set behind each convolutional layer in addition to the final convolutional layer to enhance the same distributions of training samples for pursuing better denoising performance.
  • Our denoising method is useful for enhancing the interaction quality of a multimedia teaching system between teacher and student.
The remainder of this paper is organized as follows. Section 2 lists related work about image denoising based on dual networks and dynamic networks. Section 3 provides detailed information of the proposed method. Section 4 presents analysis of our proposed method and results. Section 5 gives the conclusion of this paper.

2. Related Work

2.1. A Dual Network for Image Denoising

To extract complementary information, dual networks are developed in image denoising [13]. For instance, Tian et al. [13] presented a dual denoising network with sparse mechanism as well as DudeNet to extract complementary information to enhance denoising effects. Alternatively, Bai et al. [14] achieved a dual network via encoder–decoder and channel attention architecture to extract local and non-local information for image denoising, where image spatial details and semantic information can be obtained by a criss-cross attention. To extract more information, Holla et al. [15] used edge information to design a CNN to capture high-frequency information in image denoising. Zhang et al. fused different masks into a CNN to facilitate complementary information to suppress noise [16]. To mine more high-frequency information, Qiao et al. combined two different networks and a sharpening loss function to improve the quality of visual denoising images [17]. Liu et al. used a wavelet decomposition technique to achieve a wide CNN to prevent vanishing and exploding gradient problems [18]. To extract salient noise information, Chen et al. fused a CNN and a transformer to implement a parallel network to extract structural information and key information based on pixel relations for improving denoising effects [19]. For medical noisy image denoising, Jiang et al. [20] used residual connections and dilated convolutions to achieve a heterogeneous dual network to mine more complementary information to suppress noise. According to the mentioned illustrations, we can see that dual networks are useful for image denoising. Inspired by that, we design a dual network architecture for image denoising in this paper.

2.2. Dynamic Networks for Image Denoising

To enhance the robustness of the image denoiser, a dynamic network is created [21]. For instance, Song et al. [21] combined dynamic convolutions and residual learning operations into a CNN to dynamically adjust parameters to obtain a robust denoising network, according to different input images. Du et al. [22] exploited a dynamic attention mechanism to better extract salient information for image denoising. Alternatively, Shen et al. [23] fused a spatial module and dynamic convolution to obtain more spatial context information to obtain better denoising performance. Tian et al. [9] used dynamic convolution and wavelet transform to extract more useful information to improve denoising effects. According to the mentioned descriptions, we can see that dynamic convolution is effective in image denoising. Motivated by that, we use a dynamic convolution in this paper, according to different kernel and channel information.

3. Proposed Method

3.1. Network Architecture

The proposed 17-layer LIDNet combines a parallel and series architecture. The parallel architecture is composed of a 6-layer block called the dynamic feature extraction block (DFEB) and a 6-layer block named complementary feature extraction block (CFEB). The series architecture contains an 11-layer block called the cascaded purification block (CPB), which is shown in Figure 1. DFEB uses a dynamic convolutional layer to adaptively extract structural information, including kernel information and channel information. To extract complementary information, CFEB use several stacked convolutional layers, BN and a combination of attention mechanism and activation function to extract complementary salient information. Also, a residual learning operation is used to connect the obtained information from a parallel network. To prevent over enhancement, a 11-layer CPB is designed behind the parallel network. To construct a clean image, a residual learning operation is used to act between an input image and output image of LIDNet. This process can be shown as Equation (1).
I C = L I D N e t ( I N ) = C P B ( D F E B ( I N ) + C F E B ( I N ) ) + ( I N ) ,
where I C represents an output of LIDNet, which is regraded to a denoised image. I N denotes the input noisy image, and L I D N e t ( ) expresses a function of LIDNet. D F E B , C F E B , and C P B stand for functions of DFEB, CFEB, and CPB, respectively. + is a residual learning operation, which is also shown as ⊕ in Figure 1. Furthermore, the MSE loss function of LIDNet is introduced in Section 3.2.

3.2. Loss Function

To fairly compare with the famous denoising benchmark of DnCNN, a mean squared error (MSE) [24] is chosen as the loss function to train LIDNet. Specifically, MSE uses pairs of I N i , I C i ( 1 i n ) to train our LIDNet in a supervised way, where I N i and I C i are defined as the i-th noisy and clean image, respectively. n represents the number of image pairs in the training dataset. LIDNet also uses the popular Adam [25] to obtain reasonable parameters. The mathematical expression of the loss function is as follows:
L ( θ ) = 1 2 n i = 1 n L I D N e t ( I N i ) I C i 2 ,
where L is a loss function of MSE and θ stands for learned parameters.

3.3. Dynamic Feature Extraction Block

The first layer in DFEB consists of a convolutional and a rectified linear unit (ReLU) [26] operation. The following 4 layers are composed of a convolutional, a batch normalization (BN) operation and an adaptively parametric rectified linear unit (APReLU) [27]. And the final layer has an omni-dimensional dynamic convolution (ODConv) [28], BN and APReLU. In terms of parameter setting, the input channel number of the first convolutional operation is the same with the channel of the input images. If the input image is color, input channel number of LIDNet is 3. otherwise, input channel number of LIDNet is 1. Other numbers of input and output channels of all the layers are 64. Every size of convolutional kernels in LIDNet is set to 3 × 3 . And the output of the DFEB is fused with the output of the CFEB via a concatenation connection. Mathematical expression of the DFEB is shown as follows:
O D F E B = D F E B ( I N ) = C o n v 6 ( 4 C o n v 2 ( C o n v 1 ( I N ) ) ) = A P R e L U ( B N ( O D C o n v ( 4 A P R e L U ( 4 B N ( 4 C o n v ( R e L U ( C o n v ( I N ) ) ) ) ) ) ) ) ,
where O D F E B is the output of DFEB. D F E B ( ) expresses a function of DFEB. C o n v 1 means the first layer of the DFEB, 4 C o n v 2 means 4 stacked layers in DFEB, which form the second layer to the fifth layer, and C o n v 6 means the last layer of the DFEB. C o n v stands for a function of a convolutional operation, R e L U stands for an activation function of ReLU, B N stands for the batch normalization operation, A P R e L U stands for another activation function of APReLU, and O D C o n v stands for ODConv operation. 4 A P R e L U ( 4 B N ( 4 C o n v ( ) ) ) is the equation for the 4 C o n v 2 .

3.4. Complementary Feature Extraction Block

The lower branch in a parallel architecture is CFEB, which is responsible for extracting complementary features by a different network architecture. The first layer in CFEB consists of a convolutional and a ReLU operation. And the following 5 layers are composed of a stacked combination of convolutional, BN, and APReLU operations. As shown in Figure 1, the difference between DFEB and CFEB is mainly reflected on the last convolutional layer. Specifically, the CFEB uses a common convolution operation to replace the ODConv as the final layer in the DFEB. Input and output channel numbers of the final convolutional layer are both 64. The mathematical expression of CFEB is as follows:
O C F E B = C F E B ( I N ) = 5 C o n v 2 ( C o n v 1 ( I N ) ) = 5 A P R e L U ( 5 B N ( 5 C o n v ( R e L U ( C o n v ( I N ) ) ) ) ) ,
where O C F E B is the output of CFEB. C F E B ( ) expresses a function of CFEB. C o n v 1 means the first layer of the CFEB and 5 C o n v 2 means 5 stacked layers in the DFEB, which is as the second layer to the sixth layer in the CFEB. 5 A P R e L U ( 5 B N ( 5 C o n v ( ) ) ) is equal to 5 C o n v 2 .

3.5. Concatenated Purification Block

To refine fused structural information from DFEB and CFEB, CPB is set as the last part of LIDNet. Specifically, its first 10 layers in CPB are composed of convolutional, BN, and APReLU operations. And its last layer is simply a common convolutional operation, which is used to construct clean images. To construct a clean image, a residual learning operation is used to act between an input image and output image of LIDNet. The numbers of input and output channels are 64 except the output channel number of the final convolutional layer, which is the same as the channel of the input image. The mathematical expression of CPB is as follows:
O C P B = C P B ( O D F E B + O C F E B ) = C o n v 11 ( 10 C o n v 1 ( O D F E B + O C F E B ) ) = C o n v ( 10 A P R e L U ( 10 B N ( 10 C o n v ( O D F E B + O C F E B ) ) ) ) ,
I C = I N O C P B ,
where O C P B is the output of CPB. C P B ( ) expresses a function of CPB. 10 C o n v 1 means the 10 stacked layers in CPB, which form the first layer to the tenth layer, and C o n v 11 means the last layer of the CPB. 10 A P R e L U ( 10 B N ( 10 C o n v ( ) ) ) is equal to 10 C o n v 1 .

4. Experiments

4.1. Datasets

The video quality of many courses inevitably declines due to the impact of the environment and equipment during shooting. To achieve better performance in multimedia, we propose LIDNet to denoise these teaching images. The architecture of our LIDNet is shown in Figure 1.
For image denoising with Gaussian noise, 400 images with sizes of 180 × 180 from Ref. [29] are used to train a denoising model. Three different denoising models with noise levels of 15 and 25 can be trained, respectively. To train a blind denoising model with noise levels from 0 to 55, a blind model is trained. Specifically, patch sizes are set to 40 × 40 .
To fairly test denoising performance, public BSD68 [30], Set 12 [31], Kodak24 [32], and collected educational images from the Internet are used as test datasets. Guassian noise with noise levels of 15 and 25 is added on BSD68, Set12, Kodak24, and collected educational images from the Internet to test the denoising performance of the proposed method.

4.2. Parameter Setting

This paper has the following experimental settings. The number of training epochs is 180. The original learning rate is 1 × 10 3 and it will decline to 0.2 times when the epoch is 30, 60, and 90, respectively. Batch size is set to 128. Adam is used to optimize parameters [25], where β 1 is 0.9 and β 2 is 0.999. More parameters can be found in Ref. [13]. The LIDNet can be trained on a PC with Intel Xeon Gold 6330 Processor and one Nvidia GeForce RTX 3090. Furthermore, all the codes run on Ubuntu 20.04 with Python 3.8, PyTorch 1.11.0, and CUDA 11.7.

4.3. Network Analysis

This paper uses a parallel network architecture to extract complementary information for image denoising, where a parallel network consists of an upper network (also regarded as dynamic feature extraction block, DFEB) and lower network (also regarded as complementary feature extraction block, CFEB). It connects a serial architecture (concatenated purification block, CPB) to extract more hierarchical structural information. Also, each branch in the parallel network is composed of six layers of stacked architecture. The upper network is composed of a Conv + APReLU, four Conv + BN + APReLU, and a ODConv + BN + APReLU, where APReLU [27] is composed of an attention mechanism and a ReLU is used to extract salient information and nonlinear information. Also, ODConv [28] utilizes convolutional kernel information and channel information to dynamically learn parameters to adaptively train a denoising model for different given noisy images. ‘LIDNet without global residual connection and ODConv + BN + APReLU’ has an improvement of 0.013dB compared to ‘LIDNet with only Conv + APReLU in the upper network’ in Table 1, which describes the effectiveness of four Conv + BN + APReLU in the upper network for image denoising. Also, the denoising effect of ‘Conv + APReLU’ in the upper network is verified by ‘The combination of lower network and CPB’ and ‘LIDNet with only Conv + APReLU in upper network and without global residual connection’ in Table 1. To test the denoising performance of DFEB, we use ‘The combination of lower network and CPB’ and ‘LIDNet without global residual connection’ to conduct comparative experiments. As shown in Table 1, we can see that ‘LIDNet without global residual connection’ exceeds ‘The combination of lower network and CPB’ in terms of PSNR. That shows that DFEB in the parallel network is effective for image denoising. Additionally, to test complementarity of two sub-networks, ‘The combination of lower network and CPB’ is superior to ‘CPB’ in terms of image denoising in Table 1, which shows the superiority of a parallel network for image denoising. To prevent the interference of upper and lower networks, a serial network is set behind a parallel network to refine the obtained structural information for image denoising. Finally, a global residual connection is employed between outputs of the first layer in a lower network and the last layer in the CPB to construct clean images.

4.4. Comparisons with State of the Art

To test the effectiveness of proposed method, we choose several popular denoising methods, i.e., EPLL, BM3D, WNNM, DnCNN, image restoration CNN (IRCNN) [33], fast and flexible denoising network (FFDNet) [34], and a cascade of shrinkage fields (CSF) [35] as comparative methods on the BSD68 and Set12 to conduct experiments. As shown in Table 2, we can see that our LIDNet has obtained the best denoising result on the BSD68 for σ = 15 and σ = 25 . For instance, our LIDNet has an improvement of 0.11 dB compared to IRCNN for σ = 15 . That shows that our method is effective for gray noisy image denoising. To verify good denoising performance for a single gray noisy image, different methods on Set12 are used to conduct denoising effects. As illustrated in Table 3, we can see that our LIDNet has obtained the best denoising effect for single noisy image denoising. For instance, our LIDNet has obtained an improvement of 0.09 dB compared to a popular denoising method, i.e., WNNM for a noise level of 15. That shows that our method is a good denoising tool for low-frequency noisy image denoising. Our LIDNet has obtained an improvement of 0.06 dB compared to a popular denoising method, i.e., IRCNN for noise level of 25. That shows that our method is a good denoising tool for high-frequency noisy image denoising. According to that, we can see that our method is effective for single noisy image denoising. Furthermore, to further demonstrate the denoising performance of our LIDNet on color images, Table 4 records the denoising results from different models with different noise levels. Compared with popular methods, i.e., IRCNN, FFDNet, D-BSN, FL(NLM), and FL(BM3D), our LIDNet has also achieved improvements in denoising performance for color noisy images. This also proved the effectiveness of LIDNet in processing color noisy images.
To comprehensively test the denoising effect of our proposed method, we use qualitative analysis to measure the effects of visual images. Specifically, we choose one area of denoising images from BM3D, FFDNet, IRCNN, and LIDNet as observation areas. If the observation area is clearer, its corresponding method shows better denoising performance. As shown in Figure 2, Figure 3 and Figure 4, we can see that our LIDNet is clearer than the results of other methods. In Figure 3, other methods can obtain more incorrect texture information. Because real noisy images are difficult to obtain in the world, we choose Guassian noise added on educational images to test the performance of the proposed method for educational image denoising. In Figure 4, we can see that our method can obtain clearer detailed information for noisy educational image denoising. Thus, that not only shows that our method is superior to other methods in terms of qualitative analysis, but also that it is robust for different scenes in terms of image denoising.

5. Conclusions

Multimedia teaching systems have become a popular tool for online education. However, interacted images from a multimedia teaching system may suffer from noise. In this paper, we present a lightweight image denoising network as well as LIDNet for multimedia teaching systems. LIDNet uses a parallel network to mine complementary information. To improve robustness of the obtained denoiser, an omni-dimensional dynamic convolution is designed in one sub-network from the parallel network to automatically adjust parameters to achieve an adaptive CNN. That also enlarges the differences in network architecture, which can improve the denoising effect. To refine the obtained structural information, a serial network is set behind a parallel network. To extract more salient information, an adaptively parametric rectifier linear unit composed of an attention mechanism and a ReLU is used in LIDNet. Experiments show that our LIDNet is effective in image denoising, which can also provide assistance for multimedia teaching systems.

Author Contributions

Methodology, X.Z. and C.T.; software, X.Z., Q.Z. and H.-S.G.; validation, X.Z., C.T., Q.Z. and H.-S.G.; formal analysis, X.Z., C.T., Q.Z. and T.C.; writing—original draft, X.Z.; writing—review & editing, X.Z., C.T., Q.Z., T.C. and M.A.H.I.; supervision, Q.Z.; project administration, C.T.; funding acquisition, C.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Guangdong Basic and Applied Basic Research Foundation Grant under 2021A1515110079, in part by Shandong Natural Science Foundation under ZR2023OG074, in part by the Ideological and Political Education of Financial Decision Support System under KVSZZZ202315, in part by Collaborative Education by the Ministry of Education under 220501210164954, in part by Teaching Education Reform of NPU under 06410-23GZ230106.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

The authors would like to thank the Guangdong Basic and Applied Basic Research Foundation, Shandong Natural Science Foundation, Ideological and Political Education of Financial Decision Support System, Collaborative Education by the Ministry of Education and Teaching Education Reform for supporting this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khorov, E.; Kureev, A.; Levitsky, I.; Akyildiz, I.F. A phase noise resistant constellation rotation method and its experimental validation for NOMA Wi-Fi. IEEE J. Sel. Areas Commun. 2022, 40, 1346–1354. [Google Scholar] [CrossRef]
  2. Gu, F.; Khoshelham, K.; Valaee, S.; Shang, J.; Zhang, R. Locomotion activity recognition using stacked denoising autoencoders. IEEE Internet Things J. 2018, 5, 2085–2093. [Google Scholar] [CrossRef]
  3. Liu, P.; Huang, F.; Li, G.; Liu, Z. Remote-sensing image denoising using partial differential equations and auxiliary images as priors. IEEE Geosci. Remote Sens. Lett. 2011, 9, 358–362. [Google Scholar] [CrossRef]
  4. Zoran, D.; Weiss, Y. From learning models of natural image patches to whole image restoration. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 479–486. [Google Scholar]
  5. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  6. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  7. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a gaussian denoiser: Residual learning of deep cnn for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  8. Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef] [PubMed]
  9. Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, D. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
  10. Le, T.H.; Lin, P.H.; Huang, S.C. LD-Net: An efficient lightweight denoising model based on convolutional neural network. IEEE Open J. Comput. Soc. 2020, 1, 173–181. [Google Scholar] [CrossRef]
  11. Lin, Y.; Cai, Z.; Li, J.; Zhang, J. Lightweight Remote Sensing Image Denoising via Knowledge Distillation. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP), Shanghai, China, 26–28 September 2022; pp. 1–7. [Google Scholar]
  12. Guo, Y.; Davy, A.; Facciolo, G.; Morel, J.-M.; Jin, Q. Fast, nonlocal and neural: A lightweight high quality solution to image denoising. IEEE Signal Process. Lett. 2021, 28, 1515–1519. [Google Scholar] [CrossRef]
  13. Tian, C.; Xu, Y.; Zuo, W.; Du, B.; Lin, C.-W.; Zhang, D. Designing and training of a dual CNN for image denoising. Knowl.-Based Syst. 2021, 226, 106949. [Google Scholar] [CrossRef]
  14. Bai, Y.; Liu, M.; Yao, C.; Lin, C.; Zhao, Y. MSPNet: Multi-stage progressive network for image denoising. Neurocomputing 2023, 517, 71–80. [Google Scholar] [CrossRef]
  15. Holla, S.; Park, N.; Lee, B. EFID: Edge-Focused Image Denoising Using a Convolutional Neural Network. IEEE Access 2023, 11, 9613–9626. [Google Scholar]
  16. Zhang, D.; Zhou, F.; Jiang, Y.; Fu, Z. Mm-bsn: Self-supervised image denoising for real-world with multi-mask based on blind-spot network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 24–31 January 2023; pp. 4188–4197. [Google Scholar]
  17. Qiao, S.; Yang, J.; Zhang, T.; Zhao, C. Layered input GradiNet for image denoising. Knowl.-Based Syst. 2022, 254, 109587. [Google Scholar] [CrossRef]
  18. Liu, G.; Dang, M.; Liu, J.; Xiang, R.; Tian, Y.; Luo, N. True wide convolutional neural network for image denoising. Inf. Sci. 2022, 610, 171–184. [Google Scholar] [CrossRef]
  19. Chen, Y.; Yin, M.; Li, Y.; Cai, Q. CSU-Net: A CNN-Transformer parallel network for multimodal brain tumour segmentation. Electronics 2022, 11, 2226. [Google Scholar] [CrossRef]
  20. Jiang, X.; Jin, Y.; Yao, Y. Low-dose CT lung images denoising based on multiscale parallel convolution neural network. Vis. Comput. 2021, 37, 2419–2431. [Google Scholar] [CrossRef]
  21. Song, Y.; Zhu, Y.; Du, X. Dynamic residual dense network for image denoising. Sensors 2019, 19, 3809. [Google Scholar] [CrossRef]
  22. Du, Y.; Han, G.; Tan, Y.; Xiao, C.; He, S. Blind image denoising via dynamic dual learning. IEEE Trans. Multimed. 2020, 23, 2139–2152. [Google Scholar] [CrossRef]
  23. Shen, H.; Zhao, Z.-Q.; Zhang, W. Adaptive Dynamic Filtering Network for Image Denoising. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; pp. 2227–2235. [Google Scholar]
  24. Douillard, C.; Jézéquel, M.; Berrou, C.; Electronique, D.; Picart, A.; Didier, P.; Glavieux, A. Iterative correction of intersymbol interference: Turbo-equalization. Eur. Trans. Telecommun. 1995, 6, 507–511. [Google Scholar] [CrossRef]
  25. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 2012: 26th Annual Conference on Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–6 December 2012; Volume 25. [Google Scholar]
  27. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Dong, S.; Pecht, M. Deep residual networks with adaptively parametric rectifier linear units for fault diagnosis. IEEE Trans. Ind. Electron. 2020, 68, 2587–2597. [Google Scholar] [CrossRef]
  28. Li, C.; Zhou, A.; Yao, A. Omni-dimensional dynamic convolution. arXiv 2022, arXiv:2209.07947. [Google Scholar]
  29. Chen, Y.; Pock, T. Trainable nonlinear reaction diffusion: A flexible framework for fast and effective image restoration. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1256–1272. [Google Scholar] [CrossRef] [PubMed]
  30. Wu, X.; Liu, M.; Cao, Y.; Ren, D.; Zuo, W. Unpaired learning of deep image denoising. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 352–368. [Google Scholar]
  31. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  32. Franzen, R. Kodak Lossless True Color Image Suite. 1999. Available online: https://r0k.us/graphics/kodak/ (accessed on 15 November 1999).
  33. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning deep CNN denoiser prior for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3929–3938. [Google Scholar]
  34. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed]
  35. Schmidt, U.; Roth, S. Shrinkage fields for effective image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2774–2781. [Google Scholar]
Figure 1. Network architecture of LIDNet.
Figure 1. Network architecture of LIDNet.
Mathematics 11 03678 g001
Figure 2. Visual effects of several denoising methods on an image from Set12 for noise level of 15. (a) Original image, (b) noisy image (24.60 dB), (c) BM3D [4] (30.98 dB), (d) FFDNet [34] (31.99 dB), (e) IRCNN [33] (32.02 dB), and (f) LIDNet (Ours) (32.21 dB).
Figure 2. Visual effects of several denoising methods on an image from Set12 for noise level of 15. (a) Original image, (b) noisy image (24.60 dB), (c) BM3D [4] (30.98 dB), (d) FFDNet [34] (31.99 dB), (e) IRCNN [33] (32.02 dB), and (f) LIDNet (Ours) (32.21 dB).
Mathematics 11 03678 g002
Figure 3. Visual effects of several denoising methods on an image from BSD68 for noise level of 25. (a) Original image, (b) noisy image (20.19 dB), (c) BM3D [4] (29.49 dB), (d) FFDNet [34] (30.04 dB), (e) IRCNN [33] (30.07 dB), and (f) LIDNet (Ours) (30.14 dB).
Figure 3. Visual effects of several denoising methods on an image from BSD68 for noise level of 25. (a) Original image, (b) noisy image (20.19 dB), (c) BM3D [4] (29.49 dB), (d) FFDNet [34] (30.04 dB), (e) IRCNN [33] (30.07 dB), and (f) LIDNet (Ours) (30.14 dB).
Mathematics 11 03678 g003
Figure 4. Visual results of several denoising methods on a real teaching image for noise level of 25. (a) Original image, (b) noisy image (20.18 dB), (c) BM3D [4] (36.22 dB), (d) IRCNN [33] (36.50 dB), (e) FFDNet [34] (36.73 dB) and (f) LIDNet (Ours) (36.82 dB).
Figure 4. Visual results of several denoising methods on a real teaching image for noise level of 25. (a) Original image, (b) noisy image (20.18 dB), (c) BM3D [4] (36.22 dB), (d) IRCNN [33] (36.50 dB), (e) FFDNet [34] (36.73 dB) and (f) LIDNet (Ours) (36.82 dB).
Mathematics 11 03678 g004
Table 1. Denoising results (average PSNR (dB)) of several networks on BSD68 for noise level of 25.
Table 1. Denoising results (average PSNR (dB)) of several networks on BSD68 for noise level of 25.
Different NetworksPSNR(dB)
CPB28.937
The combination of lower network and CPB28.944
LIDNet with only Conv + APReLU in upper network and without global residual connection29.219
LIDNet without global residual connection and ODConv + BN + APReLU29.232
LIDNet without global residual connection29.237
LIDNet (Ours)29.247
Table 2. Average PSNR (dB) results of several networks on BSD68 for noise levels of 15 and 25.
Table 2. Average PSNR (dB) results of several networks on BSD68 for noise levels of 15 and 25.
MethodsEPLL [4]BM3D [5]WNNM [6]DnCNN [7]IRCNN [23]FFDNet [34]CSF [35]LIDNet (Ours)
σ = 15 31.2131.0731.3731.7331.6331.6231.2431.74
σ = 25 28.6828.5728.8329.2329.1529.1928.7429.25
Table 3. PSNR (dB) results of different methods on Set12 with noise levels of 15 and 25.
Table 3. PSNR (dB) results of different methods on Set12 with noise levels of 15 and 25.
ImagesC.manHousePeppersStarfishMonarchAirplaneParrotLenaBarbaraBoatManCoupleAverage
Noise level15
EPLL [4]31.834.1732.6431.1332.1031.1931.4233.9231.3831.9332.0031.9332.14
BM3D [5]31.9134.9332.6931.1431.8531.0731.3734.2633.1032.1331.9232.1032.37
WNNM [6]32.1735.1332.9931.8232.7131.3931.6234.2733.6032.2732.1132.1732.70
IRCNN [33]32.5534.8933.3132.0232.8231.7031.8434.5332.4332.3432.4032.4032.77
FFDNet [34]32.4335.0733.2531.9932.6631.5731.8134.6232.5432.3832.4132.4632.77
CSF [35]31.9534.3932.8531.5532.3331.3331.3734.0631.9232.0132.0831.9832.32
LIDNet (Ours)31.9335.0333.2432.2133.0931.7331.9334.5432.5732.4032.2932.4632.79
Noise level25
EPLL [4]29.2632.1730.1728.5129.3928.6128.9531.7328.6129.7429.6629.5329.69
BM3D [5]29.4532.8530.1628.5629.2528.4228.9332.0730.7129.9029.6129.7129.97
WNNM [6]29.6433.2230.4229.0329.8428.6929.1532.2431.2430.0329.7629.8230.26
IRCNN [33]30.0833.0630.8829.2730.0929.1229.4732.4329.9230.1730.0430.0830.38
FFDNet [34]30.1033.2830.9329.3230.0829.0429.4432.5730.0130.2530.1130.2030.44
CSF [35]29.4832.3930.3228.8029.6228.7228.9031.7929.0329.7629.7129.5329.84
LIDNet (Ours)30.2033.1030.8329.3930.4029.0929.5232.3930.0430.1830.0430.0830.44
Table 4. Average PSNR (dB) results of different methods on CBSD68 and Kodak24 datasets with noise levels of 15 and 25.
Table 4. Average PSNR (dB) results of different methods on CBSD68 and Kodak24 datasets with noise levels of 15 and 25.
DatasetsModels σ = 15 σ = 25
CBSD68CBM3D [5]33.5230.71
DnCNN [7]33.9831.31
IRCNN [33]33.8631.16
FFDNet [34]33.8031.18
D-BSN [30]33.5630.61
FL (NLM) [12]-31.01
FL (BM3D) [12]-31.13
LIDNet (Ours)33.9931.37
Kodak24CBM3D [5]34.2831.68
DnCNN [7]34.7332.23
IRCNN [33]34.5632.03
FFDNet [34]34.5532.11
D-BSN [30]33.7431.64
FL (NLM) [12]-32.11
FL (BM3D) [12]-32.26
LIDNet (Ours)34.5732.12
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Tian, C.; Zhang, Q.; Gan, H.-S.; Cheng, T.; Ibrahim, M.A.H. Lightweight Image Denoising Network for Multimedia Teaching System. Mathematics 2023, 11, 3678. https://doi.org/10.3390/math11173678

AMA Style

Zhang X, Tian C, Zhang Q, Gan H-S, Cheng T, Ibrahim MAH. Lightweight Image Denoising Network for Multimedia Teaching System. Mathematics. 2023; 11(17):3678. https://doi.org/10.3390/math11173678

Chicago/Turabian Style

Zhang, Xuanyu, Chunwei Tian, Qi Zhang, Hong-Seng Gan, Tongtong Cheng, and Mohd Asrul Hery Ibrahim. 2023. "Lightweight Image Denoising Network for Multimedia Teaching System" Mathematics 11, no. 17: 3678. https://doi.org/10.3390/math11173678

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop