Next Article in Journal
Automatic Estimation of Tropical Cyclone Centers from Wide-Swath Synthetic-Aperture Radar Images of Miniaturized Satellites
Previous Article in Journal
Analysis of Penalty Kick Performance in the Spanish Football League: A Longitudinal Study
Previous Article in Special Issue
Unsupervised Scene Image Text Segmentation Based on Improved CycleGAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Detail–Noise Complementary Learning for Image Denoising

1
College of Mathematics and Information Science, Nanjing Normal University of Special Education, Nanjing 210038, China
2
School of Software, Nanjing University of Information Science and Technology, Nanjing 210044, China
3
State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210023, China
4
Jiangsu Province Engineering Research Center of Advanced Computing and Intelligent Services, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(16), 7044; https://doi.org/10.3390/app14167044 (registering DOI)
Submission received: 2 July 2024 / Revised: 4 August 2024 / Accepted: 8 August 2024 / Published: 11 August 2024
(This article belongs to the Special Issue Advances in Neural Networks and Deep Learning)

Abstract

:
Deep convolutional neural networks (CNNs) have demonstrated significant potential in enhancing image denoising performance. However, most denoising methods fuse different levels of features through long and short skip connections, easily generating a lot of redundant information, thereby weakening the complementarity of different levels of features, resulting in the loss of image details. In this paper, we propose a multi-scale detail–noise complementary learning (MDNCL) network for additive white Gaussian noise removal and real-world noise removal. The MDNCL network comprises two branches, namely the Detail Feature Learning Branch (DLB) and the Noise Learning Branch (NLB). Specifically, a loss function is applied to guide the complementary learning of image detail features and noisy mappings in these two branches. This learning approach effectively balances noise reduction and detail restoration, especially when dealing with high ratios of noise. To enhance the complementarity of features between different network layers and avoid redundant information, we designed a Feature Subtraction Unit (FSU) to capture the differences in features across the DLB network layers. Our extensive experimental evaluations demonstrate that the MDNCL approach achieves impressive denoising performance and outperforms other popular denoising methods.

1. Introduction

Image denoising has become increasingly critical in computer vision tasks, such as satellite remote sensing [1], video object detection [2], industrial inspection [3], and medical imaging [4]. This is primarily due to the heightened quality requirements for input images in these domains. The objective of image denoising is to recover a clean image, denoted as x, from a model defined as x = y n , where y represents the noisy image and n represents the additive white Gaussian noise (AWGN) with a standard deviation of σ [5].
Traditional image denoising methods leverage the inherent structural characteristics of images, such as sparsity, low rank, and non-local similarity. Aharon et al. [6] addressed the limited adaptability of fixed dictionaries by successfully applying the K-singular value decomposition (K-SVD) algorithm to construct sparse dictionaries. Dabov et al. [7] introduced the block-matched three-dimensional (BM3D) filtering method, which combines non-local self-similarity and frequency-domain transformation to search for similar blocks. Wu et al. [8] proposed an enhanced Weighted Nuclear Norm Minimization (WNNM) algorithm that incorporates adaptive median filtering and the properties of WNNM to suppress noise by constraining the parameters of low-rank matrices. Although traditional image denoising methods can deliver satisfactory denoising results, they often involve computationally expensive procedures and rely heavily on prior knowledge of the image.
Compared to traditional denoising algorithms, deep convolutional neural networks (CNNs) have shown significant improvements in image denoising tasks by automatically learning network parameters from large datasets without relying on prior assumptions. DnCNN [9] generates residual images containing only noise through residual learning and improves the performance of the denoising network combined with batch normalization (BN). Based on the DnCNN model, Zhang et al. [10] successively proposed a fast and flexible denoising network (FFDNet). The input of FFDNet is noise-level estimation, which can solve the more complex real-world image denoising problem. However, as the depth of the network increases, the deep network layers may learn similar or duplicate feature representations, resulting in redundant features and degraded denoising performance [11]. Therefore, the researchers denoised images by increasing the width of the network or using skip join operations. Gurrola et al. [12] proposed a residual dense neural network (RDUNet) based on a densely connected hierarchical network for image denoising. The encoding layer and decoding layer of RDUNet are composed of densely connected convolutional layers to reuse feature maps and local residual learning, avoid gradient disappearance, and speed up the learning process. Qu et al. [13] proposed a simple and effective multiscale denoising back-projection feature fusion network to recover underwater images affected by complex degradation, which solves the limitations of traditional encoder–decoder methods in detail recovery and complex degradation processing. Liu et al. [14] constructed a wide CNN (WCNN) framework to reorganize several convolutional layers. Enough convolutional layers are trained using different loss functions so that the WCNN becomes a truly wide network and all subnetworks are trained in parallel, which greatly shortens the training time. Wu et al. [15] proposed a new dual convolutional neural network (DCANet) for blind image denoising, which is the first network to combine dual CNNs and attention mechanisms for image denoising. DCANet can suppress not only singular noise but also mixed noise and real-world noise.
Although the denoising algorithms described above have achieved good performance, most denoising methods fuse different levels of features through long and short skip connections, easily generating a lot of redundant information, thereby weakening the complementarity of different levels of features, resulting in the loss of image details [11]. Therefore, effectively retaining the detail features of an image poses a challenge in denoising tasks. To address these limitations, we propose a new multi-scale detail–noise complementary learning (MDNCL) network for AWGN and real-world noise removal. Our approach involves several steps. First, we employ a low-pass filter to extract a base layer and a detail layer from the input image. The detail layer contains edge and texture features, along with the noise mappings, while the base layer represents the approximate features of the image. The MDNCL network consists of two branches, namely the detail feature learning branch (DLB) and the noise learning branch (NLB). The detail layer serves as the input for both branches to separately learn image detail features and noisy mappings. In the DLB, we utilize a 2D Discrete Cosine Transform (DCT) to transform the detail layer from the spatial domain to the frequency domain. We also introduce the feature subtraction unit (FSU) to obtain the difference detail features between network layers. These difference features are then fused to produce rich multi-scale difference features, enhancing the complementarity between different network layers. Finally, the learned image features are combined with the base layer to reconstruct the final denoised image. We train the MDNCL network using appropriate loss functions to achieve a balance between noise removal and detail preservation. The proposed MDNCL effectively retains more image details and yields favorable denoising performance.
Overall, the contributions of the proposed MDNCL can be summarized as follows:
  • A novel multi-scale detail–noise complementary learning (MDNCL) network is introduced for the removal of AWGN and real-world noise. The MDNCL network incorporates two branches that extract rich detail features, effectively striking a balance between noise reduction and detail retention.
  • Complementary learning is approached from two perspectives. Firstly, the MDNCL network facilitates the complementary learning of detail features and noise distribution, enabling more effective noise removal while preserving important image details. Secondly, the network facilitates the complementary learning of features between different network layers within the detail feature learning branch.
  • The proposed FSU is proven to be an effective component in the MDNCL network. The FSU captures and leverages difference features within and between network layers. By avoiding the redundancy in image features and enhancing the complementarity between different network layers, the FSU further enhances the denoising performance.
The remainder of this paper is structured as follows. Section 2 presents a comprehensive review of related work. Section 3 introduces the proposed MDNCL approach in detail, explaining on its architecture and learning strategies. Section 4 provides an overview of the datasets used in our study, presents the implementation details, and reports the experimental results. Section 5 concludes the study and presents the key findings and contributions of our research.

2. Related Work

2.1. Multi-Scale Differential Feature Extraction

In the field of image denoising, the ability of a denoising algorithm to preserve detailed features from the original image is a crucial aspect to consider. Scale cues play a crucial role in capturing contextual information of images. Multi-scale differential feature extraction is particularly valuable in computer vision and image processing, as it can generate more discriminative and robust features, ultimately improving the performance of the model. Huang et al. [16] proposed the Multi-Scale Feature Subtraction Fusion (MFSF) module and the Feature Depth Supervision (FDS) module, which not only reduce redundant features but also provide additional supervision by changing the scale features in the decoder, thereby improving the training efficiency and performance of the network. Gan et al. [17] achieved a global understanding of image depth through a multi-scale structure, subtracting two feature maps at each level to extract differential features. This approach is similar to U-Net [18], where upsampling and downsampling are used to compute differences and extract multi-scale differences. In this paper, inspired by the above methods, the U-shaped structure and long and short jump connections of the DLB module are borrowed from U-Net++ [19]. The FSU in DLB can obtain different features by subtracting feature maps of different scales to obtain reliable image features and improve image denoising performance.

2.2. Deep CNNs for Image Denoising

With the development of deep learning, CNN-based methods have gradually become the mainstream in the field of image denoising. Tian et al. [20] proposed a denoising convolutional neural network (ADNet) that uses the principle of nonlinear diffusion to decompose noisy images into two parts, namely the noise part and the signal part. The noise part is removed by nonlinear diffusion, and the signal part is extracted by a deep CNN. Zhang et al. [21] proposed a robust deformable denoising CNN. The algorithm solves the problem of offset pixels in noisy image feature mapping by integrating deformation–convolution into the CNN and combines extended convolution with BN and rectified linear unit (ReLU) realized feature interaction, which improves the denoising performance of the model. In order to solve the problem of feature scaling losing high-frequency features and ignoring in-scale features, Shen et al. [22] adopted dynamic convolution to improve the learning of high-frequency and multi-scale features, achieving better denoising performance with lower computational complexity. Deep CNN models are able to learn complex image features and perform well when dealing with highly noisy images.

2.3. Multi-Branch-Based Network for Image Application

Multi-branch networks have demonstrated strong expressive power and the ability to capture nonlinear features in hidden spaces. Researchers have utilized such networks to tackle various computer vision tasks by assigning different branches to learn different contents. For instance, Reyhaneh et al. [23] proposed a three-branch self-supervised and cyclic multivariate function-based denoising network. Each branch in their network learns a specific aspect, such as the predicted clean image, the signal-dependent noise, and the signal-independent noise. Li et al. [24] developed a three-branch, omni-frequency region-adaptive network (ORNet) for denoising. Their network consists of low-frequency, medium-frequency, and high-frequency branches. Each branch enhances the corresponding frequency components to recover texture details. Wang et al. [25] implemented a simple lightweight network (LWNET) for image denoising with a two-branch structure. The decoder in the upper branch learns a quarter of the channel features, whereas the decoder in the lower branch learns three-fourths of the channel features. In our denoising network, we adopt a two-branch architecture. One branch directly learns the detail features from the detail layer, while the other branch learns the noise mappings from the detail layer. This design allows for the effective extraction and preservation of important image details.

3. Proposed Method

Many existing denoising networks struggle to properly capture the detailed textures of an image. Additionally, as the network depth increases, computational requirements and redundant image features tend to grow, leading to weakened complementarity among features in different network layers and a loss of fine-grained details. To address these issues, we propose the MDNCL approach, which consists of a detail learning branch (DLB), a Noise learning branch (NLB), and a Fusion Module (FM). The architecture of our proposed MDNCL approach is depicted in Figure 1. The DLB focuses on learning detail features, whereas the NLB emphasizes noise mappings derived from the detail layer. The FM is responsible for combining and fusing the features learned from both branches. Ultimately, we employ element-wise addition to reconstruct the clean images, incorporating the insights gained from both the detail and noise components.

3.1. Noise Learning Branch

To preserve the high-frequency information in noisy images, we begin by applying a high-pass filter to emphasize the edges and details. The high-pass filter convolves the noisy image with a high-pass filter kernel. A high-pass filter kernel is a matrix with negative center values and positive surrounding values. In a convolution operation, the high-pass filter reduces the low-frequency part of the image to highlight the high-frequency part. This process, known as the detail layer, isolates the high-frequency components. Next, we subtract the noisy image from the detail layer to obtain the base layer, which can be represented as follows:
I D e t a i l = f H F ( I N ) , I B a s e = I N I D e t a i l ,
where I N represents the noisy image, while I D e t a i l and I B a s e signify the detail layer obtained through the high-pass filter and the base layer, respectively. The base layer is calculated as the difference between the noisy image and the detail layer, and f H F denotes the function of the high-pass filter.
The NLB is responsible for learning the noise mappings to obtain the predicted detail layer. The structure of the NLB, as illustrated in Figure 1a, comprises two 1 × 1 convolution layers and six dense blocks (DBs). Each DB consists of a three-layer 3 × 3 convolutional operation followed by a ReLU and a one-layer 3 × 3 convolutional operation followed by BN and a ReLU. A detailed schematic of a DB is depicted in Figure 1b. Collectively, this process can be represented as follows:
N = 6 f D B ( C o n v ( I D e t a i l ) ) , I d = C o n v ( C o n v ( I D e t a i l ) N ) , N = C o n v ( N ) ,
where N R H × W × C and N R H × W × C denote the noise mappings, where C indicates the channel number of 64 and 1 or 3 depending on whether the noisy image is in gray or color, respectively. N is used for the loss calculation of MDNCL in Section 3.4. I d R H × W × C represents the predicted detail features learned from the NLB, where C indicates the channel number of 1 or 3. f D B denotes the function of the DB, and C o n v denotes the function of 1 × 1 convolution.

3.2. Detail Learning Branch

The detail layer obtained from the high-pass filter is typically sparser compared to the noisy image. To facilitate feature extraction, we apply the DCT to further decompose the detail layer and convert it from the spatial domain to the frequency domain. The DCT process can be mathematically represented as follows:
I d ^ = f D C T ( I D e t a i l ) ,
where f D C T represents the function of the DCT. The kernel size for the DCT is set to 8. I ^ d R H × W × C indicates the detail layer in the frequency domain, where C indicates the channel number of 64.
The DLB is designed with a U-shaped structure to directly learn the predicted detail layer. The structure of the DLB, as shown in Figure 2, consists of the following three main components: an encoder, Feature Subtraction Module (FSM), and decoder. The encoder is composed of four DBs, along with four 3 × 3 convolutional operations. To address the potential loss of image detail features during downsampling operations, four DBs are used in the encoder to adjust the channel numbers. The input channels for the four DBs are 64, 128, 256, and 512. After four encoders, the receptive field size of the feature map of the last layer is 31. The corresponding output channels for the four DBs are 128, 256, 512, and 1024. It should be noted that a 1 × 1 convolutional operation is applied to each DB layer to reduce the number of channels to 64 and, consequently, decrease the number of parameters. The encoding process can be represented as follows:
I ^ d i = C o n v ( f D B ( I ^ d ) ) ( i = 1 , 2 , 3 , 4 ) ,
where I ^ d i denotes the detail features learned from the i-th f D B . The number of channels for all detail features is 64.
The FSU emphasizes the difference and complementary information among features and removes redundant features. A series of differential features with different scales and receptive fields is calculated by connecting multiple FSUs horizontally and vertically. For inter-layer, multi-scale features, we connect multiple FSUs in a vertically connected way to capture large-span cross-layer difference features. For intra-layer, multi-scale features, we connect multiple FSUs in a horizontally connected way to naturally achieve multi-scale, intra-layer feature fusion. To capture multi-scale complementary detail features, we employ multiple FSUs that connect different layers in both horizontal and vertical directions to calculate the differential detail features. In this way, it can remove redundant information and enhance feature complementarity at different levels, thereby improving the denoising performance of the network. The structure of an FSU is illustrated in Figure 3. The FSU includes feature maps ( F A and F B ) corresponding to different layers. We extract features at different scales from these feature maps using 1 × 1 , 3 × 3 , and 5 × 5 convolutional operations. By element-wise subtraction, we obtain three differential feature maps, where each map represents the difference between the same convolutional operation applied to F A and F B . Finally, these differential feature maps are fused together using element-wise addition, resulting in a multi-scale differential feature map. The FSU effectively captures complementary detail features between F A and F B , emphasizing their differences in detail and providing rich features for the decoder. The progress of the FSU can be described as follows:
F S U B A = C o n v 1 × 1 ( F A ) C o n v 1 × 1 ( F B ) + C o n v 3 × 3 ( F A ) C o n v 3 × 3 ( F B ) + C o n v 5 × 5 ( F A ) C o n v 5 × 5 ( F B ) ,
where F S U B A represents the multi-scale difference feature map obtained from the A-th and the B-th layers. C o n v 1 × 1 represents the 1 × 1 convolutional operation, while C o n v 3 × 3 and C o n v 5 × 5 represent the 3 × 3 and 5 × 5 convolutional operations, respectively.
To aggregate the multi-scale differential features from different layers, we employ element-wise addition to enhance and fuse the difference feature maps of each layer with the corresponding feature maps from the next layer. This fusion process results in a complementary fusion feature. Subsequently, we design the decoder to decode the complementary fusion feature. The structure of the decoder, as shown in Figure 2 (decoder part), includes three feature decoders (FDs). Each FD consists of a one-layer 3 × 3 convolution, followed by a ReLU activation and another one-layer 3 × 3 convolution with batch normalization (BN) and a ReLU. This decoding process can be summarized as follows:
F D 3 = f F D ( I ^ d 4 + F S U 4 3 ) , F D 2 = f F D ( F D 3 + F S U 3 2 + F S U 4 , 3 2 ) , F D 1 = f F D ( F D 2 + F S U 2 1 + F S U 3 , 2 1 + F S U 4 , 3 , 2 1 ) ,
where f F D denotes the function of the FD. Each FD, denoted as F D i R H × W × C (for i = 1 , 2 , 3 ), represents the decoding result, with C being the number of channels (set to 64).
Based on Equation (6), we can deduce that F D 1 serves as the final output of the DLB. To transform F D 1 from the frequency domain to the spatial domain, we employ inverse Discrete Cosine Transform (DCT). The process of inverse DCT can be represented as follows:
I d = f I D C T ( F D i ) ,
where f I D C T denotes the function of inverse DCT, and we set the kernel size to 8. I d R H × W × C represents the predicted detail features learned from the DLB, where C indicates a channel number of 1 or 3.

3.3. Fusion Module

We obtain I d and I d from the two branches, corresponding to d e t a i l 1 and d e t a i l 2 in Figure 1, respectively. To obtain the final detail feature map, we utilize a fusion operation, which involves a concatenation operation and a 1 × 1 convolutional operation. Subsequently, for the reconstruction of the denoised image, we perform element-wise addition between the base layer and the final detail feature map. The specific process is outlined as follows:
I ˜ d = C o n v ( C o n c a t ( I d , I d ) ) , I ˜ c = I ˜ d + I B a s e ,
where C o n c a t and C o n v indicate the functions of concatenation and convolution operations, respectively. I ˜ d R H × W × C represents the final detail feature map, while I ˜ c R H × W × C represents the denoised image. Here, C indicates the channel number, which can be either 1 or 3.

3.4. Loss Function

To constrain the training of MDNCL, we adopt the mean square error loss, denoted as L 2 [26], as the loss function. The loss (L) can be expressed as follows:
L = L L 2 d e t a i l + μ L L 2 n o i s e ,
where L L 2 d e t a i l represents the mean squared error between the ground-truth detail and the predicted detail learned by the DLB. L L 2 n o i s e denotes the mean squared error between the ground-truth noise and the predicted noise learned by the NLB. μ represents the loss weight associated with L L 2 n o i s e . The specific expressions for L L 2 d e t a i l and L L 2 n o i s e are presented as follows:
I d = I d e t a i l N , L L 2 d e t a i l = E ( I d , I d ) [ 1 N p I d I d 2 ] , L L 2 n o i s e = E ( N , N ) [ 1 N p N N 2 ] ,
where N and I d denote the ground-truth noise mapping and ground-truth detail, respectively. N p represents the number of pixels.

4. Experiments

4.1. Dataset

  • Training datasets: To train MDNCL for AWGN removal, we utilized the Berkeley segmentation dataset (BSD) [27], which consists of 432 original natural images with dimensions of 481 × 321 . In order to enhance the training process and increase the number of training patches, each image was divided into 128 patches with dimensions of 58 × 58 , resulting in a total of 55,296 image patches. Subsequently, random data augmentation techniques were applied to the training patches to enhance the generalization ability of MDNCL. These techniques included vertical flipping, rotation, etc. For denoising in grayscale images, we trained two versions of MDNCL, namely MDNCL, which utilizes known noise levels, and MDNCL-B, which performs denoising with blind noise levels. Similarly, for color images, we conducted experiments using CMDNCL and CMDNCL-B. Known noise levels were set to σ = 15, 25, and 50, where σ represents the standard deviation. The blind noise level was set within the range of σ 0 , 55 .
    To train MDNCL for real-world noise removal, we used the Smartphone Image Denoising DATA (SIDD) sRGB track [28] with 160 scene instances, where each instance includes two pairs of high-resolution images. Each pair contained one noisy image and the corresponding clean image. In all, 320 pairs of images were used to train for real-world denoising.
  • Test datasets: For grayscale testing of MDNCL, we evaluated its denoising performance using two datasets, namely the Berkeley segmentation dataset (BSD68) [29] and Set12 [29]. BSD68 consists of 68 images, while Set12 contains 12 images. These datasets were specifically used for AWGN removal. Regarding color image testing, we employed the CBSD68 [29], McMaster [30], and Kodak24 [31] datasets. CBSD68 consists of 68 images, the McMaster dataset consists of 18 images, and Kodak24 consists of 24 images. To assess real-world denoising performance, we utilized sets from the SIDD validation dataset and the Darmstadt Noise Data (DND) dataset [32]. The DND dataset comprises 50 pairs of images but does not provide ground-truth images. To evaluate the denoising results, we uploaded the denoised images to the official DND website, where we obtained the PSNR [33] and SSIM [34] measurements.

4.2. Implementation Details

MDNCL was implemented using Python 3.7 and Pytorch 1.11.0 on an Intel(R) Core(TM) i7-9700K CPU with an Nvidia GeForce RTX 2080Ti GPU. The parameters for MDNCL are listed as follows: a batch size of 16, 90 epochs, and an initial learning rate of 0.0001 that decreases from 0.0001 to 0.00002 as the epochs progress. Optimization was carried out using the Adam algorithm, with the β 1 and β 2 parameters set to 0.9 and 0.999, respectively. Additionally, the weight ( μ ) in Equation (9) was set to 1. More details can be found in Table 1.

4.3. Comparative Experiments

To assess the denoising performance of MDNCL, we conducted a comparison with several advanced denoising methods. The reference methods used in the comparison were BM3D [7], WNNM [35], DnCNN [9], FFDNet [10], ADNet [20], BRDNet [36], BUIFD [37], CDNet [38], MWDCNN [39], DRFENet [40], CBDNet [41], AirNet [42], HI-GAN [43], and TBSN [44]. Among these methods, BM3D [7] and WNNM [35] are traditional denoising techniques, while the remaining approaches are state-of-the-art denoising methods based on deep learning. For quantitative evaluation, we utilized the PSNR and SSIM metrics to measure the performance of MDNCL relative to the other methods.
During the grayscale testing of MDNCL, we employed the Set12 and BSD68 datasets to evaluate its denoising performance. As illustrated in Table 2 and Table 3, MDNCL achieved the highest average PSNR and SSIM results on Set12 and BSD68 at noise levels of 15, 25, and 50. Specifically, when examining the “Barbara” image from the Set12 dataset (Table 3), MDNCL’s denoising performance ranked second, just behind WNNM. MDNCL also demonstrated impressive denoising performance in blind denoising (MDNCL-B) compared to other state-of-the-art denoising methods. For instance, the PSNR and SSIM results of MDNCL-B on the Set12 dataset surpassed those of both DnCNN and DRFENet, as shown in Table 3. Detailed PSNR results for each image and the average PSNR and SSIM scores for each method on the Set12 dataset at three different noise levels can be found in Table 3.
To assess the denoising performance of MDNCL on noisy color images, we conducted evaluations using the CBSD68, Kodak24, and McMaster datasets. As indicated in Table 4, CMDNCL achieved the highest PSNR results across all three datasets and surpassed those of existing reference methods. When the noise level was as high as 75, CMDNCL still obtained the best denoising index, which proves the superior denoising ability of CMDNCL. Furthermore, in the context of blind denoising on color images, we observed that MDNCL-B also demonstrated impressive denoising performance. It effectively removed noise from color images, further affirming the effectiveness of CMDNCL in various denoising scenarios.
To evaluate the denoising performance of MDNCL in removing real-world noise, we conducted tests using the SIDD validation and DND datasets. As illustrated in Table 5, MDNCL achieved the highest PSNR and SSIM results, indicating its effectiveness in removing real-world noise. This demonstrates that MDNCL is well-suited for real-world denoising applications.
After conducting quantitative analysis of MDNCL’s denoising performance, we selected several popular reference denoising methods, including BM3D, DnCNN, FFDNet, ADNet, BRDNet, and MWDCNN, for a visual qualitative comparison. We amplified and observed the differences among the methods by selecting the same areas of interest in the denoised images. Figure 4 and Figure 5 showcase the grayscale image denoising performance. One image from each of the Set12 and BSD68 datasets was chosen for visual comparison at noise levels of 15 and 25, respectively. In Figure 4, the image “House” exhibits more preserved striped details when denoised by MDNCL. In Figure 5, MDNCL demonstrates the clearest outline of the castle compared to other methods. These visual comparisons indicate that MDNCL achieves the best visual performance and preserves more detailed features compared to the reference methods.
For color image denoising, we selected two images from the CBSD68 dataset at a noise level of 25 and one image from each of the Kodak24 and McMaster datasets at a noise level of 50. In Figure 6, the selected images of zebras with many stripes and a lion with intricate hair textures from CBSD68 demonstrate that CMDNCL excels at handling texture details. Additionally, in Figure 7 and Figure 8, when the noise level is set to 50, CMDNCL effectively preserves the stripe details in the window and the outline of the cherry.
For real-world denoising, several denoised images from the SIDD dataset are shown in Figure 9, demonstrating the effective removal of real-world noise. Based on both quantitative and visual analysis, we conclude that MDNCL exhibits superior visual denoising performance compared to the reference denoising methods.

4.4. Ablation Experiment

  • Role of network framework: To verify the effectiveness of complementary learning branches and the FSU module in MDNCL, we conducted ablation experiments on the Set12 dataset. For complementary learning branches, MDNCL was compared to a corresponding network with only one branch. For the FSU module, the MDNCL was compared to a corresponding network without an FSU. Specifically, we compared MDNCL with MDNCL-NLB, which only learns image details; MDNCL-DLB, which only learns noise; and MDNCL-FSU, which does not have an FSU. To ensure a fair comparison, all networks were trained using the same framework shown in Figure 1. The quantitative results of ablation experiments are shown in Table 6, highlighting the effectiveness of MDNCL.
For visual comparison, Figure 10 shows the results of the ablation experiment. It can be concluded that when a single branch is used to learn detailed features or noise mappings, the denoising results suffer from varying degrees of detail feature loss. However, when the FSU module is removed, the feature artifacts appear in the denoised image, and the noise is not completely removed. Overall, the strong denoising performance of MDNCL is achieved through the joint action of the two branches and the FSU module.
  • Role of loss function: The μ hyperparameter in the loss function was determined by a series of experiments. We conducted experiments on the Set12 dataset, changing the value of the hyperparameter to 0.4, 0.6, 0.8, 1.0, and 1.2. To ensure fair comparisons, all networks were trained using the same framework shown in Figure 1, and quantitative results of the ablation experiments are shown in Table 7.
As can be seen from Table 7, the best denoising performance was obtained by setting the μ hyperparameter to 1.0. When the value of μ is greater than 1.0, the network pays more attention to removing noise and ignores the retention of details. When μ is less than 1, the network is not able to effectively remove noise. Therefore, it is important to strike an balance between detail retention and noise removal.
To more intuitively demonstrate why setting the μ hyperparameter to 1.0 is a reasonable choice, we compared denoised images with different μ values. As can be seen in Figure 11, when μ is less than 1, the zoomed-in detail areas, such as the white feathers, still contain a significant amount of noise. This indicates that the denoising effect is insufficient and smoothing is inadequate. On the other hand, when μ is greater than 1, the noise is reduced, but the smoothing is too strong, leading to a loss of detail information. Through this comparison, we can see that when μ equals 1.0, it achieves a good balance between noise removal and detail preservation.

4.5. Model Complexity Analysis

Running time and the number of parameters are important indicators to evaluate the complexity of a model. In this section, we compare MDNCL with BM3D [7], TWSC [45], MCWNNM [46], IRCNN [47], DnCNN-B [9], FFDNet [10], BUIFD [37], ADNet [20], SADNet [48], CycleISP [49], AirNet [42], VDN [50], AINDNet [51], VDIR [52], RDUNet [12], ATDNet [53], CTNet [54], and TBSN [44] in terms of running time and model parameters. Among these models, BM3D, TWSC, and MCWNNM were implemented in the Matlab (R2020a) environment, and other denoising methods were implemented in PyCharm (2021.2.1) software. We randomly selected images with sizes of 256 × 256 , 512 × 512 , and 1024 × 1024 at a noise level of 25. Table 8 records the average running time of 20 implementations of each tested model, and we ignore the memory transfer time between the CPU and GPU. Table 9 shows the complexity of the comparison models.
The performance of the MDNCL algorithm is highlighted on multiple dimensions, showing its superiority over the compared denoising models. For color images, the running times for each image size are 0.316 s, 0.574 s, and 2.174 s. When processing large-size color images, MDNCL is much faster than models such as AINDNet, BM3D, and MCWNNM; AINDNet took 9.573 s, BM3D took 12.788 s, and MCWNNM took 1120.332 s, indicating the efficiency of MDNCL when processing large-size images.
As can be seen from Table 9, MDNCL has 6.71 M parameters, indicating that it can still achieve excellent noise reduction performance while maintaining low computational complexity. In contrast, the number of parameters of CTNet is as high as 49.03 M. While the denoising performance of CTNet may be better, it can be prohibitive in practical applications due to its huge computing requirements. MDNCL provides a balanced solution that optimizes performance and computational efficiency for environments with limited computing resources.

5. Conclusions

In this paper, we propose a multi-scale detail–noise complementary learning (MDNCL) network for image denoising. The two-branch architecture of MDNCL aims to strike a balance between noise reduction and detail preservation, with particularly noticeable performance gains at higher noise levels. Furthermore, the proposed FSU addresses issues related to redundant information and computational complexity during the denoising process, resulting in more effective preservation of detailed features. To evaluate the denoising capabilities of MDNCL, we conducted experiments on various datasets for both grayscale and color images, as well as ablation experiments on the Set12 dataset. The obtained experimental results consistently reinforce the strong denoising performance of MDNCL. However, the proposed MDNCL also has some limitations. This method is suitable for single noise and real-world noise, but it may not be effective for complex mixed noise and noise in complex environments. We will further investigate combining the idea of generative models with MDNCL to improve denoising performance in future work.

Author Contributions

Conceptualization, Y.C. and J.J.; methodology, J.J. and M.S.; validation, M.S.; formal analysis, M.S. and Y.C.; investigation, Y.C. and J.J.; resources, Y.C.; writing—original draft preparation, M.S. and Y.C.; writing—review and editing, Y.C. and M.S.; visualization, J.J.; funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62001236, in part by the Natural Science Foundation of the Jiangsu Higher Education Institutions of China under Grant 20KJA520003, and in part by the Six Talent Peaks Project of Jiangsu Province under Grant JY-051.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available datasets were analyzed in this study. These data can be found at https://paperswithcode.com/datasets (accessed on 26 December 2023).

Acknowledgments

The authors would like to thank Haidong Yang and Yanqiong Zhang. Haidong Yang contributed to software and project administration, and Yanqiong Zhang contributed to data curation and supervision. Their work improved the quality and clarity of the paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Yang, X.; Yan, J.; Liao, W.; Yang, X.; Tang, J.; He, T. Scrdet++: Detecting small, cluttered and rotated objects via instance-level feature denoising and rotation loss smoothing. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2384–2399. [Google Scholar] [CrossRef] [PubMed]
  2. Yazdi, M.; Bouwmans, T. New trends on moving object detection in video images captured by a moving camera: A survey. Comput. Sci. Rev. 2018, 28, 157–177. [Google Scholar] [CrossRef]
  3. Zatar, W.; Chen, G.; Nghiem, H. Ultrasonic Pulse-Echo Signals for Quantitative Assessment of Reinforced Concrete Anomalies. Appl. Sci. 2024, 14, 4860. [Google Scholar] [CrossRef]
  4. Huang, Z.; Zhang, J.; Zhang, Y.; Shan, H. DU-GAN: Generative Adversarial Networks with Dual-Domain U-Net-Based Discriminators for Low-Dose CT Denoising. IEEE Trans. Instrum. Meas. 2021, 71, 1–12. [Google Scholar] [CrossRef]
  5. Tian, C.; Fei, L.; Zheng, W.; Xu, Y.; Zuo, W.; Lin, C.W. Deep learning on image denoising: An overview. Neural Netw. 2020, 131, 251–275. [Google Scholar] [CrossRef] [PubMed]
  6. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  7. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3-D transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  8. Wu, J.; Lee, X. An improved WNNM algorithm for image denoising. J. Phys. Conf. Ser. 2019, 1237, 022037. [Google Scholar] [CrossRef]
  9. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef]
  10. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef]
  11. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  12. Gurrola-Ramos, J.; Dalmau, O.; Alarcón, T.E. A Residual Dense U-Net Neural Network for Image Denoising. IEEE Access 2021, 9, 31742–31754. [Google Scholar] [CrossRef]
  13. Qu, Q.; Song, Y.; Chen, J. Denoising Multiscale Back-Projection Feature Fusion for Underwater Image Enhancement. Appl. Sci. 2024, 14, 4395. [Google Scholar] [CrossRef]
  14. Liu, G.; Dang, M.; Liu, J. True wide convolutional neural network for image denoising. Inf. Sci. 2022, 610, 171–184. [Google Scholar] [CrossRef]
  15. Wu, W.; Ge, A.; Lv, G. DCANet: Dual Convolutional Neural Network with Attention for Image Blind Denoising. arXiv 2023, arXiv:2304.01498. [Google Scholar]
  16. Huang, Z.; You, H. MFSFNet: Multi-scale feature subtraction fusion network for remote sensing image change detection. Remote Sens. 2023, 15, 3740. [Google Scholar] [CrossRef]
  17. Gan, Y.; Xu, X.; Sun, W.; Lin, L. Monocular depth estimation with affinity, vertical pooling, and label enhancement. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 224–239. [Google Scholar]
  18. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCA, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  19. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar]
  20. Tian, C.; Xu, Y.; Li, Z.; Zuo, W.; Fei, L.; Liu, H. Attention-guided CNN for image denoising. Neural Netw. 2020, 124, 117–129. [Google Scholar] [CrossRef]
  21. Zhang, Q.; Xiao, J.; Tian, C.; Chun-Wei Lin, J. A Robust Deformed Convolutional Neural Network (CNN) for Image Denoising. CAAI Trans. Intell. Technol. 2023, 8, 331–342. [Google Scholar] [CrossRef]
  22. Shen, H.; Zhao, Z.Q.; Zhang, W. Adaptive dynamic filtering network for image denoising. In Proceedings of the AAAI Conference on Artificial Intelligence, Seattle, WA, USA, 7–14 February 2023; Volume 37, pp. 2227–2235. [Google Scholar]
  23. Neshatavar, R.; Liu, P.; Pang, J.; Mei, T.; Barnes, N.; Petersson, L.; Harandi, M. CVF-SID: Cyclic multi-variate function for self-supervised image denoising by disentangling noise from image. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17583–17591. [Google Scholar]
  24. Li, X.; Zhang, K.; Zuo, W.; Zhang, L.; Zhang, L.; Zhu, X.; Zhang, Y.; Zhang, L. Learning omni-frequency region-adaptive representations for real image super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence; Association for the Advancement of Artificial Intelligence: Washington, DC, USA, 2021; Volume 35, pp. 1975–1983. [Google Scholar]
  25. Wang, J.; Lu, Y.; Lu, G. Lightweight image denoising network with four-channel interaction transform. Image Vis. Comput. 2023, 137, 104766. [Google Scholar] [CrossRef]
  26. Allen, D.M. Mean square error of prediction as a criterion for selecting variables. Technometrics 1971, 13, 469–475. [Google Scholar] [CrossRef]
  27. Li, H.; Cai, J.; Nguyen, T.N.A.; Zheng, J. A benchmark for semantic image segmentation. In Proceedings of the IEEE International Conference on Multimedia and Expo, San Jose, CA, USA, 15–19 July 2013; pp. 1–6. [Google Scholar]
  28. Abdelhamed, A.; Lin, S.; Brown, M.S. A high-quality denoising dataset for smartphone cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1692–1700. [Google Scholar]
  29. Li, H.; Xia, S.; Zhou, B.; Peng, J. The growth mechanism of grain boundary carbide in Alloy 690. Mater. Charact. 2013, 81, 1–6. [Google Scholar] [CrossRef]
  30. Zhang, L.; Wu, X.; Buades, A.; Li, X. Color demosaicking by local directional interpolation and nonlocal adaptive thresholding. J. Electron. Imaging 2011, 20, 023016. [Google Scholar]
  31. Franzen, R. Kodak lossless true color image suite. Photocd pcd0992 1999, 4, 2. [Google Scholar]
  32. Plotz, T.; Roth, S. Benchmarking denoising algorithms with real photographs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2750–2759. [Google Scholar]
  33. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the international conference on pattern recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  34. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  35. Gu, S.; Zhang, L.; Zuo, W.; Feng, X. Weighted nuclear norm minimization with application to image denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 2862–2869. [Google Scholar]
  36. Tian, C.; Xu, Y.; Zuo, W. Image denoising using deep CNN with batch renormalization. Neural Netw. 2020, 121, 461–473. [Google Scholar] [CrossRef] [PubMed]
  37. El Helou, M.; Süsstrunk, S. Blind Universal Bayesian Image Denoising with Gaussian Noise Level Learning. IEEE Trans. Image Process. 2020, 29, 4885–4897. [Google Scholar] [CrossRef] [PubMed]
  38. Quan, Y.; Chen, Y.; Shao, Y.; Teng, H.; Xu, Y.; Ji, H. Image denoising using complex-valued deep CNN. Pattern Recognit. 2021, 111, 107639. [Google Scholar] [CrossRef]
  39. Tian, C.; Zheng, M.; Zuo, W.; Zhang, B.; Zhang, Y.; Zhang, Y. Multi-stage image denoising with the wavelet transform. Pattern Recognit. 2023, 134, 109050. [Google Scholar] [CrossRef]
  40. Zhong, R.; Zhang, Q. DRFENet: An Improved Deep Learning Neural Network via Dilated Skip Convolution for Image Denoising Application. Appl. Sci. 2022, 13, 28. [Google Scholar] [CrossRef]
  41. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 1712–1722. [Google Scholar]
  42. Li, B.; Liu, X.; Hu, P.; Wu, Z.; Lv, J.; Peng, X. All-in-one image restoration for unknown corruption. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 17452–17462. [Google Scholar]
  43. Vo, D.M.; Nguyen, D.M.; Le, T.P.; Lee, S.W. HI-GAN: A hierarchical generative adversarial network for blind denoising of real photographs. Inf. Sci. 2021, 570, 225–240. [Google Scholar] [CrossRef]
  44. Li, J.; Zhang, Z.; Zuo, W. TBSN: Transformer-Based Blind-Spot Network for Self-Supervised Image Denoising. arXiv 2024, arXiv:2404.07846. [Google Scholar]
  45. Wu, W.; Ge, A.; Lv, G.; Xia, Y.; Zhang, Y. Two-stage Progressive Residual Dense Attention Network for Image Denoising. arXiv 2024, arXiv:2401.02831. [Google Scholar]
  46. Xu, J.; Zhang, L.; Zhang, D.; Feng, X. Multi-channel weighted nuclear norm minimization for real color image denoising. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1096–1104. [Google Scholar]
  47. Zhang, K.; Zuo, W.; Gu, S.; Zhang, L. Learning deep CNN denoiser prior for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3929–3938. [Google Scholar]
  48. Chang, M.; Li, Q.; Feng, H.; Xu, Z. Spatial-adaptive network for single image denoising. In Proceedings of the Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; pp. 171–187. [Google Scholar]
  49. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Cycleisp: Real image restoration via improved data synthesis. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 2696–2705. [Google Scholar]
  50. Yue, Z.; Yong, H.; Zhao, Q.; Meng, D.; Zhang, L. Variational denoising network: Toward blind noise modeling and removal. In Proceedings of theAdvances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Volume 32. [Google Scholar]
  51. Kim, Y.; Soh, J.W.; Park, G.Y.; Cho, N.I. Transfer learning from synthetic to real-noise denoising with adaptive instance normalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 3482–3492. [Google Scholar]
  52. Soh, J.W.; Cho, N.I. Variational deep image restoration. IEEE Trans. Image Process. 2022, 31, 4363–4376. [Google Scholar] [CrossRef] [PubMed]
  53. Kim, Y.; Soh, J.W.; Cho, N.I. Adaptively tuning a convolutional neural network by gate process for image denoising. IEEE Access 2019, 7, 63447–63456. [Google Scholar] [CrossRef]
  54. Tian, C.; Zheng, M.; Zuo, W.; Zhang, S.; Zhang, Y.; Lin, C.W. A cross Transformer for image denoising. Inf. Fusion 2024, 102, 102043. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the proposed method. (a) Network architecture of MDNCL. (b) Dense block (DB).
Figure 1. The overall framework of the proposed method. (a) Network architecture of MDNCL. (b) Dense block (DB).
Applsci 14 07044 g001
Figure 2. Schematic diagram of DLB.
Figure 2. Schematic diagram of DLB.
Applsci 14 07044 g002
Figure 3. Schematic diagram of FSU.
Figure 3. Schematic diagram of FSU.
Applsci 14 07044 g003
Figure 4. Denoising results of “House” from Set12 with a noise level of 15. (a) clean image; (b) noisy image; (c) BM3D/34.93 dB; (d) DnCNN/34.99 dB; (e) FFDNet/35.07 dB; (f) ADNet/35.22 dB; (g) BRDNet/35.27 dB; (h) MDNCL (Ours)/35.46 dB.
Figure 4. Denoising results of “House” from Set12 with a noise level of 15. (a) clean image; (b) noisy image; (c) BM3D/34.93 dB; (d) DnCNN/34.99 dB; (e) FFDNet/35.07 dB; (f) ADNet/35.22 dB; (g) BRDNet/35.27 dB; (h) MDNCL (Ours)/35.46 dB.
Applsci 14 07044 g004
Figure 5. Denoising results of images from BSD68 with a noise of level 25. (a) clean image; (b) noisy image; (c) BM3D/29.53 dB; (d) DnCNN/30.17 dB; (e) FFDNet/30.02 dB; (f) ADNet/30.24 dB; (g) BRDNet/30.27 dB; (h) MWDCNN/30.14 dB; and (i) MDNCL (ours)/30.39 dB.
Figure 5. Denoising results of images from BSD68 with a noise of level 25. (a) clean image; (b) noisy image; (c) BM3D/29.53 dB; (d) DnCNN/30.17 dB; (e) FFDNet/30.02 dB; (f) ADNet/30.24 dB; (g) BRDNet/30.27 dB; (h) MWDCNN/30.14 dB; and (i) MDNCL (ours)/30.39 dB.
Applsci 14 07044 g005
Figure 6. Denoising results of two images from CBSD68 with a noise level of 25. (a) clean image; (b) noisy image; (c) FFDNet/30.54 dB; (d) ADNet/30.74 dB; (e) MDNCL (Ours)/31.10 dB; (f) clean image; (g) noisy image; (h) FFDNet/29.85 dB; (i) ADNet/29.92 dB; (j) MDNCL (Ours)/30.07 dB.
Figure 6. Denoising results of two images from CBSD68 with a noise level of 25. (a) clean image; (b) noisy image; (c) FFDNet/30.54 dB; (d) ADNet/30.74 dB; (e) MDNCL (Ours)/31.10 dB; (f) clean image; (g) noisy image; (h) FFDNet/29.85 dB; (i) ADNet/29.92 dB; (j) MDNCL (Ours)/30.07 dB.
Applsci 14 07044 g006
Figure 7. Denoising results of one image from Kodak24 with a noise level of 50. (a) clean image; (b) noisy image; (c) DnCNN/26.13 dB; (d) ADNet/26.22 dB; (e) FFDNet/26.23 dB; (f) CMDNCL (Ours)/26.48 dB.
Figure 7. Denoising results of one image from Kodak24 with a noise level of 50. (a) clean image; (b) noisy image; (c) DnCNN/26.13 dB; (d) ADNet/26.22 dB; (e) FFDNet/26.23 dB; (f) CMDNCL (Ours)/26.48 dB.
Applsci 14 07044 g007
Figure 8. Denoising results of one image from the McMaster dataset with a noise level of 50. (a) clean image; (b) noisy image; (c) DnCNN/31.13 dB; (d) ADNet/31.84 dB; (e) FFDNet/31.82 dB; (f) CMDNCL (Ours)/32.08 dB.
Figure 8. Denoising results of one image from the McMaster dataset with a noise level of 50. (a) clean image; (b) noisy image; (c) DnCNN/31.13 dB; (d) ADNet/31.84 dB; (e) FFDNet/31.82 dB; (f) CMDNCL (Ours)/32.08 dB.
Applsci 14 07044 g008
Figure 9. Denoising results on SIDD. The noisy images are arranged at the top, and the denoised images are presented at the bottom.
Figure 9. Denoising results on SIDD. The noisy images are arranged at the top, and the denoised images are presented at the bottom.
Applsci 14 07044 g009
Figure 10. Denoising results of one image from Set12 with a noise level of 15. (a) clean image; (b) noisy image; (c) MDNCL-NLB/30.19 dB; (d) MDNCL-DLB/31.27 dB; (e) MDNCL-FSU/28.79 dB; (f) MDNCL/32.08 dB.
Figure 10. Denoising results of one image from Set12 with a noise level of 15. (a) clean image; (b) noisy image; (c) MDNCL-NLB/30.19 dB; (d) MDNCL-DLB/31.27 dB; (e) MDNCL-FSU/28.79 dB; (f) MDNCL/32.08 dB.
Applsci 14 07044 g010
Figure 11. Image denoising result with a noise level of 25 on Set12 when μ takes different values. (a) clean image; (b) noisy image; (c) μ = 0.6; (d) μ = 0.8; (e) μ = 1.2; (f) μ = 1.0.
Figure 11. Image denoising result with a noise level of 25 on Set12 when μ takes different values. (a) clean image; (b) noisy image; (c) μ = 0.6; (d) μ = 0.8; (e) μ = 1.2; (f) μ = 1.0.
Applsci 14 07044 g011
Table 1. Experimental parameters.
Table 1. Experimental parameters.
ParameterValue
Batch size16
Epochs90
Initial learning rate0.0001
β 1 0.9
β 2 0.999
μ 1
Xavier initializationDraw weights from a normal distribution with a mean of 0
Table 2. Average PSNR (dB)/SSIM results of different methods on BSD68 (The superior results are emphasized in bold).
Table 2. Average PSNR (dB)/SSIM results of different methods on BSD68 (The superior results are emphasized in bold).
Method σ = 15 σ = 25 σ = 50
BM3D [7]31.07/0.872928.57/0.804225.62/0.6817
WNNM [35]31.37/0.877728.83/0.810325.87/0.6975
DnCNN [9]31.72/0.890229.23/0.828826.23/0.7180
FFDNet [10]31.63/0.890129.19/0.829826.29/0.7262
ADNet [20]31.74/0.891029.25/0.828526.29/0.7211
CDNet [38]31.74/0.891529.28/0.831226.36/0.7229
BRDNet [36]31.72/0.891029.24/0.829526.28/0.7232
DRFENet [40]31.76/0.892029.26/0.830126.29/0.7230
MWDCNN [39]31.77/0.891929.28/0.830526.29/0.7229
MDNCL (ours)31.82/0.892329.30/0.831626.38/0.7242
MDNCL-B (ours)31.76/0.891929.25/0.830426.31/0.7234
Table 3. Average PSNR (dB)/SSIM results of different methods for Set12 (The superior results are emphasized in bold).
Table 3. Average PSNR (dB)/SSIM results of different methods for Set12 (The superior results are emphasized in bold).
ImageC.manHousePeppersStarfishMonarchAirplaneParrotLenaBarbaraBoatManCoupleAverage
Noise Level σ = 15
BM3D [7]31.9134.9332.6931.1431.8531.0731.3734.2633.1032.1331.9232.1032.37/0.8953
WNNM [35]32.1735.1332.9931.8232.7131.3931.6234.2733.6032.2732.1132.1732.70/0.8944
DnCNN [9]32.5934.9933.2432.1333.2531.6731.8834.5832.6132.4232.4332.4332.85/0.9029
FFDNet [10]32.4335.0733.2431.9932.6631.5731.8734.6232.5432.3832.4132.4632.77/0.9027
CDNet [38]32.6434.9433.2832.1733.3031.7231.9334.5232.6632.3632.4932.4732.87/0.9034
ADNet [20]32.8135.2233.4932.1733.1731.8631.9634.7132.8032.5732.4732.5832.98/0.9039
BRDNet [36]32.8035.2733.4732.2433.3531.8232.0034.7532.9332.5532.5032.6233.03/0.9038
DRFENet [40]32.6035.1933.4232.1933.4131.7631.9634.7232.7332.5432.4532.5532.96/0.9045
MWDCNN [39]32.5335.0933.2932.2833.2031.7431.9734.6432.6532.4932.4632.5232.91/0.9043
MDNCL (ours)32.9135.4633.5332.3133.6831.9432.0834.8133.1532.6232.5632.6833.14/0.9065
MDNCL-B (ours)32.7935.3433.4632.1833.2931.8432.0134.7332.7732.5232.5032.6333.01/0.9056
Noise Level σ = 25
BM3D [7]29.4532.8530.1628.5629.2528.4228.9332.0730.7129.9029.6129.7129.97/0.8499
WNNM [35]29.6433.2230.4229.0329.8428.6929.1532.2431.2430.0329.7629.8230.26/0.8451
DnCNN [9]30.1833.0630.8729.4130.2829.1329.4332.4430.0030.2130.1030.1230.43/0.8608
FFDNet [10]30.1033.2830.9329.3230.0829.0429.4432.5730.0130.2530.1130.2030.44/0.8638
CDNet [38]30.2132.9830.7829.5130.3529.2029.5332.4030.0530.1530.0830.1130.53/0.8646
ADNet [20]30.3433.4131.1429.4130.3929.1729.4932.6130.2530.3730.0830.2430.58/0.8637
BRDNet [36]31.3933.4131.0429.4630.5029.2029.5532.6530.3430.3330.1430.2830.61/0.8639
DRFENet [40]30.2633.4131.0729.4929.4929.1229.4632.5730.0230.2930.0730.1730.54/0.8642
MWDCNN [39]30.1933.3330.8529.6630.5529.1629.4832.6730.2130.2830.1030.1330.55/0.8645
MDNCL (ours)30.5233.6631.1829.6930.7529.3429.6832.8430.8130.4830.2430.4430.80/0.8679
MDNCL-B (ours)30.4633.5531.1129.4330.5229.2329.5932.6930.3430.3530.1630.3230.65/0.8673
Noise Level σ = 50
BM3D [7]26.1329.6926.6825.0425.8225.1025.9029.0527.2226.7826.8126.4626.72/0.7654
WNNM [35]26.4530.3326.9525.4426.3225.4226.1429.2527.7926.9726.9426.6427.05/0.7568
DnCNN [9]27.0330.0027.3225.7026.7825.8726.4829.3926.2227.2027.2426.9027.18/0.7829
FFDNet [10]27.0530.3727.5425.7526.8125.8926.5729.6626.4527.3327.2927.0827.32/0.7916
CDNet [38]27.0230.4827.4025.9327.1225.8526.5829.7026.5027.1527.3727.0527.38/0.7924
ADNet [20]27.3130.5927.6925.7026.9025.8826.5629.5926.6427.3527.1727.0727.37/0.7875
BRDNet [36]27.4430.5327.6725.7726.9725.9326.6629.7326.8527.3827.2727.1727.45/0.7898
DRFENet [40]27.1030.5727.5425.8326.8925.8326.4229.6026.5327.3127.2727.0627.33/0.7894
MWDCNN [39]26.9930.5827.3425.8527.0225.9326.4829.6326.6027.2327.2727.1127.34/0.7897
MDNCL (ours)27.6430.9627.7926.0027.1526.0826.8029.9527.4027.5527.3927.3727.67/0.7955
MDNCL-B (ours)27.5130.5327.6925.7426.9225.9126.6729.6726.7927.3627.2827.1427.43/0.7923
Table 4. Average PSNR (dB) results of different methods on the CBSD68, Kodak24, and McMaster datasets (The superior results are emphasized in bold).
Table 4. Average PSNR (dB) results of different methods on the CBSD68, Kodak24, and McMaster datasets (The superior results are emphasized in bold).
DatasetsMethod σ = 15 σ = 25 σ = 50 σ = 75
CBSD68CBM3D [7]33.52/0.92530.71/0.87227.38/0.76725.74/-
CDnCNN-B [9]33.95/0.92931.29/0.88328.01/0.790-
FFDNet [10]33.86/0.92931.18/0.88227.95/0.78926.24/0.593
ADNet [20]34.02/0.93331.34/0.88928.05/0.79726.33/0.606
BUIFD [37]33.65/0.93030.76/0.88226.61/0.777-
AirNet [42]33.92/0.93331.26/0.88828.01/0.798-
CMDNCL (ours)34.21/0.93931.53/0.89228.26/0.80226.49/0.613
CMDNCL-B (ours)34.12/0.93431.46/0.88928.22/0.80226.41/0.609
Kodak24CBM3D [7]34.28/0.91631.68/0.86828.46/0.77526.82/-
CDnCNN-B [9]34.73/0.92032.23/0.87629.02/0.791-
FFDNet [10]34.55/0.92232.11/0.87828.99/0.79427.25/0.733
ADNet [20]34.76/0.92432.26/0.88229.10/0.79827.40/0.739
BUIFD [37]34.41/0.92331.77/0.87927.74/0.786-
AirNet [42]34.68/0.92432.21/0.88229.06/0.799-
MWDCNN [39]34.91/0.92732.40/0.88629.26/0.80627.55/0.749
CMDNCL (ours)35.09/0.93432.57/0.88929.46/0.81328.61/0.759
CMDNCL-B (ours)35.01/0.93032.49/0.88529.37/0.80928.02/0.751
McMasterCBM3D [7]34.06/0.91531.66/0.87428.51/0.793-
CDnCNN-B [9]34.80/0.90432.47/0.86929.21/0.799-
FFDNet [10]34.47/0.92232.25/0.88629.14/0.815-
ADNet [20]34.93/0.92732.56/0.89429.36/0.825-
BUIFD [37]33.84/0.90131.06/0.84726.20/0.733-
AirNet [42]34.70/0.92532.44/0.89129.26/0.822-
CMDNCL (ours)35.19/0.93932.92/0.91329.69/0.839-
CMDNCL-B (ours)35.13/0.93032.86/0.90729.59/0.833-
Table 5. Denoising results of different methods on real-world noise datasets (The superior results are emphasized in bold).
Table 5. Denoising results of different methods on real-world noise datasets (The superior results are emphasized in bold).
Test DataMethodBM3D [7]WNNM [35]CBDNet [41]HI-GAN [43]TBSN [44]MDNCL
SIDDPSNR25.6525.7838.6838.4737.7839.31
SSIM0.6850.6850.8090.9000.9400.928
DNDPSNR34.5134.6738.0639.3239.0839.37
SSIM0.8510.8650.9420.9520.9450.959
Table 6. Average PSNR (dB)/SSIM results of ablation experiments on Set12 (The superior results are emphasized in bold).
Table 6. Average PSNR (dB)/SSIM results of ablation experiments on Set12 (The superior results are emphasized in bold).
Method σ = 15 σ = 25 σ = 50
MDNCL-DLB32.46/0.891330.19/0.858927.17/0.7821
MDNCL-NLB31.93/0.879429.82/0.849726.89/0.7690
MDNCL-FSU28.89/0.862329.48/0.844624.11/0.7413
MDNCL33.14/0.906530.80/0.867927.67/0.7955
Table 7. Average PSNR (dB)/SSIM results of ablation experiments on Set12 (The superior results are emphasized in bold).
Table 7. Average PSNR (dB)/SSIM results of ablation experiments on Set12 (The superior results are emphasized in bold).
μ σ = 15 σ = 25 σ = 50
0.431.90/0.841929.56/0.844926.47/0.7553
0.632.77/0.881429.46/0.857727.45/0.7690
0.832.57/0.876329.80/0.859327.54/0.7723
1.033.14/0.906530.80/0.867927.67/0.7955
1.231.78/0.880129.40/0.847026.45/0.7529
Table 8. Running time (in seconds) of the compared models on three sizes of grayscale and color noisy images.
Table 8. Running time (in seconds) of the compared models on three sizes of grayscale and color noisy images.
DeviceModel256 × 256512 × 5121024 × 1024
GrayscaleColorGrayscaleColorGrayscaleColor
BM3D [7]0.4460.5892.2743.6619.68712.788
CPUTWSC [45]12.29434.36153.011140.922220.997608.477
MCWNNM [46]-62.732-278.073-1120.332
IRCNN [47]0.0300.0300.0300.0300.0300.030
DnCNN-B [9]0.0320.0320.0370.0370.0570.057
FFDNet [10]0.0310.0300.0310.0300.0320.030
BUIFD [37]0.0350.0370.0500.0530.1120.123
ADNet [20]0.0310.0330.0350.0450.0510.093
GPUSADNet [48]0.0300.0300.0430.0440.1010.102
CycleISP [49]-0.055-0.156-0.545
AirNet [42]-0.143-0.498-2.501
VDN [50]0.1440.1620.6070.5972.3672.376
AINDNet [51]-0.531-2.329-9.573
VDIR [52]-0.385-1.622-6.690
MDNCL0.2790.3160.5740.6972.0132.174
Table 9. Complexity of different denoising methods.
Table 9. Complexity of different denoising methods.
MethodParametersFLOPs
DnCNN [9]0.56 M0.891 G
ADNet [20]0.52 M0.832 G
ATDNet [53]9.45 M-
RDUNet [12]166 M48 G
CTNet [54]49.03 M6.91 G
TBSN [44]12.97 M607.08 G
AirNet [42]8.72 M-
MDNCL6.71 M9.34 G
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cui, Y.; Shi, M.; Jiang, J. Multi-Scale Detail–Noise Complementary Learning for Image Denoising. Appl. Sci. 2024, 14, 7044. https://doi.org/10.3390/app14167044

AMA Style

Cui Y, Shi M, Jiang J. Multi-Scale Detail–Noise Complementary Learning for Image Denoising. Applied Sciences. 2024; 14(16):7044. https://doi.org/10.3390/app14167044

Chicago/Turabian Style

Cui, Yan, Mingyue Shi, and Jielin Jiang. 2024. "Multi-Scale Detail–Noise Complementary Learning for Image Denoising" Applied Sciences 14, no. 16: 7044. https://doi.org/10.3390/app14167044

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop