Next Article in Journal
Cautious Bayesian Optimization: A Line Tracker Case Study
Previous Article in Journal
Unmanned Ariel Vehicle (UAV) Path Planning for Area Segmentation in Intelligent Landmine Detection Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved BM3D Algorithm Based on Image Depth Feature Map and Structural Similarity Block-Matching

1
College of Big Data and Intelligent Engineering, Southwest Forestry University, Kunming 650224, China
2
Information Security College, Yunnan Police College, Kunming 650221, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7265; https://doi.org/10.3390/s23167265
Submission received: 19 July 2023 / Revised: 16 August 2023 / Accepted: 17 August 2023 / Published: 18 August 2023
(This article belongs to the Section Sensing and Imaging)

Abstract

:
We propose an improved BM3D algorithm for block-matching based on UNet denoising network feature maps and structural similarity (SSIM). In response to the traditional BM3D algorithm that directly performs block-matching on a noisy image, without considering the deep-level features of the image, we propose a method that performs block-matching on the feature maps of the noisy image. In this method, we perform block-matching on multiple depth feature maps of a noisy image, and then determine the positions of the corresponding similar blocks in the noisy image based on the block-matching results, to obtain the set of similar blocks that take into account the deep-level features of the noisy image. In addition, we improve the similarity measure criterion for block-matching based on the Structural Similarity Index, which takes into account the pixel-by-pixel value differences in the image blocks while fully considering the structure, brightness, and contrast information of the image blocks. To verify the effectiveness of the proposed method, we conduct extensive comparative experiments. The experimental results demonstrate that the proposed method not only effectively enhances the denoising performance of the image, but also preserves the detailed features of the image and improves the visual quality of the denoised image.

1. Introduction

In the process of image generation and transformation, noise is inevitably generated, which not only reduces the visual quality of the image but also affects subsequent advanced visual processing tasks [1]. Therefore, it is necessary to take measures to eliminate noise while preserving image details.
At present, image-processing algorithms can be roughly divided into three categories: spatial-domain denoising, frequency-domain denoising [2,3], and deep-learning denoising. Spatial-domain denoising involves directly averaging pixels on noisy images, such as Gaussian filtering [4], median filtering [5], and anisotropic filtering [6]. This type of method can effectively preserve the edge information of the image, but it is also prone to blurring the image. Frequency-domain denoising is the process of converting an image from the spatial domain to the transform domain. In the transform domain, the image is divided into two parts: low-frequency and high-frequency, and noise is more present in the high-frequency components of the image. Therefore, the purpose of denoising can be achieved by processing the high-frequency components through threshold methods [7]. In 2005, based on the nonlocal self-similarity of images, Buades, Coll, and Morel [8] proposed a spatial-domain denoising algorithm called NLM, which improved the performance of the spatial-domain denoising algorithm to a new height. To further improve the performance of the denoising algorithm, in 2007, Dabov, Foi et al. [9] proposed the BLOCK-MATCHING and 3D filtering (BM3D) algorithm that has obvious denoising effects and can effectively retain image details, combining the characteristics of nonlocal self-similarity and frequency-domain denoising. The BM3D algorithm can be simply summarized into three steps [10]: the first step is block-matching grouping, which involves dividing the noisy image into blocks. Based on certain similarity measurement criteria and set threshold, search for similar blocks of each block and combine them to form a 3D matrix cluster; The second step is to use a filtering method to filter each cluster; The third step is aggregation, which involves aggregating the filtered clusters to obtain the output image. Ref. [11] proposed quantum mechanics-based signal and image representation (QAB), which processes high and low values in signals or images by thresholding coefficients. This method has interesting potential in denoising signals and images and is very suitable for Poisson and speckle noise. Inspired by quantum multibody theory, ref. [12] proposed an image-denoising algorithm called De-QuIP. Based on patch analysis, a term similar to quantum mechanics is used to formalize the similarity measure in the local image neighborhood, effectively preserving the local structure of the real image. With the development of artificial intelligence technology, deep-learning-based denoising methods have attracted increased attention. Ref. [13] introduced a nonlocal self-similarity before network processing flow and proposed a blind-spot-based self-supervised image-denoising method. This method can significantly improve denoising performance. Ref. [14] proposed a joint patch-group-based sparse representation (JPG-SR), which can effectively integrate the local sparsity and nonlocal self-similarity of images. The objective and perceptual quality obtained by applying this method to image restoration problems is superior to many state-of-the-art methods. Ref. [15] proposed a group sparsity residual constraint with nonlocal priors (GSRC-NLP) for image restoration to address the problem of traditional GSR models being unable to faithfully estimate image sparsity. This algorithm has achieved excellent results in image denoising, image restoration, and other aspects. Ref. [16] proposed a convolutional neural network called BM3D-Net, which unrolls the BM3D algorithm into a convolutional neural network structure, and “extraction” and “aggregation” layers are utilized to model the block-matching process of traditional BM3D. The denoising effect achieved is significantly better than that of the BM3D method. Ref. [17] proposed a deep neural network called DIVA based on the theory of quantum many-body physics and designed a baseline adaptive denoising algorithm (De-QuIP). This algorithm can handle nonlocal image structures, significantly reducing training costs while improving the visual quality of denoising. Ref. [18] proposed a deep denoising network embedded in quantum interactions, incorporating nonlocal structures into the proposed CNN architecture. The key feature integration in quantum theory enables this method to effectively handle image-denoising problems. Ref. [19] proposed a nonlocal denoising convolutional neural network (NL-DeCNN) network for denoising vibration signals of rolling bearings, and nonlocal mean (NLM) was introduced to construct nonlocal blocks (NLB), which improved the denoising performance of CNN.
According to the BM3D step, block-matching is a critical procedure of the algorithm. Currently, there has been a lot of research on block-matching in the academic community. Refs. [20,21,22] have proposed a new block similarity measure to alleviate the interference of noise and ensure the accuracy of block-matching in images with high-intensity noise. Refs. [23,24] proposed an improved BM3D technology based on an adaptive threshold, which has excellent denoising performance for different types of noise in different images. In [25], to achieve more accurate block-matching, each image is divided into three different regions using the region energy of the alternating current (AC) coefficients of discrete cosine transform (DCT), and the block size is adaptively selected based on the region. It can be seen that the current block-matching improvement algorithms have achieved certain results from the perspective of improving similarity measurement criteria and improving block-matching accuracy. However, the block-matching operations involved in the above studies often directly process noisy images, lacking consideration for deep-level features of the image, such as structure, brightness, contrast, etc. This leads to the neglect of certain important features of the image, resulting in inaccurate block-matching, as a result, the visual quality of the denoised image may not necessarily be good.
In recent years, the rapid development of deep learning has provided ideas for the extraction of deep features in images. Deep learning has the advantages of autonomous learning of target features and a high recognition rate [26,27]. The RGB-D saliency object detection model established in [28] utilized ResNet-50 [29] and VGG-16 [30] to process RGB images and their corresponding depth maps, which can effectively consider the complementarity of multi-level features and generate high-quality predictive saliency maps. Ref. [31] proposed the fusion of spatial and temporal feature representations of speech emotion by parallelizing a Transformer encoder for speech emotion recognition and convolutional neural networks (CNNs). This method not only significantly improved the results of emotion recognition but also reduced computing costs. Ref. [32] proposed a multi-attention-based UNet (MA-UNet) [33] to improve the extraction ability of fine-grained features, which was applied to the semantic segmentation of remote sensing images and achieved good results. Ref. [34] and others utilized feature fusion technology to construct a pyramid structure feature map with equally rich positional and semantic information, and performed object detection on multi-scale feature layers, enhancing the network’s adaptability to scale images. It can be seen that deep learning can learn the deep-level features of images, and using these features to process other advanced visual tasks can achieve more accurate and effective results. However, there is currently little research on using deep-learning feature maps for block-matching. Therefore, inspired by the above research, we adopt a pre-trained UNet denoising network [35] to extract feature maps of different scales from noisy images. Unlike the traditional BM3D algorithm for block-matching on the noisy image, the improved BM3D algorithm proposed in this paper first performs block-matching on the feature maps. Based on the results of block-matching on each feature map, all corresponding similar blocks are searched for at the same location in the noisy image, thus obtaining a set of block-matching that considers different scale features of the noisy image.
In addition, the similarity measurement for block-matching in the traditional BM3D algorithm is based on pixels to calculate the similarity between two images, which is too simple and direct to consider the depth features of the noisy image. The SSIM [36] is an index utilized to measure the similarity between two images. It fully considers the structure, brightness, and contrast information of the image itself, which is more in line with human visual perception. Therefore, to enhance the visual quality of denoised images, we utilize SSIM to improve the traditional similarity measurement and propose a new similarity measurement criterion that considers the structural similarity of noisy images. The contributions of this paper are three-fold:
(1) A pre-trained UNet denoising network is utilized to obtain feature maps at different scales of the noisy image, and then block-matching operations of the BM3D algorithm are performed on the feature maps. The block-matching results are mapped to the corresponding positions of the noisy image, therefore obtaining block-matching groups that take into account the depth characteristics of the noisy image.
(2) Considering that SSIM has the characteristics of image structure, brightness, and contrast, an improved similarity measurement criterion based on SSIM is proposed.
(3) Based on classic image data, extensive comparative experiments and depth analysis are conducted to explain the value of considering image depth features for block-matching operations proposed in this paper.
The other parts of this paper are organized as follows. The Section 2 reviews the basic principles of the traditional BM3D algorithm, while the Section 3 introduces the methods proposed in this paper, including the modified block-matching based on noisy image depth features and the improved similarity measurement based on SSIM. The Section 4 conducts extensive experiments to verify the effectiveness of the proposed improvement method. Section 5 offers some conclusions.

2. Traditional BM3D Algorithm

The BM3D algorithm is mainly divided into two steps. The first step is the basic estimation, and the second step is the final estimation. The first and second steps both include block-matching grouping, filtering, and aggregation. The execution process of the algorithm is shown in Figure 1.

2.1. Basic Estimation

2.1.1. Block-Matching Grouping

Construct a two-dimensional sliding window with a size of N 1 × N 1 , and take image blocks based on the set step size p on the noisy image I. Set the current block being operated on as Z R . Then search for similar blocks of the reference block Z R in the N 2 × N 2 search area centered around Z R , denoted as Z Q . Define the distance d Z R , Z Q (inversely proportional to similarity) to measure the similarity between Z R and Z Q , and its expression is shown in Equation (1).
d Z R , Z Q = N 1 1 γ τ 2 D Z R , λ thr 2 D σ 2 log N 1 2 γ τ 2 D Z Q , λ thr 2 D σ 2 log N 1 2 2
where τ 2 D is the 2D unitary transformation operator (such as DCT, DFT, etc.). γ is a hard threshold operator, λ thr 2 D is a fixed threshold parameter, σ represents the standard deviation of noise, 2 represents a binomial. The definition of γ is as follows.
γ ( λ , λ thr ) = λ | λ | > λ thr 0 otherwise
The idea of finding similar blocks for reference block Z R is to select an appropriate threshold τ 1 , and if d ( Z R , Z Q ) < τ 1 , Z Q is considered similar to Z R and added to the set B R . Finally, integrate the reference block Z R and their similar block sets into a three-dimensional matrix T R . Figure 2 is a simple example of block-matching grouping on the noisy image.

2.1.2. Hard Threshold Shrinkage

Perform a three-dimensional unitary transformation τ 3 D ht (two-dimensional Bior wavelet transform and one-dimensional Haar wavelet transform) on a three-dimensional matrix T R . Three-dimensional transform domain denoising can separate the noise and real information of an image while ensuring no energy loss. Most of the real information of the image is concentrated at the vertex of the 3D matrix energy, while the noise is often concentrated at the bottom of the 3D matrix energy. Therefore, through hard threshold filtering in the transformation domain, most of the image information can be preserved while effectively reducing the noise. V R ht = τ 3 D 1 ht γ τ 3 D ht ( T R ) , λ thr 3 D σ 2 log ( N 1 2 ) is the set of estimated values for blocks in B R after hard threshold filtering, and λ thr 3 D is a fixed threshold parameter.

2.1.3. Aggregation

After hard threshold shrinkage, each block will obtain a basic estimated weight value, N R represents the number of non-zero coefficients in the matrix V R ht . The basic estimated weight value ω R basic of the reference block is shown in Equation (3).
ω R basic = 1 N R N R 1 1 N R = 0
For a pixel i that may appear in multiple blocks, it is necessary to perform a weighted average on the estimated values of these overlapping blocks to obtain the basic estimation value of i. The formula is as follows:
Y basic ( i ) = B R Z Q B R ω R basic · V R Q ht B R Z Q B R ω R basic · x Q , i I
where Z Q represents any image block containing pixel i in the set B R , and similarly, there may be multiple sets B R with image blocks containing pixel i. Therefore, it is necessary to weighted-average the estimated values of all blocks containing that pixel in the set. V R Q ht is the estimated value of the block Z Q containing pixel i in any set B R , x Q is a 0–1 variable, and the expressions for V R Q ht and x Q are as follows:
V R Q ht = V R Q ht i Z Q 0 i Z Q
x Q = 1 i Z Q 0 i Z Q

2.2. Final Estimation

2.2.1. Block-Matching Grouping

Based on the basic estimation image Y basic obtained in the first step, block-matching grouping is performed again to form a new three-dimensional matrix T R 2 . There are two three-dimensional matrices at this time, one is the three-dimensional matrix T R composed of similar blocks in the noisy image obtained in the first step, and the other is the three-dimensional matrix T R 2 composed of similar blocks in the image generated from the basic estimation.

2.2.2. Collaborative Wiener Filtering

Unlike the use of hard threshold filtering in the basic estimation stage, the final estimation utilizes Wiener filtering. The attenuation co-efficient of the Wiener filter is calculated as follows:
W R final = τ 3 D wie T R 2 2 τ 3 D wie T R 2 2 + σ 2
Then, the Wiener filtering of T R is achieved by multiplying the 3D transform co-efficient τ 3 D wie ( T R ) of the noisy image with the Wiener contraction co-efficient. Finally, the estimation value of the block is generated through inverse transformation. The expression for Wiener filtering is as follows:
V R wie = τ 3 D 1 wie W R final τ 3 D wie T R
The weight of each reference block in the final estimation stage can be defined as:
ω R final = σ 2 W R final 2 2

2.2.3. Aggregation

Similarly, it is necessary to perform a weighted average on these block estimates to obtain the final estimate of the pixel i. The formula is as follows:
Y final ( i ) = B R Z Q B R ω R final · V R Q wie B R Z Q B R ω R final · x Q , i I
where V R Q wie is the estimated value of the block Z Q containing pixel i in any set B R , and its expression is similar to Equation (5).

3. Proposed Method

The process of searching for similar blocks in the traditional BM3D algorithm is carried out on the noisy image, which does not take into account the deep-level features of the noisy image and may lead to inaccurate block-matching results. To fully consider the depth features at different scales of the noisy image and improve the accuracy of block-matching, we propose the modified block-matching based on noisy image depth features and the improved similarity measurement based on SSIM. The specific instructions are as follows.

3.1. The Modified Block-Matching Based on UNet Denoising Network Feature Maps

Given the advantages of deep-learning networks in extracting deep-level features of images, we select a pre-trained UNet denoising network to extract feature maps of the noisy image, and then search for similar blocks on the feature maps. Based on the results of block-matching on each feature map, all corresponding similar blocks are searched for at the same position in the noisy image, thus obtaining a set of similar blocks that consider the deep-level features of the noisy image. Based on this, subsequent operations on BM3D are carried out. The execution process of the improved BM3D algorithm is shown in Figure 3.
Compared with the flowchart of the traditional BM3D algorithm in Figure 1, it can be seen that the improved BM3D algorithm in this paper adds a feature extractor module before the block-matching of the traditional BM3D algorithm, which first inputs the noisy image into a pre-trained UNet denoising network to obtain feature maps at different scales of the noisy image. Figure 3 only lists 9 feature maps of different layers. The UNet denoising network consists of 4 processes, with process 1 being the head and having 9 layers with 48 channels per layer. Process 2 is down, with 20 layers and 48 channels per layer. Process 3 is up, with 36 layers and 96 channels per layer. Process 4 is last, with 10 layers and 3 channels per layer. Select a feature map from all channels in each layer and arrange it to form the feature maps of each layer of the UNet network shown in Figure 4. It can be seen that the UNet network has many layers, and the number of channels varies greatly among different process layers. The head and down processes of UNet networks are similar to the encoder structure and can compress images. The up and last processes are similar to the decoder structure, which can restore the image to the denoised image, making the image features more obvious. The last layer obtains the denoised image of the noisy image. Therefore, to fully consider the deep-level features of the noisy image, improve the accuracy of block-matching, and enhance the denoising effect of the BM3D algorithm, we select the feature maps of layers in the last process for improved block-matching. Figure 5 is a simple example of the modified block-matching module.
Unlike Figure 2, which performs a similar block search on the noisy image, as shown in Figure 5, it is assumed that three feature maps of the noisy image are selected from the UNet network, namely feature map 1, feature map 2, and feature map 3. Then search for similar blocks of reference block R on three feature maps, where the similar blocks of R in feature map 1 are block Q1 and block Q2, the similar blocks of R in feature map 2 are block Q1 and block Q3, and the similar blocks of R in feature map 3 are block Q4 and block Q5. Then, based on the matching of similar block positions on each feature map, select all non-repetitive similar blocks at the same positions on the noisy image, and use them as a set of similar blocks for reference block R. The set of similar blocks selected through the above process fully considers the impact of different scale features of the noisy image on block-matching, improving the accuracy of block-matching.
The next step is to input the 3D matrix formed by the set of similar blocks of each block into the next step, which is the same process as the traditional BM3D algorithm. Similarly, after the first step of algorithm execution, the basic estimated image needs to undergo feature map extraction and modified block-matching. Similar block sets that consider deep-level features are obtained from the basic estimated image and merged to form a 3D matrix T R 2 . Then, the 3D matrix T R obtained from the first step on the noisy image and T R 2 are inputted into the next step for subsequent operations.

3.2. The Improved Similarity Measurement Criteria Based on SSIM

From Equation (1), it can be seen that the similarity measurement in the traditional BM3D algorithm is a simple operation based on the pixels of the image, which fails to consider the structural features of the image itself. SSIM is a measure of image similarity, which effectively integrates the structure, brightness, and contrast information of the image. These features can promote human visual perception and have the advantage of considering image structural features. The expression of SSIM is shown in Equation (11). Therefore, to further enhance the correlation between the BM3D algorithm and the structural features of the image itself, and improve the visual quality of the denoised image, we utilize SSIM to improve Equation (1). The improved similarity measurement criteria are shown in Equation (12):
S S I M ( Z R , Z Q ) = ( 2 μ R μ Q + C 1 ) ( 2 σ R Q + C 2 ) ( μ R 2 + μ Q 2 + C 1 ) ( σ R 2 + σ Q 2 + C 2 )
where μ R is the average of the reference block Z R , μ Q is the average of the block Z Q , σ R 2 is the variance of the reference block Z R , σ R 2 is the variance of the block Z Q , and σ R Q is the covariance between Z R and Z Q . C 1 and C 2 are constants that maintain stability, where C 1 = 0.0004 , C 2 = 0.0036 in the experiment.
d modified ( Z R , Z Q ) = d ( Z R , Z Q ) × [ 1 S S I M ( Z R , Z Q ) ]
In Equation (12), the smaller the distance d ( Z R , Z Q ) , the closer the distance between the reference block Z R and the block Z Q , indicating a higher similarity between the two blocks. S S I M ( Z R , Z Q ) is a number between 0 and 1, and the larger it is, the higher the structural similarity between Z R and Z Q , and the smaller 1 S S I M ( Z R , Z Q ) . According to Equation (12), the distance d modified ( Z R , Z Q ) between Z R and Z Q is also smaller. In summary, the smaller d modified ( Z R , Z Q ) , the higher the similarity between Z R and Z Q . It can be seen that the improved similarity measurement formula d modified ( Z R , Z Q ) not only measures the pixel value similarity between blocks but also takes into account the inherent features such as the structure between blocks due to the introduction of the SSIM. This has a good promoting effect on improving the visual quality of the denoised image.

4. Experimental Results and Analysis

To verify the denoising performance of the improved BM3D method proposed in this paper, we conduct extensive comparative experiments, and the experimental results are presented and discussed. First, based on the combination of different layer feature maps using the UNet denoising network to improve the denoising effect of BM3D, the optimal feature map combination is selected. Second, compare the improved BM3D method proposed in this paper with other classic denoising algorithms and display the denoising effect diagram. Then, compare the denoising performance of various denoising algorithms on images with different noise levels. Finally, the contributions of different components in the proposed method are analyzed.

4.1. Data and Parameter Settings

In this experiment, we utilize classic test images for image denoising, such as Cameraman, boat, and Baboon. Each image is a Grayscale with a size of 256 × 256, including both noiseless images and images with various noise levels. We adopt pre-trained UNet network models on datasets with different noise levels and utilize these UNet models to extract feature maps of images with different noise levels for experiments. In addition, we utilize the peak signal-to-noise ratio (PSNR) and SSIM as evaluation indicators for image-denoising performance. The larger the PSNR, the better the denoising effect. The larger the SSIM, the higher the visual quality of the denoised image. The parameter settings for traditional BM3D and improved BM3D algorithm are the same, as shown in Table 1.

4.2. Selection of the Optimal Combination of Feature Maps

The UNet network has four processes: head, down, up, and last. Since feature maps of each layer in each process are similar, we select the last layer for each process to conduct experiments. A total of four layers are selected, namely head.1.block.1 (feature layer 1), down_Path.4.block.1 (feature layer 2), up_Path.4.conv_2.Block.1 (feature layer 3) and last.2 (feature layer 4). Due to the fact that each layer has multiple channels, there are many combinations of feature maps, including feature map combinations of different channels in any of the above four layers. The main purpose of this paper is to verify that the denoising performance of the improved BM3D algorithm based on deep feature maps is superior to the traditional BM3D algorithm. Therefore, for the convenience of the experiment, the way of obtaining feature map combinations in this experiment is as follows: first, select any two layers from the above four layers, with a total of 10 combinations, including 1&1 (representing feature layer 1 and feature layer 1), 1&2 (representing feature layer 1 and feature layer 2, and so on), 1&3, 1&4, 2&2, 2&3, 2&4, 3&3, 3&4, and 4&4. Then select one channel from each of these two layers and treat the selected two channels as a combination, without considering other combinations. Based on the principle of feature map combination selection above, we calculate the PSNR value of BM3D improved by combining feature maps of two channels in any two layers using an exhaustive method. By comparing the denoising effects of various combinations under these two layers, the PSNR value of the combination with the best denoising effect is selected as the optimal PSNR value for these two layers. Table 2 shows the optimal PSNR values for any two layers.
From Table 2, it can be seen that the largest terms of PSNR appear in the combination of layers 4&4. In addition, the denoising effect of the combination containing feature layer 2 is not good, especially the denoising effect of the 2&2 combination is the worst in each image. Further observation reveals that the PSNR values of various combinations exhibit a pattern of 1&2 < 1&3 < 1&4, 2&2 < 2&3 < 2&4, indicating that the denoising effect of the feature map of the later layer is better than that of the front layer. This is because the structure of the UNet denoising network is similar to the Autoencoder, and feature layer 2 is located in the down-compression process, so the feature map of this layer will become fuzzy, and the features are not obvious. Therefore, the effect of block-matching with the matching of feature layer 2 is not good. Feature layer 3 is located in the up-decoding process. After denoising by the network, the feature map of this layer begins to become clear, and the features are obvious. Feature layer 4 is located in the last layer of the last process, which is the output of the UNet network. Therefore, the feature map image of this layer is the clearest and the features are also the most obvious. Based on the above experimental results and analysis, the improved BM3D algorithm in the subsequent experiments in this paper adopts a combination of feature maps of layers 4&4.

4.3. Performance Comparison of Different Denoising Methods

To evaluate the performance of the improved BM3D algorithm proposed in this paper, we compared two classic nonlocal mean denoising algorithms, BM3D and NLM, respectively. We also selected three advanced deep-learning denoising networks, namely BM3D-Net [16], DIVA [17] and UNet. In addition, a comparison was made with a sparse encoding-based denoising algorithm NCSR [37], and the experimental results are shown in Table 3.
Table 3 shows that our proposed method has better PSNR and SSIM values than the other four methods. Compared to UNet, our method has an average increase in PSNR of 3.24 dB, and an average increase in SSIM of 0.2181 compared to NLM. Compared with BM3D and NCSR, our method has an average increase of 0.26–0.39 dB in PSNR value and 0.0076–0.011 in SSIM value. Compared with advanced DIVA and BM3D-Net, our method has some advantages in PSNR. In addition, for individual images, DIVA’s SSIM is better than our method, but overall, our method has better denoising performance. To further demonstrate that our method outperforms the other methods in terms of denoising performance, we utilize the image of Baboon as an example to visualize the denoising effects of various methods with different noise levels. In Figure 6, Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, “original” represents the original image without noise, “noisy” represents the noisy image, and the rest are the denoised images of BM3D, DIVA, NCSR, NLM, BM3D-Net and Our method in order.
Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10 show the denoising effects of six denoising algorithms on Baboon with noise levels of σ = 10 , σ = 30 , and σ = 50 , respectively. Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11 show the local denoising effects of six denoising algorithms on Baboon with noise levels of σ = 10 , σ = 30 , and σ = 50 , respectively. It can be seen that after denoising with BM3D, NCSR, NLM, and BM3D-Net algorithms, the complex texture areas such as Baboon’s nose bridge, facial hair, and beard become smoother, resulting in a certain degree of distortion. Although the image denoised by the DIVA algorithm retains a certain degree of texture, its denoising effect on noise is not particularly ideal. Our method not only effectively removes noise, but also preserves good texture and has high visual quality in the denoised image. With the increase in noise level, other algorithms’ denoised images exhibit blurry and smooth phenomena. Although our method also exhibits some degree of distortion, it still has a certain visual quality compared to others. The above experiments can verify that the proposed method for block-matching on feature maps fully considers the deep-level features of images at different scales, therefore improving the accuracy of block-matching. Moreover, the proposed method adopts SSIM to improve the similarity measurement criterion for block-matching, which further improves the visual quality of the denoised image.

4.4. Comparison of Various Algorithms under Different Noise Levels

Table 3 presents the numerical results of various denoising algorithms at noise level σ = 10 . To verify that the proposed method performs better than other methods at different noise levels, six images with noise levels of σ = 10 , σ = 20 , σ = 30 and σ = 40 are selected for the experiment. The obtained PSNR values are visualized in Figure 12.
From Figure 12, it can be seen that as the noise level increases, the denoising performance of various denoising methods decreases. This is because the more noise, the less detailed information of the image is presented, making the algorithm unable to utilize more features of the image, resulting in a decrease in denoising performance. Overall, the method proposed in this paper exhibits the best performance under various noise levels and uniformly outperforms other methods. In addition, we can see from the graph that for individual images, the PSNR value of the BM3D-Net algorithm at a noise level of σ = 10 is lower than the PSNR value at a noise level of σ = 20 . The main reason for this phenomenon is that the network should be trained separately for different noise levels. In addition, this is also related to the data. For example, differences in noise distribution and changes in image content and structure. These characteristics can also lead to abnormal performance of algorithms at certain noise levels.

4.5. Ablation Experiment

We conduct ablation experiments to better analyze the contributions of different components of the method proposed in this paper. In the following example, we test four methods with different components.
(1) BM3D: Traditional BM3D method.
(2) BM3D + SSIM: A method for improving traditional BM3D based on SSIM.
(3) BM3D + Feature: A method for improving traditional BM3D based on UNet de-noising network feature maps.
(4) BM3D + SSIM + Feature: A method for improving traditional BM3D based on SSIM and UNet denoising network feature maps.
Based on the experimental results in Table 4, the following observations can be obtained. First, the denoising effect of the improved BM3D based on SSIM is better than that of the traditional BM3D method, which is reflected in the average increase of 0.06 dB in PSNR value and 0.0019 in SSIM value. This is because the SSIM considers the structure, brightness, and contrast information of the image, so the denoised image has higher visual quality. Second, the denoising effect of the improved BM3D based on UNet denoising network feature maps is superior to the traditional BM3D method, specifically reflected in an average increase of 0.19 dB in PSNR value and 0.0055 in SSIM value. This is because block-matching on feature maps takes into account the depth features of different scales of the noisy image, resulting in more accurate block-matching results and better denoising effects. Third, the denoising effect of the improved BM3D based on SSIM and UNet denoising network feature maps is higher than the other three methods. Compared with the traditional BM3D algorithm, the average PSNR value increases by 0.26 dB, and the average SSIM value increases by 0.0076. The above experimental results verify that both the UNet denoising network feature map and the improved BM3D algorithm based on SSIM proposed in this paper can improve the denoising performance of the traditional BM3D algorithm.

4.6. Application Case

To verify the denoising potential of the denoising method proposed in this paper, we selected some real-life image data for denoising, including one satellite image, one map image, one ball image, and one calligraphy image. As shown in Figure 13, the satellite image is from the FloodNet [38] dataset, and the remaining three images are from the PolyU [39] dataset.
We add Gaussian noise with a noise level of σ = 10 to each of the four images and then utilize the improved BM3D method proposed in this paper for denoising. In Figure 13, we present the original image, noisy image, and denoised image of each image, and enlarge the local details of each image below the image. In addition, we also list the PSNR and SSIM numerical results of the noisy and denoised images at the bottom of the image. On the one hand, it can be seen from the denoised image and the enlarged image of local details that the proposed method effectively removes noise from the image and preserves the detailed features of the image clearly. On the other hand, from the numerical results, it can be seen that compared with the PSNR and SSIM values of noisy images, the proposed method has significantly improved PSNR and SSIM. From the above analysis, we can conclude that the improved BM3D method proposed in this paper also has excellent denoising effects on real-life images and has good denoising potential.

5. Conclusions

We propose an improved BM3D denoising algorithm based on the UNet denoising network feature map and Structural Similarity Index (SSIM). Compared to the traditional BM3D algorithm that directly performs block-matching on the noisy image, this method takes into account the deep-level features of different scales of the image and improves the accuracy of block-matching. Moreover, SSIM is utilized to improve the similarity measurement criterion, taking into account the structure, brightness, and contrast information of the image, further enhancing the denoising effect of the BM3D algorithm. In addition, it should be pointed out that the time complexity of the method proposed is relatively high, mainly due to the use of the SSIM to improve the similarity measurement criterion for BM3D block-matching, which involves SSIM calculation between many blocks of the image. In areas with high requirements for image quality, the proposed method can play a good role in non-real-time applications such as image restoration and image enhancement. Finally, through several experiments, the following conclusions can be drawn:
(1) Feature map combinations from different layers of the UNet denoising network are utilized to improve the block-matching process of BM3D, the denoising effect of feature map combinations containing feature layer 2 is poor, while the denoising effect of feature map combinations containing feature layer 4 is excellent.
(2) Compared to the six classic denoising algorithms of traditional BM3D, NLM, UNet, NCSR, DIVA, and BM3D-Net, the improved BM3D method proposed in this paper both has advantages in PSNR and SSIM. In addition, the proposed method outperforms other methods in denoising images with different noise levels.
(3) The block-matching process of the improved BM3D algorithm proposed in this paper fully considers the deep-level features of the image at different scales, as well as the structure, brightness, and contrast information of the image. After denoising, the detailed features of the image can be well preserved.
(4) The UNet denoising network feature map and SSIM in the proposed method in this paper both have a promoting effect on improving the denoising performance of the traditional BM3D algorithm. Using both to improve the denoising effect of the traditional BM3D algorithm is better than using either of them alone.

Author Contributions

Conceptualization, J.C. and Z.Q.; methodology, J.C., Z.Q. and H.L.; software, J.C.; validation, Z.Q. and L.H.; formal analysis, H.L. and L.H.; investigation, J.C., Z.Q. and H.L.; resources, Z.Q. and L.H.; writing—original draft preparation, J.C.; writing—review and editing, J.C., Z.Q., F.D. and H.L.; visualization, J.C.; supervision, F.D.; funding acquisition, Z.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the project of the Natural Science Foundation of China (Grant No. 12163004); the Yunnan Fundamental Research Projects (CN) (Grant Nos. 202301BD070001-008, 202001AT070135, 202101BD070001-053); the Project of Key Science Foundation of Yunnan Province under Grant No. 202101AS070007.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shan, S.; Li, Y.; Zhu, S. BM3D denoising based on minimum GCV score. In Proceedings of the 2015 International Conference on Computers, Communications, and Systems (ICCCS), Kanyakumari, India, 2–3 November 2015; pp. 154–158. [Google Scholar]
  2. Lin, T.; Hong, H.; Wu, L. Improved BM3D for Real Image Denoising. In Proceedings of the 2021 13th International Conference on Wireless Communications and Signal Processing (WCSP), Online, 20–22 October 2021; pp. 1–5. [Google Scholar]
  3. Han, X.; Sun, Q.; Li, Y.; Ye, F. A Novel Sonar Image Denoising Algorithm based on Block Matching. In Proceedings of the 2022 International Conference on Microwave and Millimeter Wave Technology (ICMMT), Harbin, China, 12–15 August 2022; pp. 1–3. [Google Scholar]
  4. Su, Y.; Shen, Z.; Long, X.; Chen, C.; Qi, L.; Chao, X. Gaussian filtering method of evaluating the elastic/elasto-plastic properties of sintered nanocomposites with quasi-continuous volume distribution. Mater. Sci. Eng. 2023, 872, 145001. [Google Scholar] [CrossRef]
  5. Mishiba, K. Fast Guided Median Filter. IEEE Trans. Image Process. 2023, 32, 737–749. [Google Scholar] [CrossRef] [PubMed]
  6. Kilic, H.; Ceyhan, S.; Gerek, O.N. A novel family of edge preserving anisotropic filters. Digit. Signal Process. 2022, 128, 103623. [Google Scholar] [CrossRef]
  7. Du, S.; Chen, S. Study on Optical Fiber Gas-Holdup Meter Signal Denoising Using Improved Threshold Wavelet Transform. IEEE Access 2023, 11, 18794–18804. [Google Scholar] [CrossRef]
  8. Buades, A.; Coll, B.; Morel, J.M. A non-local algorithm for image denoising. In Proceedings of the Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005. [Google Scholar]
  9. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef]
  10. Shao, G.; Gao, Y.; Zuo, J.; Yue, Y.; Yang, J. An Improved BM3D Method for eDNA Mieroarray Image Denoising. In Proceedings of the 2018 13th International Conference on Computer Science & Education (ICCSE), Colombo, Sri Lanka, 8–11 August 2018. [Google Scholar]
  11. Dutta, S.; Basarab, A.; Georgeot, B.; Kouamé, D. Quantum Mechanics-Based Signal and Image Representation: Application to Denoising. IEEE Open J. Signal Process. 2021, 2, 190–206. [Google Scholar] [CrossRef]
  12. Dutta, S.; Basarab, A.; Georgeot, B.; Kouamé, D. A novel image denoising algorithm using concepts of quantum many-body theory. Signal Process. 2022, 201, 108690. [Google Scholar] [CrossRef]
  13. Fang, W.; Li, H. A self-supervised CNN for image denoising with self-similarity prior. In Proceedings of the 2022 16th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 21–24 October 2022; Volume 1, pp. 66–69. [Google Scholar]
  14. Zha, Z.; Yuan, X.; Wen, B.; Zhang, J.; Zhou, J.; Zhu, C. Image Restoration Using Joint Patch-Group-Based Sparse Representation. IEEE Trans. Image Process. 2020, 29, 7735–7750. [Google Scholar] [CrossRef]
  15. Zha, Z.; Yuan, X.; Wen, B.; Zhou, J.; Zhu, C. Group Sparsity Residual Constraint with Non-Local Priors for Image Restoration. IEEE Trans. Image Process. 2020, 29, 8960–8975. [Google Scholar] [CrossRef]
  16. Yang, D.; Sun, J. BM3D-Net: A Convolutional Neural Network for Transform-Domain Collaborative Filtering. IEEE Signal Process. Lett. 2018, 25, 55–59. [Google Scholar] [CrossRef]
  17. Dutta, S.; Basarab, A.; Georgeot, B.; Kouamé, D. DIVA: Deep unfolded network from quantum interactive patches for image restoration. arXiv 2022, arXiv:2301.00247. [Google Scholar]
  18. Dutta, S.; Basarab, A.; Georgeot, B.; Kouamé, D. Deep Unfolding of Image Denoising by Quantum Interactive Patches. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; pp. 491–495. [Google Scholar]
  19. Han, H.; Wang, H.; Liu, Z.; He, Z. Non-local Denoising Convolutional Neural Network for Rolling Bearing Vibration Signal. In Proceedings of the 2020 International Conference on Sensing, Diagnostics, Prognostics, and Control (SDPC), Beijing, China, 5–7 August 2020; pp. 184–188. [Google Scholar]
  20. Feng, Q.; Tao, S.; Xu, C.; Jin, G. BM3D-GT&AD: An improved BM3D denoising algorithm based on Gaussian threshold and angular distance. IET Image Process. 2020, 14, 431–441. [Google Scholar]
  21. Ri, G.I.; Kim, S.J.; Kim, M.S. Improved BM3D method with modified block-matching and multi-scaled images. Multimed. Tools Appl. 2022, 81, 12661–12679. [Google Scholar] [CrossRef]
  22. Mahmood, S.Z.; Afzal, H.; Mufti, M.R.; Akhtar, N.; Habib, A.; Hussain, S. A Novel Method of Image Denoising: New Variant of Block Matching and 3D. J. Med. Imaging Health Inform. 2020, 10, 2490–2500. [Google Scholar] [CrossRef]
  23. Mbarki, Z.; Seddik, C.B.J.; Seddik, H. Building a modified block matching kernel based on wave atom transform for efficient image denoising. Egypt. J. Remote Sens. Space Sci. 2021, 24, 857–878. [Google Scholar] [CrossRef]
  24. Zhang, Y.S.; Zhu, S.J.; Li, Y.J. BM3D denoising algorithm with adaptive block-match thresholds. Appl. Mech. Mater. 2012, 229, 1715–1720. [Google Scholar] [CrossRef]
  25. Hou, Y.; Shen, D. Image denoising with morphology-and size-adaptive block-matching transform domain filtering. EURASIP J. Image Video Process. 2018, 2018, 59. [Google Scholar] [CrossRef] [PubMed]
  26. Zhong, X.; Xu, L.; Li, C.; An, L.; Wang, L. RFE-UNet: Remote Feature Exploration with Local Learning for Medical Image Segmentation. Sensors 2023, 23, 6228. [Google Scholar] [CrossRef]
  27. Su, S.; Shao, X.; He, L.; Lin, H.; Zuo, Y.; Qiang, Z. Face Image Completion Method Based on Parsing Features Maps. IEEE J. Sel. Top. Signal Process. 2023, 17, 624–636. [Google Scholar] [CrossRef]
  28. Liu, C.; Zhou, W.; Chen, Y.; Lei, J. Asymmetric Deeply Fused Network for Detecting Salient Objects in RGB-D Images. IEEE Signal Process. Lett. 2020, 27, 1620–1624. [Google Scholar] [CrossRef]
  29. Kumar, E.K.; Kishore, P.V.V.; Kumar, M.T.K.; Kumar, D.A.; Sastry, A.S.C.S. Three-Dimensional Sign Language Recognition with Angular Velocity Maps and Connived Feature ResNet. IEEE Signal Process. Lett. 2018, 25, 1860–1864. [Google Scholar] [CrossRef]
  30. Masood, S.; Rai, A.; Aggarwal, A.; Doja, M.N.; Ahmad, M. Detecting distraction of drivers using convolutional neural network. Pattern Recognit. Lett. 2020, 139, 79–85. [Google Scholar] [CrossRef]
  31. Ullah, R.; Asif, M.; Shah, W.A.; Anjam, F.; Ullah, I.; Khurshaid, T.; Wuttisittikulkij, L.; Shah, S.; Ali, S.M.; Alibakhshikenari, M. Speech Emotion Recognition Using Convolution Neural Networks and Multi-Head Convolutional Transformer. Sensors 2023, 23, 6212. [Google Scholar] [CrossRef]
  32. Sun, Y.; Bi, F.; Gao, Y.; Chen, L.; Feng, S. A multi-attention UNet for semantic segmentation in remote sensing images. Symmetry 2022, 14, 906. [Google Scholar] [CrossRef]
  33. He, X.; Wang, C.; Zheng, R.; Sun, Z.; Li, X. GPR image denoising with NSST-UNET and an improved BM3D. Digit. Signal Process. 2022, 123, 103402. [Google Scholar] [CrossRef]
  34. Dai, W.; Mao, Y.; Yuan, R.; Liu, Y.; Pu, X.; Li, C. A novel detector based on convolution neural networks for multiscale SAR ship detection in complex background. Sensors 2020, 20, 2547. [Google Scholar] [CrossRef]
  35. Wang, Z.; Liu, J.; Li, G.; Han, H. Blind2Unblind: Self-Supervised Image Denoising with Visible Blind Spots. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 2017–2026. [Google Scholar]
  36. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  37. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally Centralized Sparse Representation for Image Restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef]
  38. Rahnemoonfar, M.; Chowdhury, T.; Sarkar, A.; Varshney, D.; Yari, M.; Murphy, R.R. FloodNet: A High Resolution Aerial Imagery Dataset for Post Flood Scene Understanding. IEEE Access 2021, 9, 89644–89654. [Google Scholar] [CrossRef]
  39. Xu, J.; Li, H.; Liang, Z.; Zhang, D.; Zhang, L. Real-world noisy image denoising: A new benchmark. arXiv 2018, arXiv:1804.02603. [Google Scholar]
Figure 1. Flow chart of the traditional BM3D algorithm.
Figure 1. Flow chart of the traditional BM3D algorithm.
Sensors 23 07265 g001
Figure 2. A simple example of block-matching grouping (in the figure, three reference blocks are filled with red, green, and blue colors, while the rest are similar blocks that are matched to each other. The dotted arrows point to the similar blocks to be matched, while the solid arrows point to the 3D similar blocks matrix.).
Figure 2. A simple example of block-matching grouping (in the figure, three reference blocks are filled with red, green, and blue colors, while the rest are similar blocks that are matched to each other. The dotted arrows point to the similar blocks to be matched, while the solid arrows point to the 3D similar blocks matrix.).
Sensors 23 07265 g002
Figure 3. The improved BM3D algorithm flowchart based on UNet network feature maps.
Figure 3. The improved BM3D algorithm flowchart based on UNet network feature maps.
Sensors 23 07265 g003
Figure 4. Feature maps of each layer of UNet network (where 1 to 75 respectively represent the serial numbers of the feature maps of each layer of the UNet network).
Figure 4. Feature maps of each layer of UNet network (where 1 to 75 respectively represent the serial numbers of the feature maps of each layer of the UNet network).
Sensors 23 07265 g004
Figure 5. A simple example of the modified block-matching (where the dotted box represents similar block-matching on the feature map).
Figure 5. A simple example of the modified block-matching (where the dotted box represents similar block-matching on the feature map).
Sensors 23 07265 g005
Figure 6. The denoising results of Baboon with σ = 10 under six denoising algorithms (the red box represents the local details to be displayed in Figure 7, similar to the following).
Figure 6. The denoising results of Baboon with σ = 10 under six denoising algorithms (the red box represents the local details to be displayed in Figure 7, similar to the following).
Sensors 23 07265 g006
Figure 7. Local denoising results of Baboon with σ = 10 under six denoising algorithms.
Figure 7. Local denoising results of Baboon with σ = 10 under six denoising algorithms.
Sensors 23 07265 g007
Figure 8. The denoising results of Baboon with σ = 30 under six denoising algorithms.
Figure 8. The denoising results of Baboon with σ = 30 under six denoising algorithms.
Sensors 23 07265 g008
Figure 9. Local denoising results of Baboon with σ = 30 under six denoising algorithms.
Figure 9. Local denoising results of Baboon with σ = 30 under six denoising algorithms.
Sensors 23 07265 g009
Figure 10. The denoising results of Baboon with σ = 50 under six denoising algorithms.
Figure 10. The denoising results of Baboon with σ = 50 under six denoising algorithms.
Sensors 23 07265 g010
Figure 11. Local denoising results of Baboon with σ = 50 under six denoising algorithms.
Figure 11. Local denoising results of Baboon with σ = 50 under six denoising algorithms.
Sensors 23 07265 g011
Figure 12. PSNR values of various denoising methods under different noise levels. “□”: Improved BM3D; “+”: DIVA; “*”: NCSR; “∘”: BM3D; “⋄”: NLM; “x”: BM3D-Net.
Figure 12. PSNR values of various denoising methods under different noise levels. “□”: Improved BM3D; “+”: DIVA; “*”: NCSR; “∘”: BM3D; “⋄”: NLM; “x”: BM3D-Net.
Sensors 23 07265 g012
Figure 13. The denoising image results of real-life images (the red and green boxes indicate the local details to be displayed).
Figure 13. The denoising image results of real-life images (the red and green boxes indicate the local details to be displayed).
Sensors 23 07265 g013
Table 1. BM3D algorithm parameter settings.
Table 1. BM3D algorithm parameter settings.
N 1 ht N 2 ht p ht τ 1 ht λ thr 2 D λ thr 3 D λ thr N 1 wie N 2 wie p wie τ 1 wie
8813250.820.752.78163400
Table 2. Denoising results of improved BM3D algorithm under different UNet denoising network feature map combinations based on PSNR.
Table 2. Denoising results of improved BM3D algorithm under different UNet denoising network feature map combinations based on PSNR.
ImageOriginal1&11&21&31&42&22&32&43&33&44&4
Cameraman34.3934.3734.2634.2834.3733.6633.8933.9534.2834.4034.59
Man33.0232.9832.8832.9532.9832.4232.6332.7132.9233.0233.24
boat33.3233.2233.1533.2833.2832.6932.8932.9733.3033.3633.58
Baboon30.5430.5430.5330.5530.5630.3530.4730.5030.5430.5630.68
Flowers131.9632.1432.0932.1232.1431.8931.9732.0032.1432.1432.25
Computer33.5133.3433.0333.2633.3532.2632.6132.7733.2133.4633.75
Traffic33.2033.1433.0333.0533.1332.5732.7632.8433.0933.2133.30
Alley32.4232.3631.9432.3632.4131.9232.1332.2032.3532.4732.51
Trees29.6029.6229.6029.6029.6129.5129.5429.5529.5829.6429.68
Gardens30.5630.6030.5330.5730.6430.3730.4430.4730.6430.6530.70
Plaza34.1434.2034.1134.1834.3333.9234.0134.1234.3434.3834.41
pentagon31.2631.2331.0431.1831.2430.6330.8230.9131.1431.3431.47
Street134.3934.3534.2134.2134.3433.7033.9534.0234.2834.4134.61
Building233.7132.9933.0033.2733.7133.2133.3133.3933.6033.7033.72
mean32.5732.5132.3932.4932.5832.0832.2432.3132.5332.6232.75
The image noise level used in this experiment is σ = 10 , with bold values representing the maximum value of the row. The first column “original” represents the traditional BM3D algorithm, and the last row “mean” represents the average of each column.
Table 3. Denoising results of different algorithms based on PSNR and SSIM ( σ = 10 ).
Table 3. Denoising results of different algorithms based on PSNR and SSIM ( σ = 10 ).
ImageBM3DUNetNLMDIVABM3D-NetNCSROURS
Cameraman34.39/0.928530.38/0.849232.26/0.527933.89/0.924029.18/0.918934.19/0.931234.64/0.9338
Man33.02/0.913230.91/0.864131.75/0.743832.65/0.901329.51/0.894732.94/0.909733.29/0.9208
boat33.32/0.920731.43/0.882731.65/0.724833.03/0.936927.96/0.890533.19/0.915933.63/0.9295
Computer33.51/0.942230.92/0.904031.55/0.733333.18/0.937325.66/0.903433.27/0.939833.79/0.9490
Traffic33.20/0.930430.58/0.883531.32/0.694132.90/0.929530.60/0.930533.02/0.932033.36/0.9335
Baboon30.54/0.912727.89/0.846828.72/0.803629.96/0.917826.26/0.872030.37/0.902430.72/0.9189
Flowers131.96/0.874629.56/0.804130.03/0.581031.76/0.935631.64/0.886332.01/0.879232.34/0.8905
Gardens30.56/0.920926.47/0.841727.41/0.759230.04/0.911819.92/0.874430.34/0.911930.76/0.9291
Trees29.60/0.936125.18/0.836326.62/0.847028.91/0.948921.78/0.839529.32/0.926729.72/0.9397
pentagon31.26/0.904027.76/0.787829.81/0.780030.77/0.904429.42/0.875431.39/0.896831.67/0.9157
Street134.39/0.931231.92/0.895232.47/0.604534.04/0.924334.10/0.932534.25/0.931934.70/0.9376
mean32.34/0.919529.36/0.854130.33/0.709031.92/0.924727.82/0.892632.21/0.916132.60/0.9271
The numerical results in the table are A/B, where A represents PSNR, and B represents SSIM. The larger the values of A and B, the better the denoising performance of the algorithm. The last row “mean” represents the average of each column, and the bold values represent the maximum value of that row.
Table 4. Denoising results of different components of the method proposed in this paper based on PSNR and SSIM ( σ = 10 ).
Table 4. Denoising results of different components of the method proposed in this paper based on PSNR and SSIM ( σ = 10 ).
ImageBM3DBM3D + SSIMBM3D + FeatureBM3D + Feature + SSIM
Cameraman34.39/0.928534.42/0.929034.59/0.932334.64/0.9338
Man33.02/0.913233.06/0.915133.24/0.919433.29/0.9208
boat33.32/0.920733.37/0.923033.58/0.927733.63/0.9295
Computer33.51/0.942233.56/0.943033.75/0.948133.79/0.9490
Traffic33.20/0.930433.25/0.931433.30/0.931733.36/0.9335
Baboon30.54/0.912730.59/0.914530.68/0.917630.72/0.9189
Flowers131.96/0.874632.06/0.878432.25/0.887132.34/0.8905
Gardens30.56/0.920930.62/0.922930.70/0.926930.76/0.9291
Trees29.60/0.936129.63/0.937229.68/0.938429.72/0.9397
pentagon31.26/0.904031.37/0.908531.47/0.910031.67/0.9157
Street134.39/0.931234.46/0.932534.61/0.935934.70/0.9376
mean32.34/0.919532.40/0.921432.53/0.925032.60/0.9271
The numerical results in the table are A/B, where A represents PSNR, and B represents SSIM. The larger the values of A and B, the better the denoising performance of the algorithm. The last row “mean” represents the average of each column, and the bold values represent the maximum value of that row.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cao, J.; Qiang, Z.; Lin, H.; He, L.; Dai, F. An Improved BM3D Algorithm Based on Image Depth Feature Map and Structural Similarity Block-Matching. Sensors 2023, 23, 7265. https://doi.org/10.3390/s23167265

AMA Style

Cao J, Qiang Z, Lin H, He L, Dai F. An Improved BM3D Algorithm Based on Image Depth Feature Map and Structural Similarity Block-Matching. Sensors. 2023; 23(16):7265. https://doi.org/10.3390/s23167265

Chicago/Turabian Style

Cao, Jia, Zhenping Qiang, Hong Lin, Libo He, and Fei Dai. 2023. "An Improved BM3D Algorithm Based on Image Depth Feature Map and Structural Similarity Block-Matching" Sensors 23, no. 16: 7265. https://doi.org/10.3390/s23167265

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop