Next Article in Journal
Linear Discriminant Analysis-Based Motion Classification Using Distributed Micro-Doppler Radars with Limited Backhaul
Previous Article in Journal
Character Recognition of Components Mounted on Printed Circuit Board Using Deep Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Light Field Compression Efficiency through BM3D-Based Denoising Using Inter-View Correlation

1
Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Korea
2
Department of Information and Communication Engineering, Inha University, Incheon 22212, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(9), 2919; https://doi.org/10.3390/s21092919
Submission received: 19 February 2021 / Revised: 7 April 2021 / Accepted: 19 April 2021 / Published: 21 April 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Multi-view or light field images have recently gained much attraction from academic and commercial fields to create breakthroughs that go beyond simple video-watching experiences. Immersive virtual reality is an important example. High image quality is essential in systems with a near-eye display device. The compression efficiency is also critical because a large amount of multi-view data needs to be stored and transferred. However, noise can be easily generated during image capturing, and these noisy images severely deteriorate both the quality of experience and the compression efficiency. Therefore, denoising is a prerequisite to produce multi-view-based image contents. In this paper, the structural characteristics of linear multi-view images are fully utilized to increase the denoising speed and performance as well as to improve the compression efficiency. Assuming the sequential processes of denoising and compression, multi-view geometry-based denoising is performed keeping the temporal correlation among views. Experimental results show the proposed scheme significantly improves the compression efficiency of denoised views up to 76.05%, maintaining good denoising quality compared to the popular conventional denoise algorithms.

1. Introduction

Multi-view images or light field data, acquired from a multi-camera array by capturing a plurality of images of different viewpoints, are used in various applications, including the synthesis of a high-resolution image with a high dynamic range [1] or the reconstruction of an occluded area through synthetic aperture photography (SAP) [2]. The light field information of dense multi-view images is frequently used in such applications as post-refocusing [3], super-resolution [4], occlusion reconstruction [5] and depth estimation. Recently, light field-based virtual reality (VR) recognizes an interactive and immersive “walk-around space” by rendering an image of an arbitrary viewpoint through a combination of pixels [6,7]. During the image capturing step, noise can be generated for various reasons. Noise appears in various ways due to various causes, such as the amount of light reaching the camera sensor or reflecting off the dust in the air. This type of noise can arise in far more diverse patterns in multi-view images than in single-view images. In particular, since the multi-view is photographed from multiple points or multiple cameras, the pattern of noise generated may be more diverse. It takes a very long time to remove the noise generated in a large amount of multi-view. The above has been added to the manuscript. When a user walks around in a light field or a multi-view based virtual space wearing a head-mounted display (HMD), the display is close to the eyes such that noise may severely deteriorate the quality of the experience. In addition, high-quality multi-view VR contents with a resolution exceeding 4 K require far more overhead for the storage and transmission of data compared to the conventional high-resolution images or the popular 360-degree video. Noise decreases the temporal correlation of the pseudo sequence, consisting of multi-view images (hereinafter referred to as multi-view sequence); therefore, the compression efficiency is reduced and transmission in an insufficient network bandwidth becomes difficult. As a result, denoising is a prerequisite to produce multi-view-based image contents for next generation immersive media to satisfy both the expected image quality and the compression efficiency.
A tremendous amount of works has been done in an effort to reduce noise in images. Table 1 summarizes the previous works. These research efforts have been extended from single-view image to videos and multi-view images. For denoising in single-view images, many studies based on the non-local mean (NLM) [8] algorithm have been suggested. Block-matching and 3-dimensional (3D) filtering (BM3D) [9,10], referring to an enhanced version of NLM, show excellent performance, but the processing time can be very long due to high computational complexity. Recently, denoising using a convolutional neural network (CNN) has been studied [11,12,13,14]. Studies of a self-guided network (SGN) [11] and FFDNet [14] focus on computational efficiency considering the trade-off between the inference speed and the denoising performance. Research efforts on removing real noise, as opposed to synthesized gaussian noise, are also being attempted [12]. Lefkimmiatis et al. construct two networks, a CNN layer as a core component and a layer for the non-local self-similarity properties of natural images, to handle a wide range of noise levels [13].
In studies involving videos, denoising techniques are based on the feature of consecutive frames or views having a high correlation. The Video BM3D (VBM3D) algorithm [15] increases the denoising quality by performing block matching not in a single image but in temporally neighboring frames. In the VBM4D algorithm [16], the motion vectors (MVs) between frames are tracked to obtain the trajectories by which to identify and group similar blocks. As filtering is conducted on the blocks in the trajectories, the temporal consistency can be fairly well maintained. However, because the tracking is limited to the horizontal and vertical directions, the denoising performance is poor for multi-view images with complicated textures. Execution times are much longer than in BM3D because the multi-view geometry is not actively considered. Research involving the grouping of similar patches using the optical flow shows denoising results that exceed that of the VBM4D, but the computational complexity is also very high [17]. Video noise is removed by training information between one frame and multiple frames [18], and a tensor approach was applied in a video denoising method called sparse and low-rank tensor (SALT) [19].
As multi-view applications have attracted attention, denoising research in this area is also increasing. Miyata et al. [20] propose an algorithm for the rapid denoising of multi-view images using the NLM with a plane sweeping (PS). In work by Zhou et al. [21], one of multiple views is set as the target view. The 3D focus image stack for the target view is made from other views. The disparity map obtained from the 3D focus image stacks is then used to generate a clean target view. When a densely arranged 2-dimensional (2D) camera array is used, a highly accurate depth map can be obtained from the 3D focus image stack. However, it does not work well for views with severe noise or wide view distances. Many multi-view algorithms using a CNN have also been proposed. Multi-view image denoising algorithm based on convolutional neural network (MVCNN) [22] that represents an enhancement of earlier work was also proposed [21]. Noisy 3D focus image stacks are fed into a denoising convolutional neural network (DnCNN) [23] and a disparity map is calculated from the denoised 3D focus image stacks. However, these types of approaches in which multiple views are used as references for one target view are aimed at denoising a single-view image utilizing the abundant information, but they do not take full advantage of the structural characteristics of multi-view images. A long time is required to remove the noise from all the views, because the process should be repeated as many times as the number of views. Meanwhile, light field denoising by sparse 5-dimensional (5D) transform domain collaborative filtering (LFBM5D) is a noise removal technique for light field (LF) images captured with a 2D camera array [24]. It performs patch search and grouping operation in both the angular and spatial domains targeting light field images that are densely acquired. The angular domain refers to sub-aperture images, whereas the spatial domain refers to a single image. The denoising performance is improved by securing many similar patches from many sub-aperture images. LFBM5D has also been applied to remove artifacts in super resolution studies of LF images [25]. Furthermore, Chen et al. used a CNN to remove noise from LF images based on an anisotropic parallax analysis (APA) [26]. For real-time LF denoising, 4-dimensional (4D) linear and shift-invariant hyper fan filters have been proposed and implemented in hardware [27]. However, the correlation among the denoised multi-view sequences is not maintained because compression was not considered. Consequently, the results showed significantly decreased compression efficiency.
Table 1. Previous works of denoising algorithms.
Table 1. Previous works of denoising algorithms.
Target ApplicationsDenoising Algorithms
Single viewNLM [8]
BM3D [9,10]
SGN [11]
Real image denoising [12]
Universal denoising networks [13]
FFDNet [14]
VideoVBM3D [15]
VBM5D [16]
With optical flow estimation [17]
Frame-to-frame training [18]
SALT [19]
Multi-viewNLM with PS [20]
With occlusion handling [21]
MVCNN [22]
LFBM5D [24,25]
APA [26]
An epipolar plane image (EPI) is a typical means of expressing the structure of multi-views. In an EPI, the relative motion between cameras or objects is expressed as lines that have different slopes depending on the depth. In comparison with regular photo images, an EPI has a very well-defined structure and there have been many studies of EPI estimation [28,29,30,31,32,33,34,35]. Depth estimation is the typical example in which an EPI is utilized actively. Zhang et al. proposed the EPI-based depth estimation that utilized a spinning parallelogram operator (SPO) for light field camera images with noise and occlusion [30]. Meanwhile Calderon et al. use a two-step EPI estimation for light field images [31]. The initial estimation, conducted using a pixel by pixel approach, is done first. The aggregation process is then performed in order to ensure the same disparity for the same object. In the field of video coding, EPI-based motion estimation (ME) has been widely adopted. Li et al. estimate the dense disparity at a high accuracy by using the structural information obtained from a light field [32]. Lu et al. increase the video coding efficiency by using the epipolar geometry in multi-view images captured from various angles for an object at the center, where the epipolar line of a multi-view image is obtained, after which the disparity vector is acquired by searching over the line [33]. Similarly, other studies [34,35] also propose EPI-based disparity or ME schemes. However, a study to combine image processing such as denoising with an EPI structure has not been seriously attempted thus far.
This paper takes advantage of the structural characteristics of linear multi-view images to increase the denoising speed and performance as well as to improve the compression efficiency of the denoised multi-view sequence. The key contributions of the proposed scheme are as follows. First, the 3D filtering and the multi-view geometry-based EPI estimation are effectively combined to speed up the multi-view denoising. Second, to increase the accuracy of the EPI estimation in the noisy multi-view, the consistence of disparity vector is considered both in the inter- and intra-views. To do this, the search is performed in units of a 1-dimensional (1D) window and the search range boundary is carefully determined in the inter-views. Third, compression-friendly EPI is estimated instead of pursuing the true EPI. This is to reflect the characteristics of the ME of video coding which performs a block-based search and considers the trade-off between bitrate and accuracy (which equals quality). The proposed scheme effectively increases the compression efficiency when the noisy multi-view images must be stored and transmitted.
The remainder of this paper is organized as follows. Section 2 describes the entire flow of the proposed algorithm, which combines EPI and BM3D. Section 3 explains the details of the EPI estimation process and the experimental results are provided in Section 4. Section 5 gives the conclusions.

2. Overview

The flow in Figure 1 shows the BM3D-like denoising process. The left side gives the single-view-oriented techniques of BM3D. In the first search step, the non-local search is carried out to find the patches that are similar to the target patch T. The patches found are grouped. In the second filtering step, the individual groups undergo 3D transformation and hard-thresholding to remove high-frequency signals, assuming that noise is present at the high frequency. Here, the weight is determined according to how much the noise is removed. After the inverse transformation, the filtered patches are aggregated according to weight. This produces the denoised view images.
There are some difficulties when BM3D is directly applied to multi-view images. First, the searching and grouping of similar blocks through block matching take too much time. This processing time increases in proportion to the number of views in multi-view images. Second, 2D block matching based on the sum of absolute difference (SAD) results in low accuracy in a noisy image. Third, view-by-view denoising cannot maintain consistency among the denoised views, resulting in a reduction of the multi-view sequence compression efficiency. VBM3D and VBM4D, based on BM3D, also have similar limitations.
The right side in Figure 1 shows the proposed scheme. In the multi-view images obtained from a linear 1D camera array, corresponding pixels or patches are found at the same height of all views, that is, in the epipolar line. Therefore, the general 2D blocking matching is re-placed with epipolar line searching (ELS) in the unit of a 1D window. In the case of a 1D window (search window), a relatively small range of 11 pixels is used in consideration of complexity and execution time. Given the target pixel, a set of the corresponding pixels from multiple views forms a straight line, which is hereinafter referred to as the corresponding pixel line (linecor-pixel). Grouping is performed in a 2D patch unit along the linecor-pixel and the eight patches are transferred to the filtering step because the computational complexity of the 3D transformation increases sharply according to the number of patches. The subsequent steps are the same as for BM3D. In the last step, the inter-view aggregation is performed on the multiple views.
The key ideas with regard to how the proposed combination of patch searching and 3D transform resolves the difficulties explained in existing research are as follows.
  • Searching time accounts for nearly half of the total time in BM3D-like approaches such as BM3D, VBM3D, VBM4D and LFBM5D. In this paper, both the search area and the difference calculation are changed from 2D to 1D considering the characteristics of 1D linear geometry. Thus, the searching speed is increased.
  • The naïve block-matching search used in the conventional BM3D-like approaches does not work well with noise. For an efficient search in the noisy views, the proposed 1D window-based search using EPI, which accurately reflects the characteristics of multi-view, reduces the risk of incorrect patches. In addition, the maximum disparity is reflected in the ELS range of the initial views (denoted as SR1step hereinafter) as meaningful searching guidance for noisy images. Subsequently, the remaining views limit the search range (denoted by SR2step hereinafter) around the best point found in the previous views to prevent an incorrect outcome due to noise.
  • Most existing denoising algorithms simply consider the SAD in block search process. However, the proposed ELS plays the role of pre-processing for the block-based ME to be carried out for the denoised multi-view compression. When searching for the correspondence in each view, the slope similarity of the linecor-pixel with adjacent pixels as well as the pixel difference is considered together to reflect the rate-distortion optimization of ME. This has a positive effect in the subsequent compression stage of denoised multi-views.
  • Aggregation using weighted summation helps denoised results to be spatially consistent within the image in BM3D-based approaches. Hereinafter, this is referred to as spatial aggregation. In this paper, both spatial and temporal aggregations are adopted. The correlation between the views is increased by performing inter-view level temporal aggregation from the filtering results for the patches of each view. This contributes to increasing the compression efficiency.

3. Proposed Compression-Friendly Denoising

3.1. Fast and Noise-Resistanti EPI Estimation

ELS as proposed in this paper consists of three steps for initial views, remaining views and unmatched pixels.
In the first step, EPI estimation is conducted with the first and second views from the left side of the given multi-views. In Figure 2, v1 and v2 represent the first and second views, respectively, whereas X and Y correspondingly denote the width and height of the views. T1, a pixel of v1, is set as the denoising target pixel. To perform ELS on v2, searching begins with the reference pixel r2 located at coordinates identical to those of T1. In a noisy view, EPI estimation in pixel units with respect to the Y-component may be inaccurate. To take the adjacent pixels into consideration together, ELS is performed in a 1D window unit with a window size of (2w + 1). Considering v2 is located to the right of v1, ELS is performed toward the left of the reference pixel r2 in the range of the SR1step which is set to the maximum disparity that can be varied depending on the distance between the camera and the closest object in multi-views. The pixel c2 with the minimum cost, found in v2, becomes the correspondence of T1. The horizontal difference between T1 and c2 is referred to as the disparity vector (DV), and the line connecting the corresponding pixels, shown in Figure 2, is referred to as the linecor-pixel. Repeating the same process with respect to the X × Y area of v1 gives EPI of the two views.
There are two criteria for judging the correspondence in ELS. First, the difference from the target pixel should be small so as to increase the denoising effect by securing similar patches in the filtering step of Figure 1. Second, the DV should be as similar as the MV which will be obtained from ME in the video compression to be conducted after denoising. Equation (1) briefly shows Jmotion, which is the cost function when the ME searches for the optimally matched block in video compression such as H.264 [36] or HEVC [37]. D(T, R) represents the difference between the current target block, T, and the reference block, R, whereas RMVD indicates the number of bits occurring when coding an MV. This is usually replaced by the MV difference between the current and predicted MV. λmotion represents the Lagrangian multiplier [38,39]. The minimum cost is determined considering both the rate and distortion.
Jmotion = D(T, R) + λmotion·RMVD.
For ME friendly ELS, the cost function in (2), which is designed by modifying (1), is used to search for the correspondence. D(T, R) is the sum of the squared errors between the current window containing a target pixel and the reference window on the epipolar line of the next view. RDVD is the difference between the current and predicted DVs. The predicted DV is obtained from the median DV of the top and left pixels of the current view already searched, whereas the current DV is the disparity vector between the current and reference windows. In many existing block matching-based noise denoising algorithms, only the difference between blocks is used as a metric to determine the block similarity. This is because filtering on similar blocks greatly influences the denoising performance. However, this paper noted that in ME of video coding, consistent MV along with the similarity of pixel values is an important criterion for block matching. Therefore, RDVD is added in (2) to reflect such spatial correlation of video coding to block matching of denoising. By considering the pixel difference and RDVD together, it is possible to achieve not only denoising performance but also compression efficiency in the multi-view compression due to enhanced spatial correlation. Also, a point with a small RDVD value is preferred to follow the condition in ME which considers the rate-distortion trade-off as shown in (1).
cost = D(T, R) + λ·RDVD.
In the second step of ELS, the EPI of all remaining views, starting with v3, is obtained. In Figure 3, the specific epipolar lines of the four views (v1, v2, v3 and v4) are stacked. Each square denotes a pixel. The gray and white pixels represent the background and the foreground in the view, respectively. The box with the thick line represents the 1D window with w = 3 to be used in searching. In the first step shown in Figure 2, the correspondence c2 has already been found in v2 for the target pixel T1 of v1. If d1_2, the DV of c2, is used as the predicted DV, the initial reference pixel r3 for ELS in v3 is obtained by subtracting d1_2 from the x coordinate of c2, which is c2(x). The cost is calculated for the range of SR2step pixels indicated by the dotted lines containing r3. Out of these, the pixel having the lowest cost is determined using (2) to be the correspondence c3. For ELS of v4, the DV d2_3 is used as the predicted DV. The initial reference pixel r4 is calculated from c3(x) − d2_3. By using the DV in the directly prior view as the predicted DV, setting SR2step such that it is very small is reasonable. Thus, the search range of the second step can be significantly reduced compared to the range of SR1step in the first step. The same process is carried out for all views except v1 and v2.
After the first and second steps, the drawing linecor-pixel on all views given the target pixels of v1 is complete. However, in other views except v1, there may be non-corresponding (nc) pixels which do not belong to any group. If these nc pixels are excluded from filtering, the denoising performance will decrease. To solve this, nc pixels are set as new target pixels. Then, ELS is performed for them. Since this process is done to include nc pixels in 8 × 8 patch-based filtering, the target pixels are determined by sampling rather than using all nc pixels. ELS is performed for the views vt−1 and view vt+1, which are respectively the previous view and the next view of the vt to which the target pixel belongs. Here, the predicted DV is obtained by referring to the linecor-pixel that is closest in the horizontally left direction of the target pixel in vt. Based on the predicted DV, initial reference pixels are determined in vt−1 and vt+1, and searching is performed in the adjacent range of SR2step. The ELS processes are conducted toward the right and left from the reference pixels in vt−1 and in vt+1, respectively.

3.2. Temporal Aggregation

In this paper, temporal aggregation is additionally applied along with spatial aggregation to improve the correlation between multi-views further. Spatial aggregation is the same as the existing BM3D-like approach. In a single view, denoised patches are returned to their position through weighted summation reflecting the degree of noise removal. In temporal aggregation, a similar operation is performed at the inter-view level. The application of temporal aggregation is determined by means of a binary confidence judgment. This binary confidence is based on the cost in (2) used in the EPI estimation process. A low cost indicates that the difference and disparity variance of the patches on the linecor-pixel are sufficiently small, meaning that the confidence of the estimated EPI is high. Therefore, when the cost value is lower than the threshold, temporal aggregation is performed. The threshold is set empirically. Temporal aggregation is performed between views by reflecting weights according to denoising results of patches from each view, contributing to the improvement of the compression efficiency.

4. Performance Evaluation

The proposed scheme in this paper is evaluated in terms of denoise performance, compression efficiency and computational complexity. Figure 4 shows ten datasets used in the experiment. Champagne, Pantomime and Dog have 80 views, whereas Wall and Piano have 96 and 64 views, respectively. Kendo and Balloons consist of four views [40,41]. Ohta, Sideboard and Cotton datasets are additionally used to test the performance when the proposed scheme is applied to not only 1D but also 2D multi-view. Ohta, Sideboard and Cotton have 25 and 81 views, respectively [42]. Noise is added to each test sequence using the additive white gaussian noise (AWGN) of sigma values of 15, 25 and 35. For HEVC encoding, pseudo sequences are created from multi-view images and HM-16.10 reference software [43] with the low-delay-P-Main (LD) configuration. Quantization parameters (QPs) are set to 22, 27, 32 and 37. All the experiments are conducted on an Intel i7-8700K running 3.70 GHz and 64 GB memory. The superiority of the proposed scheme is compared with BM3D [9], VBM4D [16], MS-EPLL [44], 3D-focus [21], MVCNN [22] and LFBM5D [25]. The codes used for the proposed scheme and for performance comparison are uploaded to https://github.com/Digital-System-Design-Lab/Compression-Aware-Multi-view-Denoising.git (accessed on 21 April 2021) [45]. In the configuration of BM3D, VBM4D and LFBM5D, only hard thresholding is applied without Wiener filtering steps. The patch size is 8 × 8 pixels. The maximum number of patches are set to 8 for Champagne, Pantomime, Dog, Wall and Piano, and 4 for Balloons, Kendo, Ohta, Sideboard and Cotton. The search area is 32 pixels horizontally and vertically with a one-pixel spacing distance. Spacing is a major factor that determines the trade-off between processing time and denoising performance. Commonly, spacing = 3 is used to reduce complexity, whereas spacing = 1 is used when denoising performance is important. The settings used in this experiment follow the previous BM3D-like approaches.
Table 2 shows the denoise performance when using the conventional denoising algorithms and the proposed scheme. Only Champagne, Pantomime, Dog, Wall, Piano, Balloons and Kendo, which are horizontally linear multi-view datasets acquired with a 1D camera array, were tested. From the third to ninth columns, the quality of the denoised multi-view images was compared to the original clean images using the peak-signal-to-noise ratio (PSNR). Given a noise-free m × n monochrome image I and its noisy approximation K, MSE is defined as follows:
MSE = 1 m   n i = 0 m 1 j = 0 n 1 [ I ( i , j ) K ( i , j ) ] 2
The PSNR can be calculated as follows:
PSNR = 10 · log 10 (   MAX I 2 MSE   ) .
Among seven algorithms, MS-EPLL and the proposed scheme show the highest PSNR when sigma = 15, whereas the MS-EPLL is best when sigma = 25 and 35. VBM4D is about 0.85 dB lower than BM3D even though more views are used for denoising. In block matching of VBM4D, simplified searching decreases the denoising performance because there is a risk of finding wrong patches in a complex image. MS-EPLL shows higher PSNR than BM3D and proposed scheme in most test sequences. However, blur due to filtering is noticeable. Therefore, its PSNR is lower than the proposed scheme for the small sigma or for the high detail views such as Piano and Dogs. Using 3D-focus is basically an algorithm that assumes a light field camera to shoot densely. Thus, its performance is good in multi-view, which is very tightly shot such as Piano. However, in the general multi-view with a wide camera distance, disparity maps are not well estimated, resulting in poor noise reduction. MVCNN also suffers from similar problems with 3D-focus. As an incorrect disparity map is generated, a strong blur occurs. Since LFBM5D does not fully utilize horizontally linear structural information, the accuracy is somewhat low when searching for similar patches in noisy views. This problem is more pronounced in datasets such as Balloons and Champagne with wider view distance. As the sigma of AWGN is larger, blur appears at the edge and serious artifacts such as color contamination occur due to grouping of wrong patches. As shown in Table 2, it was confirmed that the PSNR is lower than that of the proposed scheme which fully utilizes a multi-view structure. From the tenth to fifteenth columns of Table 2, the compression efficiency of the denoised multi-view images is measured with the Bjontegaard Delta Bitrate (BDBR) [46], which is calculated using the bitrate used during the transmission and the PSNR between already denoised input views in the encoder side and the reconstructed views in the decoder side. The BDBR in Table 2 show the reduced bitrate of various denoising schemes compared to BM3D. In Pantomime, having a wide range of black background and little object movement, both VBM3D and the proposed scheme which consider the correlation between views, achieve a bitrate reduction compared to BM3D. However, in other sequences, the proposed scheme shows the highest compression efficiency. For complex and noisy sequences, the BDBR performance of the proposed scheme is much improved. In VBM4D, it is difficult to denoise the multi-view with high details or wide camera distance. Thus, the compression efficiency is also poor. In MS-EPLL, as the sigma value increases, the compression efficiency increases significantly. Although temporal correlation among views is not considered in MS-EPLL, the increased similarity between views from the severe blur can help video coding. In the case of 3D-focus, the compression efficiency is not good because the noise is not removed well and the color is distorted. In the case of MVCNN, the denoising performance is greatly improved compared to 3D-focus. The compression performance is also high due to the blurred result. The LFBM5D’s BDBR performance seems to be good, but it is due to blur caused by wrong patch grouping. This appears stronger as the sigma value increases.
Table 3 shows the multi-view denoising performance as acquired from a 2D camera array in terms of the objective image quality and compression efficiency. This experiment shows that the proposed scheme can easily be extended to a 2D light field by estimating EPI horizontally as well as vertically. The PSNR value of the objective image quality measures the difference between the clean original multi-view and the denoised multi-view with each algorithm. In the case of BDBR, a multi-view denoised image with BM3D was used as a reference. Among a number of denoising algorithms for 2D multi-view, the performance of the proposed scheme was verified through a comparison with LFBM5D, which shows excellent denoising results. The 2D multi-view was converted to a spiral pseudo sequence and input to the HEVC encoder. The Ohta, Sideboard and Cotton datasets were used. Ohta has 5 × 5 = 25 views, whereas, in Ohta2, the views were down sampled to 3 × 3 = 9 views so as to assume an environment with a wide view distance. Sideboard and Cotton used 9 × 9 = 81 views. AWGNs with sigma = 15, 25 and 35 were used. The average PSNR of LFBM5D is 29.46 dB, whereas that of the proposed is 30.97 dB, which is approximately 1.51 dB higher compared to LFBM5D. In the case of objective image quality, it was verified that the proposed scheme outperforms LFBM5D in all datasets and under all sigma conditions. In the case of BDBR, the average BDBR of LFBM5D was −18.41%, whereas that of the proposed is −33.76%. LFBM5D shows high compression efficiency in datasets with narrow view distances, such as Ohta, Sideboard and Cotton in this case, but not in datasets with wide view distances, such as Ohta2. When the proposed scheme was applied, the BDBR performance was excellent for all datasets regardless of the distance between the views.
For a fair comparison with a deep learning based approach optimized for low resolution images, Table 3 and Table 4 show the PSNR and BDBR of denoised multi-views in four algorithms of BM3D, MS-EPLL, MVCNN and the proposed scheme when the resolution of seven datasets is reduced to 1/4. In the patch based denoising algorithm, the lower resolution increases the spatial complexity of the image patches, which can degrade the denoising performance. In Table 4, the average PSNR of each algorithm is 32.92 dB for BM3D, 33.36 dB for MS-EPLL, 22.25 dB for MVCNN and 33.18 dB for the proposed scheme. In Table 5, compared to BM3D, the average BDBR for each algorithm is +6.83% for MS-EPLL, −79.54% for MVCNN, and −25.78% for the proposed scheme. MVCNN has good compression efficiency due to excessive blur, but its PSNR value is very low. MS-EPLL even increases the BDBR value at low sigma value. As shown in Table 3 and Table 4, the proposed scheme is superior to other algorithms when considering both denoising performance and compression efficiency even in low resolution images.
Figure 5 and Figure 6 show the subjective quality of denoised multi-view densely acquired with a 1D camera array. After adding noise with sigma = 25 to Kendo and Piano, various denoising algorithms were applied. In the result of BM3D and the proposed scheme, the edge is clear, and the noise is removed quite well. In VBM4D, many noises are not removed properly. The probability of grouping the wrong patches is high because patch searching is done in the vertical direction. In the MS-EPLL, blurs are observed strongly in the human face in Figure 5 and the brick wall in Figure 6. Many details disappear. Inaccurate disparity estimation of 3D-focus not only degrades denoising performance but also causes color distortion. In the multi-view with a large camera distance such as Kendo in Figure 5, the disparity map of MVCNN is inaccurate and thus, blur is observed. In the case of Piano in Figure 6 which consists of dense views, the quality looks relatively clean, but it also has a lot of blurred edges.
Figure 7 shows the subjective quality of Ohta’s denoised multi-view. Here, Figure 7a shows a clean view without added noise, whereas Figure 7b is the result of denoising with LFBM5D when the AWGNs with sigma = 15, 25, and 35 are added. Additionally, Figure 7c results when the proposed scheme is applied. Since LFBM5D did not find similar patches well, noise that could not be removed and broken edges were observed. These artifacts are more severe as the sigma value becomes higher and the distance between the view’s increases. Meanwhile, areas with small disparity values, such as background areas, are well denoised, but strong blurring is noticeable. Consequently, the PSNR performance is poor, but it is advantageous to increase the compression efficiency. In the case of what was proposed, because similar patches were well searched, the denoising performance was quite good with clear edges.
Table 6 shows improved BDBR results when both spatial and temporal aggregation are used. The third column shows the results when only spatial aggregation is applied, whereas the fourth column presents the results when both spatial and temporal aggregations are applied together, which is identical to the fifteenth column in Table 2. When temporal aggregation is added, it can be seen that the compression efficiency is further improved by −10.54% in BDBR.
Table 7 shows the execution time of the tested codes of each algorithm. For BM3D, two code types, each implemented in C++ [9] and Matlab [10], were used. The implementation of the two codes differs slightly, but the denoising performance of the C++ version is much better than that of the Matlab version. VBM4D [22], MS-EPLL [44], 3D-focus and MVCNN were implemented in Matlab, and LFBM5D and the proposed scheme were implemented in C++. In this experiment, the average execution time of the Matlab version of BM3D was set to 1.00 when processing for images with resolutions of 1024 × 768 and 1280 × 960, and the execution times of different algorithms were compared relatively. VBM4D, MS-EPLL, 3D-focus and MVCNN require 3.59 times, 1618.90 times, 127.33 times and 29.33 times longer execution compared to the execution times of the Matlab version of BM3D. Meanwhile, BM3D’s C++ version is 9.00 times longer, LFBM5D is 35.27 times longer, and the proposed scheme is 1.80 times longer. The proposed scheme achieves high denoising performance and compression efficiency while having a much faster execution speed than most other algorithms.
Table 8 shows the results of a comparison of the computational complexity required for denoising. First, for the block searching, which is the most time-consuming step during the denoising process, the second column in Table 8 is obtained from the denoising of the Champagne sequence and indicates the number of subtraction operations per view when 80 views 1280 × 960 in size are processed. For the block-matching in BM3D, the number of subtractions is calculated as h × w × sp2 × s2, where h and w denote the height and width, respectively. sp2 represents the patch size, whereas s2 represents the search area. For the block-matching in VBM4D, the number of subtractions comes to h × w × ht × sp2 × s. Here, ht is the number of patches found in other views considering the temporal correlation in VBM4D. 3D-focus selects the best five windows and additionally searches for similar windows around them. The required computations are h × w × (5 × sp2 + s2). For the block-matching in LFBM5D, the number of subtractions is calculated as h × w × sp2 × nsim. nsim is a 2D search window to find similar patches within a single view. In the proposed EPI-based search, the total is h × w × (sw × s1 + sw × s2). Here, sw denotes the 1D window size, whereas s1 and s2 represent the search range in the first and second steps, respectively. In summary, VBM4D and the proposed scheme show the highest and the lowest level of search complexity. Based on the 1K resolution image, computations for searching are calculated. CNN-based denoising is excluded. The amounts of computation for searching when using BM3D, VBM4D, 3D-focus and LFBM5D are approximately 1677, 15,859, 34 and 131 times larger, respectively, than that in the proposed scheme. In the third column, the filtering complexity is compared using the number of patches because the filtering time is proportional to the number of patches. The number of patches required per view is lowest with VBM4D and highest with 3D-focus.

5. Conclusions

Until recently, many works to remove noise in images have been actively conducted, but the structural characteristics of multi-views cannot be effectively utilized. The proposed scheme in this paper improves the consistency of denoised multi-view as well as the performance of noise reduction by combining an excellent denoising algorithm BM3D with EPI reflecting the correlation of multi-views acquired from camera array. Experimental results also show that improved consistency has a significant positive effect on compression efficiency. In the future, additional intensive study is required through algorithmic and architectural co-development for fast processing of massive multi-view.

Author Contributions

Investigation, C.-E.R.; Supervision, C.-E.R.; Validation, C.-E.R.; Writing—original draft, Y.-N.J. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was result of the research project supported by the R&D program of MOTIE/KEIT [No. 10077609, Developing Processor-Memory-Storage Integrated Architecture for Low Power, High Performance Big Data Servers] and also supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT & Future Planning (NRF-2021R1A2C2008946).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The codes used for the proposed scheme and for performance comparison are uploaded to https://github.com/Digital-System-Design-Lab/Compression-Aware-Multi-view-Denoising.git (accessed on 21 April 2021).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Xiong, Z.; Cheng, Z.; Peng, J.; Fan, H.; Liu, D.; Wu, F. Light field super-resolution using internal and external similarities. In Proceedings of the 2017 IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 1612–1616. [Google Scholar]
  2. Xiao, Z.; Si, L.; Zhou, G. Seeing beyond foreground occlusion: A joint framework for SAP based scene depth and appearance reconstruction. IEEE J. Sel. Top. Signal Process. 2017, 11, 979–991. [Google Scholar] [CrossRef]
  3. Wang, Y.; Yang, J.; Guo, Y.; Xiao, C.; An, W. Selective light field refocusing for camera arrays using bokeh rendering and superresolution. IEEE Signal Process. Lett. 2019, 26, 204–208. [Google Scholar] [CrossRef]
  4. Wang, Y.; Wang, L.; Yang, J.; An, W.; Yu, J.; Guo, Y. Spatial-angular interaction for light field image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 November 2020; pp. 290–308. [Google Scholar]
  5. Wang, Y.; Wu, T.; Yang, J.; Wang, L.; An, W.; Guo, Y. DeOccNet: Learning to See Through Foreground Occlusions in Light Fields. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 118–127. [Google Scholar]
  6. Jung, H.; Lee, H.-J.; Rhee, C.E. Ray-space360: An extension of ray-space for omnidirectional free viewpoint. In Proceedings of the IEEE International Symposium on Circuits and Systems, Florence, Italy, 27–30 May 2018; pp. 1–5. [Google Scholar]
  7. Jung, H.; Lee, H.-J.; Rhee, C.E. Flexibly Connectable Light Field System for Free View Exploration. IEEE Trans. Multimed. 2017, 22, 980–991. [Google Scholar] [CrossRef]
  8. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 60–65. [Google Scholar]
  9. Lebrun, M. An Analysis and Implementation of the BM3D Image Denoising Method. Image Process. Line 2012, 2, 175–213. [Google Scholar] [CrossRef] [Green Version]
  10. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image denoising by sparse 3d transform-domain collaborative filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  11. Gu, S.; Li, Y.; Van Gool, L.; Timofte, R. Self-guided network for fast image denoising. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 2511–2520. [Google Scholar]
  12. Anwar, S.; Barnes, N. Real image denoising with feature attention. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3155–3164. [Google Scholar]
  13. Lefkimmiatis, S. Universal denoising networks: A novel CNN architecture for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 3204–3213. [Google Scholar]
  14. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a fast and flexible solution for CNN-based image denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [Green Version]
  15. Dabov, K.; Foi, A.; Egiazarian, K. Video denoising by sparse 3d transform-domain collaborative filtering. In Proceedings of the European Signal Processing Conference, Poznan, Poland, 2–7 September 2007; pp. 145–149. [Google Scholar]
  16. Maggioni, M.; Boracchi, G.; Foi, A.; Egiazarian, K. Video denoising using separable 4-D nonlocal spatiotemporal transforms. SPIE Electron. Imaging 2011, 7870, 787003-1–787003-11. [Google Scholar]
  17. Buades, A.; Lisani, J.-L.; Miladinović, M. Patch-based video denoising with optical flow estimation. IEEE Trans. Image Process. 2016, 25, 2573–2586. [Google Scholar] [CrossRef] [PubMed]
  18. Ehret, T.; Davy, A.; Morel, J.-M.; Facciolo, G.; Arias, P. Model-blind video denoising via Frame-To-Frame training. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 11369–11378. [Google Scholar]
  19. Wen, B.; Li, Y.; Pfister, L.; Bresler, Y. Joint adaptive sparsity and low-rankness on the fly: An online tensor reconstruction scheme for video denoising. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 241–250. [Google Scholar]
  20. Miyata, M.; Kodama, K.; Hamamoto, T. Fast multiple-view denoising based on image reconstruction by plane sweeping. In Proceedings of the IEEE Visual Communications and Image Processing, Valletta, Malta, 7–10 December 2014; pp. 462–465. [Google Scholar]
  21. Zhou, S.; Hu, Y.H.; Jiang, H. Patch-based multiple view image denoising with occlusion handling. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, New Orleans, LA, USA, 5–9 March 2017; pp. 1782–1786. [Google Scholar]
  22. Zhou, S.; Hu, Y.H.; Jiang, H. Multi-View Image Denoising Using Convolutional Neural Network. Sensors 2019, 19, 2597. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual learning of deep CNN for image denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [Green Version]
  24. Alain, M.; Smolic, A. Light field denoising by sparse 5D transform domain collaborative filtering. In Proceedings of the IEEE 19th International Workshop on Multimedia Signal Processing, Luton, UK, 16–18 October 2017; pp. 1–6. [Google Scholar]
  25. Alain, M.; Smolic, A. Light field super-resolution via LFBM5D sparse coding. In Proceedings of the IEEE International Conference on Image Processing, Athens, Greece, 7–10 October 2018; pp. 2501–2505. [Google Scholar]
  26. Chen, J.; Hou, J.; Chau, L.-P. Light field denoising via anisotropic parallax analysis in a CNN framework. IEEE Signal Process. Lett. 2018, 25, 1403–1407. [Google Scholar] [CrossRef] [Green Version]
  27. Premaratne, S.U.; Liyanage, N.; Edussooriya, C.U.S.; Wijenayake, C. Real-time light field denoising using a novel linear 4-D hyperfan filter. IEEE Trans. Circuits Syst. I 2020, 67, 2693–2706. [Google Scholar] [CrossRef]
  28. Wu, G.; Liu, Y.; Dai, Q.; Chai, T. Learning sheared EPI structure for light field reconstruction. IEEE Trans. Image Process. 2019, 28, 3261–3273. [Google Scholar] [CrossRef]
  29. Shin, C.; Jeon, H.-G.; Yoon, Y.; Kweon, I.S.; Kim, S.J. EPINET: A fully-convolutional neural network using epipolar geometry for depth from light field images. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 4748–4757. [Google Scholar]
  30. Zhang, S.; Sheng, H.; Li, C.; Zhang, J.; Xiong, Z. Robust depth estimation for light field via spinning parallelogram operator. Comput. Vision Image Underst. 2016, 145, 148–159. [Google Scholar] [CrossRef]
  31. Calderon, F.C.; Parra, C.A.; Niño, C.L. Depth map estimation in light fields using an stereo-like taxonomy. In Proceedings of the IEEE Symposium on Image, Signal Processing and Artificial Vision, Armenia, Colombia, 17–19 September 2014; pp. 1–5. [Google Scholar]
  32. Li, J.; Lu, M.; Li, Z.-N. Continuous depth map reconstruction from light fields. IEEE Trans. Image Process. 2015, 24, 3257–3265. [Google Scholar] [PubMed] [Green Version]
  33. Lu, J.; Cai, H.; Lou, J.-G.; Li, J. An epipolar geometry-based fast disparity estimation algorithm for multiview image and video coding. IEEE Trans. Circuits Syst. Video Technol. 2007, 17, 737–750. [Google Scholar] [CrossRef]
  34. Micallef, B.W.; Debono, C.J.; Farrugia, R.A. Fast disparity estimation for multi-view video plus depth coding. In Proceedings of the Visual Communications and Image Processing, Tainan, Taiwan, 6–9 November 2011; pp. 1–4. [Google Scholar]
  35. Li, X.M.; Zhao, D.B.; Ma, S.W.; Gao, W. Fast disparity and motion estimation based on correlations for multiview video coding. IEEE Trans. Consum. Electron. 2008, 54, 2037–2044. [Google Scholar] [CrossRef] [Green Version]
  36. Wiegand, T.; Sullivan, G.J.; Bjontegaard, G.; Luthra, A. Overview of the H.264/AVC video coding standard. IEEE Trans. Circuits Syst. Video Technol. 2003, 13, 560–576. [Google Scholar] [CrossRef] [Green Version]
  37. Sullivan, G.J.; Ohm, J.-R.; Han, W.-J.; Wiegand, T. Overview of the high efficiency video coding (HEVC) standard. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1648–1667. [Google Scholar] [CrossRef]
  38. Sullivan, G.J.; Wiegand, T. Rate-distortion optimization for video compression. IEEE Signal Process. Mag. 1998, 15, 74–90. [Google Scholar] [CrossRef] [Green Version]
  39. Li, X.; Oertel, N.; Hutter, A.; Kaup, A. Advanced Lagrange multiplier selection for hybrid video coding. In Proceedings of the 2007 IEEE International Conference on Multimedia and Expo (ICME), Beijing, China, 2–5 July 2007; pp. 364–367. [Google Scholar]
  40. Nagoya University Multi-View Sequences. Available online: http://www.fujii.nuee.nagoya-u.ac.jp/multiview-data/ (accessed on 24 January 2018).
  41. Multi-View Dataset. Available online: http://sydlab.net/wordpress/wp-content/uploads/2019/10/dataset.zip (accessed on 21 April 2021).
  42. Middlebury Stereo Vision Page. Available online: http://vision.middlebury.edu/stereo/data/ (accessed on 21 June 2017).
  43. Joint Collaborative Team on Video Coding (JCT-VC). High Efficiency Video Coding Reference Software. High Efficiency Video Coding Test Model 16.10 (HM16.10.). Available online: http://hevc.hhi.fraunhofer.de/ (accessed on 13 September 2018).
  44. Papyan, V.; Elad, M. Multi-scale patch-based image restoration. IEEE Trans. Image Process. 2016, 25, 249–261. [Google Scholar] [CrossRef] [PubMed]
  45. Source Codes Used in the Experiments of this PAPER. Available online: https://github.com/Digital-System-Design-Lab/Compression-Aware-Multi-view-Denoising.git (accessed on 21 April 2021).
  46. Bjøntegaard, G. Calculation of Average PSNR Differences between RD-Curves. In Proceedings of the Report VCEG-M33, ITU-T SG16/Q6, 13th Video Coding Experts Group (VCEG) Meeting, Austin, TX, USA, 2–4 April 2001; Available online: https://github.com/tbr/bjontegaard_etro (accessed on 21 April 2021).
Figure 1. Flow chart of the proposed scheme.
Figure 1. Flow chart of the proposed scheme.
Sensors 21 02919 g001
Figure 2. Initial EPI estimation to obtain the predicted slope of linecor-pixel.
Figure 2. Initial EPI estimation to obtain the predicted slope of linecor-pixel.
Sensors 21 02919 g002
Figure 3. Fast search based on predicted slope.
Figure 3. Fast search based on predicted slope.
Sensors 21 02919 g003
Figure 4. Dataset. (a) Champagne; (b) Pantomime; (c) Dog; (d) Wall; (e) Piano; (f) Kendo; (g) Balloons; (h) Ohta; (i) Sideboard; (j) Cotton.
Figure 4. Dataset. (a) Champagne; (b) Pantomime; (c) Dog; (d) Wall; (e) Piano; (f) Kendo; (g) Balloons; (h) Ohta; (i) Sideboard; (j) Cotton.
Sensors 21 02919 g004
Figure 5. Subjective quality of kendo. (a) Full image of original; (b) original; (c) BM3D; (d) VBM4D; (e) MSEPLL; (f) 3D focus; (g) MVCNN; (h) proposed.
Figure 5. Subjective quality of kendo. (a) Full image of original; (b) original; (c) BM3D; (d) VBM4D; (e) MSEPLL; (f) 3D focus; (g) MVCNN; (h) proposed.
Sensors 21 02919 g005
Figure 6. Subjective quality of Piano. (a) Full image of original; (b) original; (c) BM3D; (d) VBM4D; (e) MSEPLL; (f) 3D focus; (g) MVCNN; (h) LFBM5D; (i) proposed.
Figure 6. Subjective quality of Piano. (a) Full image of original; (b) original; (c) BM3D; (d) VBM4D; (e) MSEPLL; (f) 3D focus; (g) MVCNN; (h) LFBM5D; (i) proposed.
Sensors 21 02919 g006aSensors 21 02919 g006b
Figure 7. Subjective quality of Ohta. (a) Original image; (b) LFBM5D when sigma = 15, 25 and 35; (c) proposed when sigma = 15, 25 and 35.
Figure 7. Subjective quality of Ohta. (a) Original image; (b) LFBM5D when sigma = 15, 25 and 35; (c) proposed when sigma = 15, 25 and 35.
Sensors 21 02919 g007
Table 2. Comparison of performance for noisy multi-views.
Table 2. Comparison of performance for noisy multi-views.
SequenceσPSNR (dB)BDBR (%)
BM3DVBM4DMS-EPLL3D-FocusMVCNNLFBM5DProposedVBM4DMS-EPLL3D-FocusMVCNNLFBM5DProposed
Champagne
(1280 × 960)
1538.0737.4038.0625.9133.3634.4336.08−9.22+6.00−22.47−43.07+9.00−14.97
2535.2334.7235.7825.0631.0830.9033.93−6.58−6.73−2.75−51.97+17.81−20.98
3533.3332.5734.2023.7928.1229.3232.18−0.88−26.92+12.83−55.81+9.94−25.32
Pantomime
(1280 × 960)
1538.9538.9338.9734.0332.9036.6637.88−51.43+15.25+36.92−92.33−49.85−48.01
2536.2036.4636.8129.4531.7232.4335.39−44.80−7.90+26.37−62.47−64.46−58.29
3534.4634.3235.2722.7430.4531.0233.23−54.02−48.38+62.43−68.22−77.88−51.64
Dog
(1280 × 960)
1536.0535.2535.6633.3430.6133.6435.39−19.83+29.80−33.55−61.76−37.74−52.88
2533.1832.8433.4530.0530.2830.5033.11−1.63+4.37+16.44−64.24−57.75−60.86
3531.1330.9131.9226.0929.9328.3231.35+47.82−25.89+13.24−69.20−61.28−60.77
Wall
(1024 × 768)
1536.9035.9937.8433.7432.3837.7138.15−8.08+1.03+189.44−27.04−55.88−48.61
2533.8133.4235.4731.6032.0734.6035.38+19.00−36.03+406.64−33.56−75.27−65.42
3531.6131.3133.8329.8431.7632.4433.06+148.10−62.21+167.57−41.74−83.41−76.05
Piano
(1024 × 768)
1537.4235.8937.2935.6031.1737.8838.20−13.02+35.02+522.99−11.58−50.05−39.49
2534.4633.1434.9032.3830.8734.2735.35+10.33−4.02+131.63−19.55−72.39−55.59
3532.3631.0133.1530.2430.5631.8633.07+130.98−41.80+127.47−34.40−79.18−65.27
Kendo
(1024 × 768)
1537.7335.8838.1433.4731.2336.2836.71+5.70−23.47+1891.84−37.20−39.03−28.46
2534.8233.1336.2130.1430.9533.1134.68+47.63−43.35+2218.42−46.27−41.51−39.68
3532.6930.9734.8127.8330.6630.7133.05+162.93−58.75+1902.10−57.39−41.56−47.12
Balloons
(1024 × 768)
1536.8434.5636.9132.1231.1334.8636.88+3.10−14.68+1071.66−37.88−21.85−30.60
2533.9631.9234.7728.9730.5431.7334.03+37.70−30.12+1220.68−45.56−30.30−42.88
3531.9229.8633.2726.8030.1428.4232.30+100.48−47.50+1060.70−54.66−35.73−49.21
Table 3. Comparison of performance for noisy 2D multi-views.
Table 3. Comparison of performance for noisy 2D multi-views.
2D
Sequence
σPSNR (dB)BDBR (%)
LFBM5DProposedLFBM5DProposed
Ohta
(384 × 288)
1533.5834.17−40.00−25.93
2529.7931.21−46.15−30.24
3527.5229.31−43.21−31.39
Ohta2
(384 × 288)
1529.7133.64+40.45−10.75
2526.4530.99+84.44−11.18
3524.4029.21+129.02−10.14
Sideboard
(512 × 512)
1534.0833.92−48.22−20.90
2530.9130.63−60.90−36.17
3528.7727.15−69.21−40.67
Table 4. Comparison of denoising performance for noisy multi-views (low resolution).
Table 4. Comparison of denoising performance for noisy multi-views (low resolution).
SequenceσPSNR (dB)
BM3DMS-EPLLMVCNNProposed
Champagne
(640 × 480)
1536.3336.0219.5935.83
2533.7033.3718.9733.36
3531.5231.6218.1931.37
Pantomime
(640 × 480)
1537.3537.2224.0037.20
2534.7334.6322.7834.51
3532.4032.8821.2532.27
Dog
(640 × 480)
1534.2433.4123.4834.44
2531.8130.8422.1531.85
3529.9929.1020.8429.89
Wall
(512 × 384)
1537.3436.3024.5737.76
2534.5833.7123.0434.47
3532.5732.0421.3732.11
Piano
(512 × 384)
1536.7535.4324.3637.98
2533.8832.7022.7734.66
3531.8530.9521.0932.34
Kendo
(512 × 384)
1534.2936.4025.3633.74
2530.9334.0123.6630.92
3529.5932.3721.9229.63
Balloons
(512 × 384)
1532.9434.7924.1632.72
2531.2032.2222.6131.27
3528.4430.5121.0429.70
Cotton
(512 × 512)
1530.3030.77−47.96−47.01
2529.4830.56−58.82−63.30
3528.5730.02−60.34−77.45
Table 5. Comparison of compression efficiency for noisy multi-views (low resolution).
Table 5. Comparison of compression efficiency for noisy multi-views (low resolution).
SequenceσBDBR (%)
MS-EPLLMVCNNProposed
Champagne
(640 × 480)
15+15.94−79.30−8.97
25+7.31−84.80−15.52
35−13.03−88.78−20.52
Pantomime
(640 × 480)
15+36.13−69.17−21.72
25+19.17−81.63−28.50
35−16.65−89.32−33.21
Dog
(640 × 480)
15+80.70−85.09−28.40
25+50.12−91.92−35.59
35+0.67−94.77−35.37
Wall
(512 × 384)
15+43.53−63.18−21.32
25+7.02−80.57−30.45
35−30.01−89.55−29.00
Piano
(512 × 384)
15+67.55−65.07−36.21
25+18.96−83.73−48.27
35−29.12−91.97−47.86
Kendo
(512 × 384)
15+3.15−58.87−22.19
25−40.66−80.85−21.46
35−37.12−70.09−10.17
Balloons
(512 × 384)
15+14.29−60.74−8.91
25−7.93−74.25−16.88
35−46.64−86.79−20.83
Table 6. Comparison of performance for noisy multi-views when using both spatial and temporal aggregation.
Table 6. Comparison of performance for noisy multi-views when using both spatial and temporal aggregation.
SequenceσBDBR (%)
Only SpatialSpatial + Temporal
Champagne
(1280 × 960)
15−9.23−14.97
25−19.32−20.98
35−25.38−25.32
Pantomime
(1280 × 960)
15−30.38−48.01
25−46.67−58.29
35−55.87−51.64
Dog
(1280 × 960)
15−35.00−52.88
25−16.43−60.86
35−49.92−60.77
Wall
(1024 × 768)
15−25.42−48.61
25−40.50−65.42
35−46.90−76.05
Piano
(1024 × 768)
15−37.10−39.49
25−57.28−55.59
35−66.83−65.27
Kendo
(1024 × 768)
15−22.93−28.46
25−34.98−39.68
35−39.48−47.12
Balloons
(1024 × 768)
15−20.55−30.60
25−31.83−42.88
35−38.02−49.21
Table 7. Comparison of processing time.
Table 7. Comparison of processing time.
MatlabC++
BM3D1.009.00
VBM4D3.59-
MS-EPLL1618.90-
3D-focus127.33-
MVCNN29.33-
LFBM5D-35.27
Proposed-1.80
Table 8. Comparison of processing time.
Table 8. Comparison of processing time.
SearchingNumber of Patches
BM3D100.60 × 1078.49 × 106
VBM4D951.58 × 1072.21 × 106
MS-EPLL--
3D-focus2.06 × 10798.30 × 106
MVCNN--
LFBM5D7.86 × 10710.78 × 106
Proposed0.06 × 1074.34 × 106
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jin, Y.-N.; Rhee, C.-E. Improved Light Field Compression Efficiency through BM3D-Based Denoising Using Inter-View Correlation. Sensors 2021, 21, 2919. https://doi.org/10.3390/s21092919

AMA Style

Jin Y-N, Rhee C-E. Improved Light Field Compression Efficiency through BM3D-Based Denoising Using Inter-View Correlation. Sensors. 2021; 21(9):2919. https://doi.org/10.3390/s21092919

Chicago/Turabian Style

Jin, You-Na, and Chae-Eun Rhee. 2021. "Improved Light Field Compression Efficiency through BM3D-Based Denoising Using Inter-View Correlation" Sensors 21, no. 9: 2919. https://doi.org/10.3390/s21092919

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop