Next Article in Journal
Carbonized Cotton Fabric-Based Flexible Capacitive Pressure Sensor Using a Porous Dielectric Layer with Tilted Air Gaps
Next Article in Special Issue
Data Clustering Using Moth-Flame Optimization Algorithm
Previous Article in Journal
Investigation of Microwave Ablation Process in Sweet Potatoes as Substitute Liver
Previous Article in Special Issue
Adversarial Gaussian Denoiser for Multiple-Level Image Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Denoising and Motion Artifact Removal Using Deformable Kernel Prediction Neural Network for Color-Intensified CMOS

1
Key Laboratory of Photo-Electronic Imaging Technology and Systems, School of Optics and Photonics, Ministry of Education of China, Beijing Institute of Technology, Beijing 100081, China
2
Science and Technology on Low-Light-Level Night Vision Laboratory, Xi’an 710065, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(11), 3891; https://doi.org/10.3390/s21113891
Submission received: 6 May 2021 / Revised: 3 June 2021 / Accepted: 3 June 2021 / Published: 4 June 2021

Abstract

:
Image intensifiers are used internationally as advanced military night-vision devices. They have better imaging performance in low-light-level conditions than CMOS/CCD. The intensified CMOS (ICMOS) was developed to satisfy the digital demand of image intensifiers. In order to make the ICMOS capable of color imaging in low-light-level conditions, a liquid-crystal tunable filter based color imaging ICMOS was developed. Due to the time-division color imaging scheme, motion artifacts may be introduced when a moving target is in the scene. To solve this problem, a deformable kernel prediction neural network (DKPNN) is proposed for joint denoising and motion artifact removal, and a data generation method which generates images with color-channel motion artifacts is also proposed to train the DKPNN. The results show that, compared with other denoising methods, the proposed DKPNN performed better both on generated noisy data and on real noisy data. Therefore, the proposed DKPNN is more suitable for color ICMOS denoising and motion artifact removal. A new exploration was made for low-light-level color imaging schemes.

1. Introduction

Since the color level that a human eye can distinguish is hundreds of times greater than the grayscale level [1], color imaging can make full use of the sensitivity of the human eye to color information, thereby improving its ability of scene understanding. Therefore, color imaging in low-light-level conditions has outstanding advantages in the areas of the military, security, etc., compared with traditional grayscale imaging, and it shows a wide range of application prospects. Low-light-level imaging techniques which use an image intensifier as the core imaging device have been used in advanced military night-vision for many years. An image intensifier can still imaging under an environmental illuminant level of 10−4 lx. A typical grayscale image of an image intensifier is shown in Figure 1a. The image intensifier is a direct-view imaging device, the phosphor screen of which is directly observed by the human eye. The image intensifier has a higher sensitivity than a CMOS/CCD. However, a CMOS/CCD has digital outputs which can be conveniently processed and stored. In order to combine their advantages, an intensified CMOS/CCD (ICMOS/ICCD) was proposed by coupling the phosphor screen of the image intensifier to the CMOS/CCD sensor. Generally, the ICMOS/ICCD has better imaging performance compared with low-light-level imaging devices such as sCMOS and EMCCD. With the development of the sensitivity of third-generation image intensifiers, it is of great significance to expand their application to low-light-level color imaging through the modulation of filtering elements such as a filter wheel [2,3,4] and a liquid-crystal tunable filter (LCTF) [5,6], whereby it may be possible for low-light-level color imaging techniques to expand their extreme working illuminant level to 10−3 lx or even 10−4 lx.
The performance of a low-light-level imaging device is not only related to the characteristics of the imaging device itself, but also the image denoising algorithm. Image denoising algorithms based on convolutional neural networks (CNNs) have achieved good denoising results in recent years. Unlike the imaging scheme of a CMOS/CCD, which simultaneously obtains various color information at different spatial pixel locations through a color filter array (CFA), due to the separation of the photosensitive surface and the digital sampling surface, the ICMOS/ICCD has to obtain various visible-band color information through filtering elements in a time-division manner. Although obtaining color information in a time-division manner can obtain full-resolution color images without the need to design additional demosaicing algorithms, if there is a motion target in the scene or if image device shaking exists, color-channel motion artifacts would be introduced, as shown in Figure 1b. Therefore, when designing denoising algorithms for time-division color imaging devices, it is necessary to take the motion artifacts into consideration.
Research on low-light-level color image denoising has explored a wide variety of technical routes. Owing to the development of neural network techniques, CNN-based low-light-level color image denoising methods with satisfactory denoising results have been proposed in recent years. However, commonly used color image denoising algorithms are mainly designed for single-frame denoising and, thus, do not consider motion artifacts. Although multi-frame color image denoising algorithms need to consider the matching of moving targets between adjacent frames, they are all aimed at the CFA-based color imaging scheme, which is different from the motion artifacts in a time-division color imaging scheme in terms of image characteristics and correction strategies; thus, they cannot be directly applied to ICMOS/ICCD denoising. Considering the relevance to the research in this article, we mainly introduce two types of methods: spatial-filter-based denoising and CNN-based denoising.
(1)
Spatial-filter-based denoising methods
Buades et al. [7] proposed a nonlocal means (NLM) denoising method, which selects pixels with similar structures in the image to perform a weighted average filtering operation, circumventing the limitations of the local information of images. Dabov et al. [8,9] proposed the BM3D denoising method, which used block matching and aggregation strategies to obtain three-dimensional image blocks, before performing denoising on the basis of collaborative Wiener filtering. BM3D remains one of the benchmark algorithms in the area of image denoising. Knaus et al. [10] proposed a progressive image denoising method that iteratively processes the original noisy image and the denoised image to obtain a better result.
(2)
CNN-based denoising methods
Mao et al. [11] proposed the denoising network RED with skip connections, inspired by the auto-encoder network UNet proposed by Ronneberger et al., for medical image segmentation [12]. Zhang et al. [13] proposed the denoising network DnCNN, the structure of which was based on the VGG [14] network, whereas batch normalization and a residual learning strategy were also introduced. Tai [15] proposed the denoising network MemNet, the long-term memory of which was set up through the skip connections between memory blocks. Liu et al. [16] proposed the denoising network WIN, which learned the noise prior of noisy images. Lefkimmiatis et al. proposed the denoising networks NLCNN [17] and UDNet [18], which embedded nonlocal operation units into the network. Kligvasser et al. [19] proposed the basic block of a denoising network named xUnit, which was trained with fewer parameters by replacing the convolutional units with xUnits. Ulyanov et al. [20] performed image denoising through network structure learning instead of weight learning. Cha et al. [21] proposed the denoising network FC-AIDE, the parameters of which were fine-tuned by combining a regularization method and learned image priors. Liu et al. [22] proposed the denoising network MWCNN, which replaced the pooling and up-convolution operations with wavelet operations. Mildenhall et al. [23] proposed the kernel prediction network (KPN) for multi-frame denoising. Image misalignment was added to training data to simulate camera motion, such that small misalignment could be corrected during denoising. Plotz et al. [24] proposed the differentiable relaxation KNN and nonlocal denoising network N3Net. Liu et al. [25] proposed the nonlocal denoising network NLRN, which achieved end-to-end training of a nonlocal network. Zhang et al. [26] proposed FFDNet, which could handle a wide range of noise levels and spatial variation noise. Lehtinen et al. [27] trained denoising networks using noisy images without using noise-free images. Liu et al. [28] proposed the BayerUnify network for raw color image denoising instead of sRGB color image denoising. Wang et al. [29] proposed the EDVR network for video denoising, which introduced submodules with different functions and adopted a residual learning strategy. Guo et al. [30] proposed the blind denoising network CBDNet, which introduced a noise level estimation subnet. Marinc et al. [31] proposed multi-KPN (MKPN), which trained multiple filter kernels with different spatial sizes and fused them when performing inference, thereby increasing the calculation efficiency of the MKPN. However, this network can still only handle small misalignment, as with KPN. Xu et al. [32,33] proposed the STPAN, which can be used for single-frame denoising and video denoising. STPAN learned deformable kernels instead of fixed-shape kernels, while it could be applied to solve large misalignment of adjacent frames in motion video scenes. Tassano et al. [34] proposed the DVDNet for video denoising, which introduced optical flow estimation for temporal denoising and relied heavily on the accuracy of optical flow estimation. The authors later improved the network and proposed the FastDVDNet [35], replacing the explicit optical flow estimation with a learning process. Tim et al. [36] proposed a training data generation method using an unprocessing approach for denoising. Ronnachal et al. [37] proposed a data generation method based on an image processing pipeline.
Denoising networks based on kernel prediction strategies such as KPN, MKPN, and STPAN have the ability to simultaneously perform image denoising and image registration. KPN and MKPN can only handle small misalignment of less than 16 pixels. Since STPAN adopts deformable kernel prediction, it has the potential to solve large misalignment between frames; however, it is mainly aimed at simultaneous color imaging schemes based on CFA. Due to the simultaneous imaging of each color channel, motion artifacts only exist between adjacent frames instead of color channels of a single frame. STPAN requires a burst of images as input and output of a denoised single-frame-prediction through sampling pixels among input images. For color ICMOS images with motion artifacts between adjacent color channels, STPAN can only treat the input single-frame color image as continuous three-frame grayscale images and output a single-frame grayscale prediction. Hence, it cannot be directly used to solve motion artifacts between adjacent color channels caused by time-division imaging. Therefore, there is currently no denoising algorithm suitable for LCTF-ICMOS images with color-channel motion artifacts. Considering that STPAN can solve large misalignment, we propose a denoising network named DKPNN that can adapt to color-channel motion artifacts using STPAN as inspiration.
A deformable kernel prediction neural network (DKPNN) is proposed for joint denoising and motion artifact removal in this article. A data generation strategy for color-channel motion artifacts is also presented, and the proposed DKPNN was trained using the generated data. This network can simultaneously complete denoising and motion artifact removal. We set up an LCTF-ICMOS experimental color imaging system to verify the effectiveness of the DKPNN. The innovations of this article are as follows:
(1)
A data generation strategy for color-channel motion artifacts;
(2)
A DKPNN which achieves joint denoising and motion artifact removal, suitable for color ICMOS/ICCD;
(3)
An LCTF-ICMOS low-light-level color imaging system with high integration.
The remainder of this article is arranged as follows: Section 2 details each step of the proposed DKPNN for joint denoising and motion artifact removal; Section 3 describes the setup of the LCTF-ICMOS experimental imaging system; Section 4 presents the validation results of the proposed method and discusses its effectiveness; and Section 5 concludes the paper and discusses future works.

2. Joint Denoising and Motion Artifact Removal Using DKPNN for Color LCTF-ICMOS

The pipeline and network structure of the proposed DKPNN are shown in Figure 2. The proposed DKPNN contains two parts: an offset-prediction subnet and a weight-prediction subnet. The offset-prediction subnet predicts the position of sampling pixels in each channel of the input color image, and the weight-prediction subnet predicts the weights of the resampled pixels. A noisy image with color-channel motion artifacts is first input to the offset-prediction subnet with a UNet structure. The offset-prediction network simultaneously outputs the predicted sampling positions of the three-dimensional deformable kernel in the spatiotemporal domain and the extracted features of the input color image. The resampled pixels are then calculated after applying trilinear interpolation to the predicted sampling positions. Next, the resampled pixels, the input image, and the extracted features are concatenated and are input to the weight-prediction subnet, which is composed of three convolutional layers. The weight-prediction subnet then outputs predicted weights. Lastly, the denoised image with motion artifacts removed is calculated through weighted averaging of the resampled pixels with the predicted weights.

2.1. Joint Denoising and Motion Artifact Removal Using DKPNN

2.1.1. Offset-Prediction Subnet

The essence of the image denoising algorithm is to utilize the zero-mean statistical characteristic of image noise. An estimate of a noise-free image can be obtained by averaging multiple irrelevant samplings of the same image signal. Hence, the effectiveness of denoising comes from the averaging operation in a statistical sense. A general image model containing noise can be expressed as
I ( x , y ) = g ( x , y ) + σ ( g ( x , y ) ) ξ ( x , y ) ,
where I(x,y) denotes the noisy image signal observed by the imaging system, x and y denote the spatial location of a pixel, g(x,y) denotes the unknown noise-free ground truth image signal, whereby the goal of the denoising algorithm is to estimate g(x,y), ξ (x,y) denotes the signal-independent zero-mean random noise with a standard deviation of 1, and σ denotes the overall noise distribution of the image as a function of g(x,y).
If the adopted noise model is signal-independent additive white Gaussian noise (AWGN), then the image noise is a zero-mean normal distribution with variance σ 0 2 , as expressed below.
σ ( g ( x , y ) ) = σ 0 ,
σ 0 ξ ( x , y ) N ( 0 , σ 0 2 ) ,
where N denotes a normal distribution.
If the adopted noise model is signal-dependent Poisson–Gaussian noise (PGN), then it is expressed as
σ ( g ( x , y ) ) · ξ ( x , y ) P ( λ ) N ( 0 , σ p · g ( x , y ) + σ g 2 ) ,
where P denotes a Poisson distribution, which is generally estimated by Gaussian distribution as shown in Equation (4) [38,39,40], while σ p and σ g denote the signal-dependent noise level and signal-independent noise level, respectively. In Equation (1), if the expectation of ξ ( x , y ) , E ( ξ ( x , y ) ) = 0 , then
E ( I ( x , y ) ) = g ( x , y ) .
In traditional temporal averaging denoising algorithms, the estimation of the noise-free image is e ( x , y ) = E ( I ( x , y ) ) . However, this has obvious limitations. First, the image sequence must be input instead of a single-frame image. Secondly, if there are moving objects in the scene or if the device is shaking, motion artifacts may be introduced. In spatial filter-based single-frame image denoising, the estimation of the noise-free image satisfies
e ( x , y ) = i = 1 n I ( x + Δ x i , y + Δ y i ) w ( x , y , i ) ,
where Δ x i and Δ y i denote the offsets of pixel (x,y), which represents the sampling process of spatial pixels, i.e., the selection strategy of pixel positions which would be involved in the weighted average operation. In typical two-dimension rectangular filter kernels, ( Δ x i , Δ y i ) is the eight-neighbor offset of target pixel (x,y), i.e., Δ x i { 1 , 0 , 1 } and Δ y i { 1 , 0 , 1 } . The filtering process of a typical rectangular filter kernel is shown in Figure 3a, where n denotes the total number of pixels in the filter kernel, and w (x,y,i) denotes the corresponding weight of the filter kernel.
As shown in Equation (6), the key to the spatial filtering denoising operation lies in the design of the pixel sampling strategy and corresponding weights. Spatial filter-based denoising methods design w (x,y,i) artificially, whereas CNN-based denoising methods learn w (x,y,i) in a data-driven manner and achieve better results when using a large number of convolution kernels. Generally, the filter kernels learned by a CNN have fixed parameters and are image content independent. The same filtering operation is used for all image scenes; thus, it is not flexible. Therefore, we adopted the deformable kernel prediction strategy (as shown in Figure 3b) proposed in [33], which simultaneously predicts the spatial pixel sampling location and the corresponding weights. Instead of directly predicting ( Δ x i , Δ y i ) in Equation (6), the sub-pixel accuracy offsets relative to a typical rectangular filter kernel were predicted, which further reflects the idea of deformation, as expressed below.
e ( x , y ) = i = 1 n I ( x + Δ x i + Δ u i , y + Δ y i + Δ v i ) w ( x , y , i ) .
If the offset between adjacent color channels is also taken into consideration and if the deformable kernel prediction is extended to color channels, then
e ( x , y , c ) = i = 1 n I ( x + Δ x i , y + Δ y i , c + Δ c i ) w ( x , y , c , i ) ,
where c denotes the index of the color channel, i.e., c { 0 , 1 , 2 } .
We adopted the same strategy as [33], whereby, instead of directly predicting ( Δ x i , Δ y i , Δ c i ), the offsets ( Δ u i , Δ v i , Δ k i ) relative to a three-dimensional cubic filter kernel were predicted, i.e.,
e ( x , y , c ) = i = 1 n I ( x + Δ x i + Δ u i , y + Δ y i + Δ v i , c + Δ c i + Δ k i ) w ( x , y , c , i ) .
The learning procedure of three-dimension deformable sampling locations is shown in Figure 4. A size of 3 × 3 × 3 was adopted as the kernel size of the three-dimensional deformable kernel for each color channel, i.e., n = 27.
After the prediction of the offsets, the input image is resampled using a trilinear interpolation method according to the predicted offsets, as expressed below.
I ( x + Δ x i + Δ u i , y + Δ y i + Δ v i , c + Δ c i + Δ k i ) = I ( p , q , j ) = i = 0 7 s i I i ,
where Ii denotes the nearest eight vertices of the position that needs to be resampled, while si denotes the corresponding sampling coefficient. They satisfy
{ I 0 = I ( x 0 , y 0 , c 0 ) I 1 = I ( x 0 , y 0 , c 1 ) I 2 = I ( x 0 , y 1 , c 0 ) I 3 = I ( x 0 , y 1 , c 1 ) I 4 = I ( x 1 , y 0 , c 0 ) I 5 = I ( x 1 , y 0 , c 1 ) I 6 = I ( x 1 , y 1 , c 0 ) I 7 = I ( x 1 , y 1 , c 1 )                 { s 0 = ( x 1 p ) ( y 1 q ) ( c 1 j ) s 1 = ( x 1 p ) ( y 1 q ) ( j c 0 ) s 2 = ( x 1 p ) ( q y 0 ) ( c 1 j ) s 3 = ( x 1 p ) ( q y 0 ) ( j c 0 ) s 4 = ( p x 0 ) ( y 1 q ) ( c 1 j ) s 5 = ( p x 0 ) ( y 1 q ) ( j c 0 ) s 6 = ( p x 0 ) ( q y 0 ) ( c 1 j ) s 7 = ( p x 0 ) ( q y 0 ) ( j c 0 ) ,
where: x0 and x1 denote the pixel locations that are closest to the pixel location to be sampled along the x-direction; y0 and y1 denote the pixel locations that are closest to the pixel location to be sampled along the y-direction; and c0 and c1 denote the pixel locations that are closest to the pixel location to be sampled along the color-channel direction. The resampling process through the trilinear interpolation method is shown in Figure 5.
The offset-prediction subnet predicts offsets relative to a three-dimensional cubic filter kernel in each color channel, which is a typical pixel-wise prediction task. The UNet is an encoder-decoder network extracting multi-scale features, where the same-scale features between encoder and decoder are connected with skip connections. It performs well in a wide range of pixel-wise prediction tasks. Therefore, the offset-prediction subnet adopted an auto-encoder structure similar to UNet. Multi-scale image features are also extracted while performing offsets prediction. By inputting the extracted multi-scale image features along with the resampled pixels into the weight-prediction subnet, there is no need to re-extract image features when performing weight prediction, thereby reducing the depth of the weight-prediction subnet and reducing the amount of model parameters. Therefore, the extracted features were also made output by the offset-prediction subnet. To achieve this goal, the first 21 convolution layers of the offset-prediction subnet, namely L1–L21, adopt the same number of feature channels as UNet. The 256-channel feature of L21 is divided into two parts of 128-channel feature, which are respectively used as 128-channel image feature output and offsets output through L22–L24 and L25–L27. For a three-dimensional cubic filter kernel with n sampling pixels, offsets along three directions are predicted for each sampling pixel and the input image contains three color channels, thus, the number of feature channels of the final output by L27 is n × 3 × 3. The number of channels in each convolutional layer of the offset-prediction subnet is shown in Table 1.

2.1.2. Weight-Prediction Subnet

Since the predicted weights are to be calculated with the resampled pixels, jointly inputting the resampled pixels along with input image and extracted image features to the weight-prediction subnet is beneficial for performing weight prediction. Therefore, the input image, resampled pixels, and image features were concatenated as concatenated features and were input to the weight-prediction subnet. Normal CNN-based denoising algorithms directly learn the mapping between noisy and noise-free image pairs, and the learned filter kernels are fixed and scene-independent. Unlike the normal training strategy, the weights predicted by the proposed weight-prediction subnet were referred to the input image scene and image features. The learned kernels were, thus, scene-dependent. Moreover, since the image features were extracted by the offset-prediction subnet, a shallow model is enough for weight prediction, which reduced the amount of model parameters. Finally, a structure containing three convolutional layers was adopted. The number of channels of L1 and L2 are both 64. Since the resampled pixels has n × 3 channels, the number of channels of L3 should also be n × 3. The number of channels in each convolutional layer is shown in Table 2. After the weighs were predicted, the final denoised image with motion artifacts removed could be calculated according to Equation (9).

2.2. Data Generation Method for Color-Channel Motion Artifacts

Since it is difficult to collect a large number of images containing color-channel motion artifacts, we used the open-access video dataset Adobe240 [41] with different kinds of motion objects to generate the needed data. The generated data were used to train the proposed DKPNN. The Adobe240 dataset contains two types of motion: object motion and device motion. It contains 133 video clips with a frame rate of 240 fps and a resolution of 1280 × 720. In order to simulate data with color-channel motion artifacts, we extracted the R, G, and B channels of every three adjacent frames of the original 240 fps video and combined them as the color channels of a single frame to generate an 80 fps video. In order to generate training data with color-channel motion artifacts of different severity, the original video frame rates were down sampled to 120 fps and 80 fps, respectively, and the same data generation strategy was adopted to obtain 40 fps and 20 fps videos. The generated videos with different frame rates were used to represent data describing small motion, medium motion, and big motion to train the network. The generated images with different motion severity are shown in Figure 6.
When generating training data, box filtering was first adopted to down sample the training image four times (two times each horizontally and vertically) to reduce the impact of other kinds of artifacts such as compression artifacts. Then, an inverse gamma correction, as used in [36], was adopted for both ground truth images and generated images to convert the image signal back into a linear response. The inverse gamma correction process can be expressed as
Γ 1 ( I ) = m a x ( I , ε ) 2.2 ,
where I denotes the input image, and ε denotes a constant factor preventing numerical instability during training; ε = 10−8 was adopted in this article.
After inverse gamma correction, image noise, described by the PGN model shown in Equation (4), was added to the image as the noisy input of the network, where σ p was randomly sampled from [10−4, 10−2] and σ g was randomly sampled from [10−3, 10−1.5].

2.3. Training

We adopted the same loss function as in [33], i.e., L1 loss with a grouped annealing term. The n × 3 sampling points were divided into q groups { Φ 0 , Φ 1 , , Φ q 1 } . At the early stage of training, the annealing term was the main loss. Each group of sampling points was stimulated to independently predict the pixel offset, the idea of which is similar to the Dropout [42] strategy. On one hand, the group strategy can prevent the training parameters from heavily relying on a few sampling points. On the other hand, due to the reduction in the number of sampling points, the grouped sampling points are stimulated by different color channels to obtain a better result of removing motion artifacts, as expressed below.
e i ( x , y , c ) = k j Φ i I ( x + Δ x j + Δ u j , y + Δ y j + Δ v j , c + Δ c j + Δ k j ) w ( x , y , c , j ) .
Then, the total loss function can be expressed as
L ( e , g ) = Γ 1 ( e ) Γ 1 ( g ) 1 + η γ t i = 1 q Γ 1 ( e i ) Γ 1 ( g ) 1 ,
where e denotes the predicted denoised image with motion artifacts removed, g denotes the ground truth image, 1 denotes L1 loss, η and γ are hyperparameters of the annealing term, and t denotes the current training step; η = 100 , γ = 0.9998 , and q = 3 were adopted in this article.
The relationship between the coefficient of the annealing term and the number of training steps is shown in Figure 7. At the early stage of training, the coefficient of the annealing term was much greater than 1, and the annealing loss was the main loss for optimization. As the number of training steps increased, the coefficient of the annealing term gradually decreased. After 23,000 training steps, the coefficient was about 1, exhibiting the same effect as L1 loss. After 35,000 training steps, the coefficient of the annealing term was much lower than 1; thus, L1 loss was the main loss for optimization.
The training data generated according to the method described in Section 2.2 were randomly cropped as 128 × 128 image blocks during training. Random horizontal flips and random vertical flips were adopted as a data augmentation strategy. The batch size was set to 16. The AdamW [43] optimizer was used to optimize the network parameters, and the weight decay was set to 0.9. The initial learning rate of the network was 1 × 10−4, while it was 1 × 10−5 after 40,000 steps. The relationship between the number of training steps and the learning rate is also shown in Figure 7. The network was trained for 50,000 steps on an NVIDIA GTX1080Ti, which took about 4.2 days.

3. Experimental Setup

The LCTF-ICMOS imaging system was improved on the basis of the results in [6]. The system was mainly composed of an objective lens, the LCTF and its controller, the GaAsP third-generation image intensifier, the relay lens, a low-light-level CMOS, and a PC. The imaging process of the LCTF-ICMOS is shown in Figure 8a. The radiant energy of the illuminant is reflected by the scene, and the reflected light is received by the objective lens of the system, which then passes through the LCTF and is focused on the photocathode of the image intensifier. Then, the phosphor screen outputs a monochrome image, which is captured by the CMOS. The LCTF sequentially changes the R, G, and B filter states driven by the synchronization signal, and the CMOS synchronizes capturing the phosphor screen of the image intensifier to obtain the corresponding R, G, and B images. After an RGB cycle, the captured adjacent R, G, and B frames are reconstructed as the final color output.
A PENTACON 50 mm fixed focus lens with a maximum F number of 1.8 was adopted as the objective lens. The tri-color LCTF of Meadowlark Optics was adopted. This LCTF contains a liquid-crystal module and a controller, which can switch R, G, and B filter states. The spectral transmittance functions of the R, G, and B filter states are shown in Figure 9a. The GaAsP third-generation image intensifier was adopted, the spectral sensitivity of which is shown in Figure 9b. The integral sensitivity of the image intensifier was 1525 μA/lm, the brightness gain was greater than 9000 cd/m2/lx, and the resolution was 54 lp/mm. The relative response function of the LCTF and image intensifier is shown in Figure 9c. A Canon EF-S 35 mm Is STM lens was adopted as the relay lens. The Photonis NOCTURN low-light-level monochrome CMOS was adopted with a resolution of 1280 × 1024, a pixel size of 9.7 μm, a maximum frame rate of 100 Hz, a dynamic range of more than 60 dB, and a readout noise of less than 4e. The final LCTF-ICMOS imaging system setup is shown in Figure 8b.

4. Results and Discussion

4.1. Generated Noisy Data

Since the ground truth data could be obtained for the generated noisy data, full-reference image quality assessment (FRIQA) metrics were adopted to evaluate the denoising results of each algorithm, namely, the peak signal-to-noise-ratio (PSNR) [44] and structural similarity index measure (SSIM) [45].
In order to demonstrate the effectiveness of the proposed method, the results were compared with those obtained using CBM3D [8], DnCNN [13], fDnCCNN [13], and FFDNet [26]. Because these denoising algorithms were not designed for color-channel motion artifact removal, the dense optical flow estimation (FE) method proposed by Farnebäck et al. [46] was adopted after denoising of the compared algorithms for color-channel registration. Then, these images were compared with the results of the proposed DKPNN.
Noise was added to the test set images using the PGN model shown in Equation (4), and two sets of noise levels were tested, where σ p = 4 × 10−3, σ g = 3 × 10−2 and σ p = 2 × 10−3, σ g = 2 × 10−2, respectively. The results of each algorithm on the generated test set data with different noise levels are similar and the results on lower noise level are shown in Figure 10. Note that we manually blurred the human faces in Figure 10 for privacy protection. The PSNR and SSIM metrics for each algorithm on the test set data with different noise levels are shown in Table 3.
The denoising results of the algorithms compared on the test set data were generally acceptable. However, if the compared algorithms did not use FE for color-channel registration, they would have been unable to deal with the color-channel motion artifacts, as shown in Figure 10c. After FE was applied, severe blur was introduced after color-channel registration, and the image details were lost.
For the generated test set data from Scene 1, the compared algorithms performed well at the edges of houses with rich details, but introduced severe blur in the regions with the lawn, sky, and moving bicycle. By contrast, the proposed DKPNN could better correct the color-channel motion artifacts and maintain image details in all the above-mentioned regions.
For the generated test set data from Scene 2, fDnCNN, FFDNet, and the proposed DKPNN achieved good denoising results, whereas residual noise could be observed in the images processed by CBM3D and DnCNN. The compared algorithms after applying FE introduced severe blur or artifacts in the details of the trophy, the reflection of the trophy, and the edge of the iron box, whereas the proposed DKPNN could better maintain the image details.
For the generated test set data from Scene 3, the compared algorithms successfully performed denoising. After applying FE, blur or artifacts were introduced on the edges of the moving bicycle, house, and plants, whereas the proposed DKPNN better maintained the image details.
For the generated test set data from Scene 4, fDnCNN, FFDNet, and the proposed DKPNN performed well in denoising the billiard table; however, residual noise could be observed in the images processed by CBM3D and DnCNN. After applying FE, the compared algorithms failed to perform color-channel registration on the moving white billiard ball and introduced severe artifacts on the billiard table. Since the white billiard ball had few details and features, FE led to poor results. Although the proposed DKPNN also presented some artifacts, its performance was superior to the compared algorithms.
Hence, the results using the generated test set data showed that, compared with other denoising methods, the proposed DKPNN performed better in denoising and color-channel motion artifact removal.

4.2. LCTF-ICMOS Noisy Data

LCTF-ICMOS noisy data under four different environmental illuminant levels using a D65 illuminant were captured using the experimental system setup described in Section 4. The four illuminant levels were 5 × 10−2 lx, 1 × 10−2 lx, 5 × 10−3 lx, and 1 × 10−3 lx. The captured scenes included static targets and motion targets. Static targets were color checkers, books, dolls, iron boxes, glass mosaic decorative mats, tank models, etc. Motion targets were two moving toy birds. LCTF-ICMOS was run at 10 fps under different illuminant levels.
Since it is difficult to obtain the ground truth images for real noisy data, no-reference image quality assessment (NRIQA) metrics were introduced to evaluate the denoising results, namely, the roughness ρ [47] and the root-mean-square contrast (RMSC) [48]. The equation describing ρ is as follows:
ρ ( I ) = h 1 I 1 + h 2 I 1 I 1 ,
where I denotes the input image, h1 denotes the horizontal filter template [1, −1], h2 denotes the vertical filter template [1, −1] T, and 1 denotes the L1 norm. Generally, a larger value of ρ indicates a rougher image, i.e., a higher image noise level.
The equation describing RMSC is as follows:
RMSC = 1 M N j = 1 N i = 1 M ( I ( i , j ) m e a n ( I ) ) 2 ,
where i and j denote the row and column indices of the pixel, respectively; M and N denote the numbers of image rows and columns, respectively; and mean () denotes the mean value calculation of image pixels. A smaller value of RMSC indicates a lower image noise level.
The two NRIQA metrics were calculated after converting color images to grayscale images. The denoised results using real noisy data under different illuminant levels are shown in Figure 11. The evaluation metrics for each algorithm are shown in Table 4.
Since the NRIQAs could not effectively distinguish image details from noise during the assessment, we adopted the strategy of manually selecting flat areas in the image, such as the colored checker patches and the gray background panels, so that the NRIQAs could better evaluate the results. The selected flat areas are shown in the green boxes in Figure 12.
As the environmental illuminant levels decreased, the performance of each denoising algorithm on LCTF-ICMOS real noisy data gradually worsened due to the sharp decrease in the image signal-to-noise ratio. As shown in Figure 11, the image characteristics of LCTF-ICMOS real noisy data are different from the Adobe240 dataset used for network training, which could worsen the performance of proposed DKPNN on both image denoising and motion artifact removal to some extent.
In terms of denoising, the adopted training image noise model is the PGN model as shown in Equation (4). The PGN model can simulate the imaging noise of CMOS/CCD sensors well, however, it is different from the image noise characteristics of LCTF-ICMOS. As shown in Figure 11a, bright spot noise can be observed in LCTF-ICMOS images. The bright spot noise is introduced by the microchannel plate in the image intensifier. The difference in noise model worsened the performance of each denoising algorithms in varying degrees. DnCNN, fDnCNN, and FFDNet performed denoising through learning the mapping between noisy and clear image pairs. Using this training strategy, when the noise models or image characteristics of training and test data are different, the performance would be worsened significantly. The denoising strategy of proposed DKPNN is essentially to aggregate similar pixels in the input image and perform weighted average, which limits the output space through directly manipulating the input pixels. This strategy weakens the impact of noise model differences. As shown in Figure 11f, the proposed DKPNN performed better than the compared algorithms in the flat areas. In order to solve the problems mentioned above, we would further study the image noise model of LCTF-ICMOS and train the DKPNN with the new noise model in future works.
In terms of color-channel motion artifact removal, the proposed DKPNN does not produce as satisfactory results on the ICMOS data as on the generated data. Firstly, the differences between ICMOS data and training Adobe240 data caused the decrease of offsets prediction accuracy of the offset-prediction subnet, thereby affecting the performance of motion artifact removal. Secondly, the difference in noise model also affects the offsets prediction accuracy. Similarly, it also affects the matching accuracy of FE algorithm. Finally, the toy bird moving targets used in the experiment has few detailed features in the ICMOS images. Similar to the billiard ball target in scene 4 of Figure 10, it is challenging to remove motion artifacts from images of such objects. However, the proposed DKPNN still performed better than the compared algorithms in maintaining the edges and details of the moving toy birds. Moreover, FE would introduce severe artifacts at the edges of input images. In order to solve the problems mentioned above, we would capture images of different still scenes using the LCTF-ICMOS setup, and generate images with motion artifacts using the method described in Section 2.2 for network training in future works.
The offset-prediction subnet of the proposed DKPNN adopted the trilinear interpolation method when calculating resampled pixels. The interpolation in the direction of the color channel caused a decrease in color saturation. As shown in scene 3 of Figure 10h, the human hands are grayish compared with the ground truth color image. Though the weight-prediction subnet would give small weights to resampled pixels with large color bias, it is still inevitable that color bias would be introduced especially in areas with severe motion artifacts. To solve this problem, designing a new loss function with a color-bias term or introducing additional color restoration modules utilizing local characteristics of input image and the denoised image can be considered.
Therefore, the proposed DKPNN performed better than the compared algorithms on both generated data and real noisy data. It can be further used to resolve the challenges related to the noised face recognition problem [49].

5. Conclusions

Image intensifiers have been used internationally as advanced military night-vision devices for many years due to their superb imaging performance in low-light-level conditions. ICMOS/ICCD was developed by coupling the phosphor screen of the image intensifier to the CMOS/CCD sensor, thereby digitizing the output of the image intensifier. An LCTF-based color imaging ICMOS experimental system was set up in this article; however, due to the time-division color imaging scheme, if there were moving targets in the scene, color-channel motion artifacts were introduced in the final image output. To solve this problem, DKPNN was proposed for joint denoising and motion artifact removal, while a data generation method for color-channel motion artifacts was also proposed to train the proposed DKPNN. The results show that, compared with other denoising algorithms, the proposed DKPNN performed better on both generated noisy data and LCTF-ICMOS real noisy data, whereby it could better maintain the image details. Since the trilinear interpolation method was adopted for pixel resampling in the proposed DKPNN, it lost color performance while applying motion artifacts removal. The size of the LCTF-ICMOS imaging system is relatively large due to the coupling strategy of the relay lens, while the incident energy is partially reduced. Therefore, in future work, corresponding color restoration methods should be studied. The training strategy could also be improved by trying more advanced data augmentation methods based on generative adversarial networks [50]. The phosphor screen in the image intensifier can be removed and a more effective coupling method can be used to further improve the signal-to-noise ratio and greatly reduce the size of the imaging system.

Author Contributions

Conceptualization, Z.H., L.L., W.J., and X.W.; methodology, Z.H.; software, Z.H.; validation, Z.H. and L.L.; investigation, Z.H. and G.J.; data collection, Z.H., X.L., and H.W.; writing—original draft preparation, Z.H.; writing—review and editing, L.L., W.J., and X.W.; project administration, L.L.; funding acquisition, X.W. and G.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Pre-Study Foundation of the Army Armament Department of China, grant number 6140414050327, and the National Natural Science Foundation of China, grant number 62031018. The APC was funded by the Beijing Institute of Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Toet, A.; Hogervorst, M.A. Progress in color night vision. Opt. Eng. 2012, 51, 010901. [Google Scholar] [CrossRef]
  2. ColorPath CCNVD Color Capable Night Vision. Available online: https://armament.com/tenebraex/ (accessed on 25 April 2021).
  3. Chen, Y.; Hu, W.; Wu, D.; He, Y.; Zhang, D.; Zhang, Y. Spectrum matching optimization study on tri-band color night vision technology. Opt. Instrum. 2016, 38, 313–319. [Google Scholar]
  4. Wu, H.; Tao, S.; Gu, G.; Wang, S. Study on color night vision image fusion method based on quadruple-based images. Acta Photonica Sin. 2017, 46, 175–184. [Google Scholar]
  5. Kriesel, J.; Gat, N. True-Color. Night Vision Cameras, Optics and Photonics in Global Homeland Security III; International Society for Optics and Photonics: Orlando, FL, USA, 2007; p. 65400D. [Google Scholar]
  6. Yuan, T.; Han, Z.; Li, L.; Jin, W.; Wang, X.; Wang, H.; Bai, X. Tunable-liquid-crystal-filter-based low-light-level color night vision system and its image processing method. Appl. Opt. 2019, 58, 4947–4955. [Google Scholar] [CrossRef]
  7. Buades, A.; Coll, B.; Morel, J.-M. A non-local algorithm for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 2, pp. 60–65. [Google Scholar]
  8. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  9. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K.O. Image restoration by sparse 3D transform-domain collaborative filtering. In Image Processing: Algorithms and Systems VI; International Society for Optics and Photonics: San Jose, CA, USA, 2008; p. 681207. [Google Scholar]
  10. Knaus, C.; Zwicker, M. Progressive Image Denoising. IEEE Trans. Image Process. 2014, 23, 3114–3125. [Google Scholar] [CrossRef] [PubMed]
  11. Mao, X.J.; Shen, C.; Yang, Y.B. Image Restoration Using Very Deep Convolutional Encoder-Decoder Networks with Symmetric Skip Connections. arXiv 2016, arXiv:1603.09056. [Google Scholar]
  12. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Munich, Germany, 5–9 October 2015, Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  13. Zhang, K.; Zuo, W.; Chen, Y.; Meng, D.; Zhang, L. Beyond a Gaussian Denoiser: Residual Learning of Deep CNN for Image Denoising. IEEE Trans. Image Process. 2017, 26, 3142–3155. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  15. Tai, Y.; Yang, J.; Liu, X.; Xu, C. MemNet: A persistent memory network for image restoration. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4549–4557. [Google Scholar]
  16. Liu, P.; Fang, R. Learning pixel-distribution prior with wider convolution for image denoising. arXiv 2017, arXiv:1707.09135. [Google Scholar]
  17. Lefkimmiatis, S. Non-local color image denoising with convolutional neural networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5882–5891. [Google Scholar]
  18. Lefkimmiatis, S. Universal denoising networks: A novel CNN architecture for image denoising. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 19–21 June 2018; pp. 3204–3213. [Google Scholar]
  19. Kligvasser, I.; Shaham, T.R.; Michaeli, T. xUnit: Learning a spatial activation function for efficient image restoration. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2433–2442. [Google Scholar]
  20. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Deep image prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9446–9454. [Google Scholar]
  21. Cha, S.; Moon, T. Fully convolutional pixel adaptive image denoiser. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 17 October–2 November 2019; pp. 4160–4169. [Google Scholar]
  22. Liu, P.; Zhang, H.; Zhang, K.; Lin, L.; Zuo, W. Multi-level wavelet-CNN for image restoration. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 773–782. [Google Scholar]
  23. Mildenhall, B.; Barron, J.T.; Chen, J.; Sharlet, D.; Ng, R.; Carroll, R. Burst denoising with kernel prediction networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 2502–2510. [Google Scholar]
  24. Plötz, T.; Roth, S. Neural nearest neighbors networks. Adv. Neural Inf. Process. Syst. 2018, 31, 1087–1098. [Google Scholar]
  25. Liu, D.; Wen, B.; Fan, Y.; Loy, C.C.; Huang, T.S. Non-local recurrent network for image restoration. arXiv 2018, arXiv:1806.02919. [Google Scholar]
  26. Zhang, K.; Zuo, W.; Zhang, L. FFDNet: Toward a Fast and Flexible Solution for CNN-Based Image Denoising. IEEE Trans. Image Process. 2018, 27, 4608–4622. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Lehtinen, J.; Munkberg, J.; Hasselgren, J.; Laine, S.; Karras, T.; Aittala, M.; Aila, T. Noise2noise: Learning image restoration without clean data. arXiv 2018, arXiv:1803.04189. [Google Scholar]
  28. Liu, J.; Wu, C.-H.; Wang, Y.; Xu, Q.; Zhou, Y.; Huang, H.; Wang, C.; Cai, S.; Ding, Y.; Fan, H.; et al. Learning raw image denoising with bayer pattern unification and bayer preserving augmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 2070–2077. [Google Scholar]
  29. Wang, X.; Chan, K.C.; Yu, K.; Dong, C.; Loy, C.C. EDVR: Video restoration with enhanced deformable convolutional networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Long Beach, CA, USA, 16–17 June 2019; pp. 1954–1963. [Google Scholar]
  30. Guo, S.; Yan, Z.; Zhang, K.; Zuo, W.; Zhang, L. Toward convolutional blind denoising of real photographs. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–17 June 2019; pp. 1712–1722. [Google Scholar]
  31. Marinc, T.; Srinivasan, V.; Gul, S.; Hellge, C.; Samek, W. Multi-kernel prediction networks for denoising of burst images. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 2404–2408. [Google Scholar]
  32. Xu, X.; Li, M.; Sun, W. Learning deformable kernels for image and video denoising. arXiv 2019, arXiv:1904.06903. [Google Scholar]
  33. Xu, X.; Li, M.; Sun, W.; Yang, M.-H. Learning Spatial and Spatio-Temporal Pixel Aggregations for Image and Video Denoising. IEEE Trans. Image Process. 2020, 29, 7153–7165. [Google Scholar] [CrossRef]
  34. Tassano, M.; Delon, J.; Veit, T. DVDNET: A fast network for deep video denoising. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1805–1809. [Google Scholar]
  35. Tassano, M.; Delon, J.; Veit, T. FastDVDnet: Towards real-time deep video denoising without flow estimation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 1351–1360. [Google Scholar]
  36. Brooks, T.; Mildenhall, B.; Xue, T.; Chen, J.; Sharlet, D.; Barron, J.T. Unprocessing images for learned raw denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Taipei, Taiwan, 22–25 September 2019; pp. 11036–11045. [Google Scholar]
  37. Jaroensri, R.; Biscarrat, C.; Aittala, M.; Durand, F. Generating training data for denoising real rgb images via camera pipeline simulation. arXiv 2019, arXiv:1904.08825. [Google Scholar]
  38. Foi, A.; Trimeche, M.; Katkovnik, V.; Egiazarian, K. Practical Poissonian-Gaussian noise modeling and fitting for single-image raw-data. IEEE Trans. Image Process. 2008, 17, 1737–1754. [Google Scholar] [CrossRef] [Green Version]
  39. Hasinoff, S.W.; Durand, F.; Freeman, W.T. Noise-optimal capture for high dynamic range photography. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 553–560. [Google Scholar]
  40. Hasinoff, S.W. Photon, Poisson Noise. 2014. Available online: http://people.csail.mit.edu/hasinoff/pubs/hasinoff-photon-2011-preprint.pdf (accessed on 25 April 2021).
  41. Su, S.; Delbracio, M.; Wang, J.; Sapiro, G.; Heidrich, W.; Wang, O. Deep video deblurring for hand-held cameras. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 237–246. [Google Scholar]
  42. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  43. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  44. Hore, A.; Ziou, D. Image quality metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  45. Wang, Z.; Simoncelli, E.; Bovik, A. Multiscale structural similarity for image quality assessment. In Proceedings of the Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; pp. 1398–1402. [Google Scholar] [CrossRef] [Green Version]
  46. Farnebäck, G. Two-frame motion estimation based on polynomial expansion. In Proceedings of the Scandinavian Conference on Image Analysis, Halmstad, Sweden, 29 June–2 July 2003; Springer: Berlin/Heidelberg, Germany, 2003; pp. 363–370. [Google Scholar]
  47. Hayat, M.M.; Torres, S.N.; Armstrong, E.; Cain, S.C.; Yasuda, B. Statistical algorithm for nonuniformity correction in focal-plane arrays. Appl. Opt. 1999, 38, 772–780. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  48. Peli, E. Contrast in complex images. J. Opt. Soc. Am. A 1990, 7, 2032–2040. [Google Scholar] [CrossRef] [PubMed]
  49. Adjabi, I.; Ouahabi, A.; Benzaoui, A.; Jacques, S. Multi-Block Color-Binarized Statistical Images for Single-Sample Face Recognition. Sensors 2021, 21, 728. [Google Scholar] [CrossRef] [PubMed]
  50. Khaldi, Y.; Benzaoui, A. A new framework for grayscale ear images recognition using generative adversarial networks under unconstrained conditions. Evol. Syst. 2020, 1–12. [Google Scholar] [CrossRef]
Figure 1. Imaging from image intensifiers and color-channel motion artifacts introduced by time-division color imaging: (a) typical monochromatic image from an image intensifier; (b) color-channel motion artifacts.
Figure 1. Imaging from image intensifiers and color-channel motion artifacts introduced by time-division color imaging: (a) typical monochromatic image from an image intensifier; (b) color-channel motion artifacts.
Sensors 21 03891 g001
Figure 2. The pipeline and network structure of the proposed DKPNN.
Figure 2. The pipeline and network structure of the proposed DKPNN.
Sensors 21 03891 g002
Figure 3. The filtering processes of a typical rectangular filter kernel and a deformable filter kernel: (a) rectangular filter kernel; (b) deformable filter kernel.
Figure 3. The filtering processes of a typical rectangular filter kernel and a deformable filter kernel: (a) rectangular filter kernel; (b) deformable filter kernel.
Sensors 21 03891 g003
Figure 4. The learning procedure of three-dimensional deformable sampling locations.
Figure 4. The learning procedure of three-dimensional deformable sampling locations.
Sensors 21 03891 g004
Figure 5. The resampling process based on trilinear interpolation.
Figure 5. The resampling process based on trilinear interpolation.
Sensors 21 03891 g005
Figure 6. The generated images with different motion severity: (a) ground truth; (b) small motion; (c) medium motion; (d) big motion.
Figure 6. The generated images with different motion severity: (a) ground truth; (b) small motion; (c) medium motion; (d) big motion.
Sensors 21 03891 g006
Figure 7. Relationship between the number of training steps and the coefficient of the annealing term and the learning rate.
Figure 7. Relationship between the number of training steps and the coefficient of the annealing term and the learning rate.
Sensors 21 03891 g007
Figure 8. LCTF-ICMOS color imaging system: (a) imaging process of the system; (b) system setup with shell opened.
Figure 8. LCTF-ICMOS color imaging system: (a) imaging process of the system; (b) system setup with shell opened.
Sensors 21 03891 g008
Figure 9. Spectral sensitivity function of LCTF-ICMOS: (a) spectral transmittance function of LCTF; (b) spectral sensitivity function of image intensifier; (c) relative response function of LCTF and image intensifier.
Figure 9. Spectral sensitivity function of LCTF-ICMOS: (a) spectral transmittance function of LCTF; (b) spectral sensitivity function of image intensifier; (c) relative response function of LCTF and image intensifier.
Sensors 21 03891 g009
Figure 10. Denoising results on generated test set data: (a) input noisy image with color-channel motion artifacts; (b) ground truth; (c) CBM3D; (d) CBM3D + FE; (e) DnCNN + FE; (f) fDnCNN + FE; (g) FFDNet + FE; (h) proposed DKPNN.
Figure 10. Denoising results on generated test set data: (a) input noisy image with color-channel motion artifacts; (b) ground truth; (c) CBM3D; (d) CBM3D + FE; (e) DnCNN + FE; (f) fDnCNN + FE; (g) FFDNet + FE; (h) proposed DKPNN.
Sensors 21 03891 g010aSensors 21 03891 g010b
Figure 11. Results using real noisy data under different illuminant levels, with scene 1 on the top row and scene 2 on the bottom row: (a) noisy input; (b) CBM3D + FE; (c) DnCNN + FE; (d) fDnCNN + FE; (e) FFDNet + FE; (f) proposed DKPNN.
Figure 11. Results using real noisy data under different illuminant levels, with scene 1 on the top row and scene 2 on the bottom row: (a) noisy input; (b) CBM3D + FE; (c) DnCNN + FE; (d) fDnCNN + FE; (e) FFDNet + FE; (f) proposed DKPNN.
Sensors 21 03891 g011aSensors 21 03891 g011b
Figure 12. The selected flat areas for NRIQA calculation: (a) scene 1; (b) scene 2.
Figure 12. The selected flat areas for NRIQA calculation: (a) scene 1; (b) scene 2.
Sensors 21 03891 g012
Table 1. The number of channels in each convolutional layer of the offset-prediction subnet.
Table 1. The number of channels in each convolutional layer of the offset-prediction subnet.
Convolutional LayersL1–3L4–6L7–9L10–12L13–15L16–18L19–21L22–24L25–26L27
Number of channels64128256512512512256128128n × 3 × 3
Table 2. The number of channels in each convolutional layer of the weight prediction subnet.
Table 2. The number of channels in each convolutional layer of the weight prediction subnet.
Convolutional LayersL1L2L3
Number of channels6464n × 3
Table 3. The evaluation metrics on the test set data for each algorithm.
Table 3. The evaluation metrics on the test set data for each algorithm.
Noise LevelMetricsBM3DDnCNNfDnCNNFFDNetBM3D + FEDnCNN + FEfDnCNN + FEFFDNet + FEDKPNN
σ p = 2 × 10 - 3
σ g = 2 × 10 - 2
PSNR22.6823.0222.8822.8425.0626.5026.5826.5129.33
SSIM0.7210.7320.7260.7250.8180.8310.8340.8320.887
σ p = 4 × 10 - 3
σ g = 3 × 10 - 2
PSNR21.5622.3322.0122.1124.3425.7825.8025.7428.68
SSIM0.7120.7220.7150.7160.8060.8200.8260.8230.878
Table 4. Evaluation metrics for each algorithm under different illuminant levels.
Table 4. Evaluation metrics for each algorithm under different illuminant levels.
MetricsSceneIlluminant Levels (lx)Algorithms
Noisy InputBM3DDnCNNfDnCNNFFDNetProposed
ρ15 × 10−20.04120.01170.02480.00790.00790.0059
1 × 10−20.05770.01280.02860.01030.01030.0081
5 × 10−30.07190.01440.02910.01110.01080.01
1 × 10−30.07780.01480.03020.01140.01130.0108
25 × 10−20.04360.00990.02690.00590.00580.0042
1 × 10−20.06010.01530.02730.01120.01130.0086
5 × 10−30.07210.01540.03060.0130.01310.0111
1 × 10−30.08070.01650.03080.01330.01320.0125
RMSC15 × 10−28.12875.67517.2995.63615.71685.4267
1 × 10−210.81236.55648.9726.09546.20025.4395
5 × 10−314.26218.732711.32977.98258.2066.9467
1 × 10−317.238811.876914.081410.949211.31089.9249
25 × 10−28.84643.70977.01862.31802.20601.9741
1 × 10−211.98455.55699.26523.78493.79552.1758
5 × 10−313.53076.282610.03094.52144.58672.5442
1 × 10−315.34557.851711.78875.82535.71433.5552
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Han, Z.; Li, L.; Jin, W.; Wang, X.; Jiao, G.; Liu, X.; Wang, H. Denoising and Motion Artifact Removal Using Deformable Kernel Prediction Neural Network for Color-Intensified CMOS. Sensors 2021, 21, 3891. https://doi.org/10.3390/s21113891

AMA Style

Han Z, Li L, Jin W, Wang X, Jiao G, Liu X, Wang H. Denoising and Motion Artifact Removal Using Deformable Kernel Prediction Neural Network for Color-Intensified CMOS. Sensors. 2021; 21(11):3891. https://doi.org/10.3390/s21113891

Chicago/Turabian Style

Han, Zhenghao, Li Li, Weiqi Jin, Xia Wang, Gangcheng Jiao, Xuan Liu, and Hailin Wang. 2021. "Denoising and Motion Artifact Removal Using Deformable Kernel Prediction Neural Network for Color-Intensified CMOS" Sensors 21, no. 11: 3891. https://doi.org/10.3390/s21113891

APA Style

Han, Z., Li, L., Jin, W., Wang, X., Jiao, G., Liu, X., & Wang, H. (2021). Denoising and Motion Artifact Removal Using Deformable Kernel Prediction Neural Network for Color-Intensified CMOS. Sensors, 21(11), 3891. https://doi.org/10.3390/s21113891

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop