Next Article in Journal
Optimisation of Deep Learning Small-Object Detectors with Novel Explainable Verification
Previous Article in Journal
Cortical Correlates of Increased Postural Task Difficulty in Young Adults: A Combined Pupillometry and EEG Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Low-Light Image Enhancement via Retinex-Style Decomposition of Denoised Deep Image Prior

1
Department of Basic Sciences, Shanxi Agricultural University, Jinzhong 030801, China
2
School of Mathematics and Statistics, Qilu University of Technology (Shandong Academy of Sciences), Jinan 250353, China
3
School of Mathematical Sciences, Dalian University of Technology, Dalian 116024, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(15), 5593; https://doi.org/10.3390/s22155593
Submission received: 8 June 2022 / Revised: 10 July 2022 / Accepted: 24 July 2022 / Published: 26 July 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Low-light images are a common phenomenon when taking photos in low-light environments with inappropriate camera equipment, leading to shortcomings such as low contrast, color distortion, uneven brightness, and high loss of detail. These shortcomings are not only subjectively annoying but also affect the performance of many computer vision systems. Enhanced low-light images can be better applied to image recognition, object detection and image segmentation. This paper proposes a novel RetinexDIP method to enhance images. Noise is considered as a factor in image decomposition using deep learning generative strategies. The involvement of noise makes the image more real, weakens the coupling relationship between the three components, avoids overfitting, and improves generalization. Extensive experiments demonstrate that our method outperforms existing methods qualitatively and quantitatively.

1. Introduction

With the great breakthrough of deep learning in the field of computer vision technology, image processing has been widely used in many fields, e.g., face recognition [1], defect detection [2], medical image retrieval [3], traffic information systems [4], and text recognition [5]. Image defects can be attributed to uncontrolled factors such as insufficient lighting conditions and non-uniform lighting during image capture. These unfavorable elements can be disturbed by backlighting, underexposure, and night-time conditions. Low-light images are usually noisy, low-contrast, color-distorted, and quality-impaired. These shortcomings not only result in an unpleasant visual experience but also affect the performance of many computer vision systems, e.g., for image recognition, object detection, and image segmentation.
Image enhancement has a wide range of applications in different fields, e.g., underwater images [6], high-speed railway images [7], and robot vision [8]. In general, there are two ways to improve the image quality. One is to improve the hardware performance of photographic equipment and the other is to process the obtained image. However, the former has disadvantages such as manufacturing difficulties, high cost, and complicated technology. Therefore, in practical applications, improving the quality of low-light images through enhancement algorithms is of great significance. Low-light image enhancement has two main purposes: improving contrast and suppressing noise. The enhanced image is more suitable for human observation and computer vision systems.
Related studies on low-light image enhancement are reviewed, including those using conventional methods and deep learning methods. Traditional low-light enhancement methods include methods based on histogram equalization (HE) and the Retinex model. Histogram equalization is a method of using an image histogram to adjust contrast in the field of image processing (BPDHE [9], DHE [10], histogram modification [11]). HE methods may increase the contrast of noise and reduce the contrast of useful signals. In view of the shortcomings of the HE method, many improved versions have been proposed, e.g., clipped AHE [12], CLAHE [13], CVC [14], and contrast enhancement algorithm [15].
Retinex model-based methods decompose low-light images into reflection and illumination components [16]. Given a low-light image S, it can be decomposed into S = R I , where S represents the low-light image, R represents the reflectance, I represents the illumination map, and ⊙ represents the dot product operation. In addition, many improved versions of Retinex models have been derived from the Retinex theory, including the single-scale Retinex model [17], the multi-scale Retinex model [18], the naturalness preserved enhancement algorithm [19], the fusion-based enhancing method [20], and illumination map estimation [21]. There are also some algorithms based on the variational Retinex model, e.g., the variational Retinex model formulated as a quadratic optimization problem [22], the variational framework for Retinex introducing a bright channel [23], the variational Retinex model based on the L 2 -norm [24], the hybrid L 2 - L p variational model with bright channel prior [25], and the maximum-entropy-based Retinex model [26]. Based on the computational complexity of variational methods, the disadvantage of this method is that processing images is time-consuming.
With the development of artificial intelligence, deep learning methods have also been widely used in the field of low-light image enhancement. Lore et al. [27] proposed a method of enhancing natural low-light images using a stacked sparse denoising autoencoder. Tao et al. [28] introduced a CNN method utilizing multi-scale feature maps to perform low-light image enhancement. Ignatov et al. [29] proposed a residual convolutional network that combines the composite perceptual error functions of content, color, and texture losses to improve the color and detail sharpness of the image. Shen et al. [30] put forward a convolutional neural network that directly learns the end-to-end mapping between dark and bright images for low-light image enhancement. Gharbi et al. [31] introduced a neural network architecture using input/output image pairs to perform image augmentation in real time and with full-resolution images. Wei et al. [32] designed a deep network called Retinex-Net based on the Retinex model, including Decom-Net for decomposition and Enhance-Net for lighting adjustment. Wang et al. [33] proposed a convolutional neural network based on the global prior information generated in the encoder–decoder network to enhance images. Chen et al. [34] presented a fully end-to-end convolutional network for processing low-light images using raw image data. Chen et al. [35] proposed an unpaired learning method for image enhancement based on a bidirectional generative adversarial network (GAN) framework. Zhang et al. [36] constructed an efficient network (KinD) trained on paired images shot under different exposure conditions. Wang et al. [37] proposed a neural network for enhancing underexposed photos by introducing intermediate lighting into the network to correlate the input with the expected enhancement result. Jiang et al. [38] proposed an unsupervised generative adversarial network trained with unpaired images. Yang et al. [39] suggested a semi-supervised learning method for low-light image enhancement based on a deep recursive band network (DRBN). Lv et al. [40] presented an end-to-end lightweight network for non-uniform illumination image enhancement that retains the advantages of the Retinex model and overcomes its limitations. Wang et al. [41] proposed the Deep Lightening Network (DLN) composed of several lightening back-projection (LBP) blocks to estimate residuals between low-light and normal-light images and the residual between low and normal light images. Zhu et al. [42] proposed the Edge-Enhanced Multi-Exposure Fusion Network (EEMEFN), which includes a multi-exposure fusion module and an edge enhancement module to enhance extremely low-light images. Liu et al. [43] obtained a Retinex-inspired Unrolling with Architecture Search (RUAS), where a cooperative architecture search was used to discover low-light prior architectures from a compact search space, and reference-free losses were used to train the network. Li et al. [44] presented a progressive–recursive image enhancement network (PRIEN) that uses a recursive unit to progressively enhance the input image. Zhang et al. [45] proposed dynamic fields to learn and make inferences from a single image, and then enforce temporal consistency. Fu et al. [46] suggested a novel unsupervised low-light image enhancement network (LE-GAN) based on generative adversarial networks using unpaired low-light/normal-light images for training. Zhao et al. [47] proposed a unified deep zero-reference framework termed RetinexDIP for enhancing low-light images; however, noise was not considered in the decomposition process. Liu et al. [48] proposed the Retinex-based fast algorithm (RBFA) to achieve low-light image enhancement. Liang et al. [49] proposed a low-light image enhancement model based on deep learning. Li et al. [50] presented a low-light image enhancement method based on a deep symmetric encoder–decoder convolutional network. Han et al. [51] proposed a DIP based on a noise-robust super resolution method. Ai and Kwon [52] used attention U-Net for extreme low-light image enhancement. Zhao et al. [53] proposed a multi-path interaction network to improve the quality of the image.
In this paper, we propose a novel RetinexDIP method to enhance images. Noise components are introduced into our network, and three components are generated by the DIP network. The involvement of noise makes the image more real, weakens the coupling relationship between the three components, avoids overfitting, and improves generalization. The illumination map is obtained by iterating and adjusting the input noise, and then the enhanced image is generated based on the Retinex model. Our training process is a zero-reference process and does not require any paired or even unpaired data, which is similar to existing methods (EnlightenGAN [38], CycleGAN [54], Zero-DCE [55]). The novel RetinexDIP method can be applied to various poorly lit environments and has good generalization. The loss function in this paper is composed of four parts: the spatial reconstruction loss, illumination-consistency loss, reflectance loss, and illumination smoothness loss. The experimental results show that the normal light images generated by our method are natural and clear and the method has excellent performance according to both visual observation and objective evaluation indicators. The main contributions of this paper are as follows:
  • We propose a novel noise-added RetinexDIP method to enhance images.
  • Three components are generated by the DIP network.
  • The zero-reference process avoids the risk of overfitting and improves generalization.
  • The experimental results show that our method significantly outperforms some current state-of-the-art methods.
The rest of the paper is organized as follows. Section 2 details our proposed approach. Section 3 presents the experimental results, and the last section concludes the paper.

2. Materials and Methods

Given a low-light image S and considering noise, the image can be decomposed into:
S = R I + N
or
S = ( R + N ) I ,
where S represents the low-light image, R represents the reflectance, I represents the illumination map, N denotes the noise, and ⊙ represents the dot product operation. Adding hand-crafted priors to components makes the components more coupled. Deep Image Prior (DIP) means that complex prior knowledge does not need to be introduced, as it can be encoded in the structure of the neural network itself  [56]. In practical problems, it is difficult to find pairs of low-light and normal images. Therefore, generative models are becoming more and more important.
In this paper, we implement image decomposition based on Retinex theory and generative strategies, taking into account the noise factor. The overall framework of this method is shown in Figure 1. As can be seen from Figure 1, there are three encoder–decoder networks (DIP1, DIP2, and DIP3) in the model. These DIP networks are all convolutional operations. DIP1 is used to generate noise N, and DIP2 and DIP3 are used to generate the reflectance R and the latent illumination I. All three DIP networks use white Gaussian noise as input and obey z 1 , 2 , 3 N ( 0 , σ 2 ) , where σ 2 represents the variance of the Gaussian distribution. The noise is obtained via random sampling and has the same size as the image.
To evaluate the quality of the augmented images, the following four types of losses were employed to train our model.
Reconstruction Loss. The reconstruction loss is defined according to the following form:
l r e c = g I ( z 3 ) ( g R ( z 2 ) + g N ( z 1 ) ) S 0 2 2 ,
where N is the noise generated by DIP1, denoted by g N , R is the latent reflectance generated by DIP2, denoted by g R , and I is the illumination generated by DIP3, denoted by g I . S 0 is the observed image.
Illumination-consistency Loss. As in [47], we also consider the illumination-consistency loss, which is defined as
l i c = g I ( z 3 ) I 0 1 ,
where I 0 is the initial illumination obtained by
I 0 ( p ) = max c { R , G , B } S c ( p )
for every pixel p.
Reflectance Loss. In this paper, the reflectance R is considered, and the total variation (TV) constraint [57] is defined as
l r e f = g R ( z 2 ) 1 ,
where ∇ denotes the first-order operator containing a horizontal component h and a vertical component v .
Illumination Smoothness Loss. We also use the illumination reflection gradient-weighted TV constraint, defined as
l i s = W g I ( z 3 ) 1 ,
where W is the weight matrix. According to the weight strategy in [21], it is set via:
W h , v 1 | h , v I 0 | + ϵ ,
where ϵ is a small decimal to ensure that the denominator is not 0.
Combining the four losses, we minimize the objective function as follows:
arg min I , R , N g I ( z 3 ) ( g R ( z 2 ) + g N ( z 1 ) ) S 0 2 2 + λ 1 g I ( z 3 ) I 0 1 + λ 2 g R ( z 2 ) 1 + λ 3 W g I ( z 3 ) 1 ,
where λ 1 , λ 2 , and λ 3 are the balance parameters.
The enhanced image S is composed of noise N, reflectance R, and the latent illumination I.
S = ( R + N ) I ,
or
S = g I ( z 3 ) ( g R ( z 2 ) + g N ( z 1 ) ) .
Next, enhancement using only estimated illumination is described. There are two commonly used composition strategies: one is to remove the illumination component, considering the reflectance as the enhancement result, i.e.  S ^ = S / I , and the other is to adjust the illumination and reconstruct the result with the reflectance, i.e.  S ^ = I ^ R . In this paper, we use a variant of the former strategy, that is, S ^ = S / I ^ (refer to [21] for details).
We adjust the illumination distribution of decomposition using the gamma correction I ^ = I γ , where γ is the correction factor. To sum up, the enhanced result is given by:
S ^ = S c / I ^ , c { R , G , B } .
The whole operation process is shown in Algorithm 1.
Algorithm 1: our algorithm
Sensors 22 05593 i001

3. Experiment

In this section, the experimental parameter settings, public low-light image datasets, and performance metrics are introduced. The results of our approach with different methods are also be discussed.

3.1. Settings

We implement our framework using PyTorch on an NVIDIA 2080Ti GPU. The model experimental parameters were set as follows: λ 1 = 1 , λ 2 = 0.0001 , λ 3 = 0.5 , δ = 0.01 , γ = 0.5 , and K = 300 . We use six public datasets with low-light images for the experiments, including DICM [15], Fusion [58], LIME [21], MEF [59], NPE [19], and VV (https://sites.google.com/site/vonikakis/datasets (accessed on 1 June 2022)).

3.2. Performance Criteria

In this paper, we measure the experimental results from visual observations and objective evaluation indicators. The following evaluation indicators were used.
Natural Image Quality Evaluator (NIQE). The inspiration for NIQE is based on constructing a series of features used to measure image quality and using these features to fit a multivariate Gaussian model. In the evaluation process, the distance between the image feature model parameters (to be evaluated) and the pre-established model parameters is used to determine the image quality. A lower NIQE score indicates better preservation of naturalness. For details, refer to [60].
No-reference Image Quality Metric for Contrast Distortion (NIQMC). NIQMC is defined as a simple linear fusion of global and local quality measures [61]. A higher NIQMC score represents better image contrast.
Colorfulness-Based Patch-Based Contrast Quality Index (CPCQI). CPCQI is a color-based PCQI metric that evaluates the enhancement effect between input and enhanced output in terms of mean strength, signal strength, and signal structure components [62]. A larger CPCQI value indicates a higher contrast ratio.

3.3. Results

In this section, we show the effectiveness of the proposed method. We compare it with six other methods, i.e., LIME [21], NPE [19], SRIE [63], KinD [36], Zero-DCE [55], and RetinexDIP [47].
The specific process for our method is shown in detail step by step in Figure 2.
First, we evaluate the different methods qualitatively. As shown in Figure 3, Figure 4, Figure 5 and Figure 6, we select local regions and zoom in on them for intuitive comparison with other methods. The following conclusions can be drawn from the observation of Figure 3. The enhancement effect of the NPE, SRIE, and KinD methods is not obvious. The LIME and RetinexDIP methods produce over-enhancement effects in these regions. The processing result of Zero-DCE has unnatural color. Our method yields natural exposure and clear details. In Figure 4, it can be seen that our method enhances the image and the edges are clearly visible. The result of KinD has an unnatural color. By considering Figure 5, it can be seen that the method proposed in this paper does not have the problems of overexposure and artifacts when improving the contrast. From Figure 6, it can also be concluded that our method improves the contrast effectively and maintains the natural color at the same time.
In the following, we compare the proposed method with other methods quantitatively. The red, green, and blue scores represent the top three in the corresponding dataset, respectively. Table 1 presents the NIQE metrics of different methods on the six datasets. Notably, a lower NIQE score indicates better preservation of naturalness. Our method achieves the best results on the MEF and VV datasets and the second-best results on the average of the six datasets and LIME. Table 2 presents the NIQMC metrics of the different methods on the six datasets. A higher NIQMC score represents better image contrast. Our method is in the top three for DICM, LIME, MEF, NPE, VV, and the average of the six datasets. Table 3 presents the CPCQI of the different methods on the six datasets. A larger CPCQI value indicates a higher contrast ratio. Our method achieves the best results on DICM, Fusion, NPE, and the average of the six datasets, and also performs well on several other datasets.
As shown in Table 4, the runtimes of different methods were compared. In the experiment, we compare the runtimes of three traditional methods (LIME, NPE, SRIE) and three deep learning methods (KinD, Zero-DCE, RetinexDIP) to that of our method, with eight different input image sizes. Compared with NPE, SRIE, and RetinexDIP, we find that our method is more efficient on high-resolution images. Unlike traditional methods such as NPE and SRIE, the proposed method uses the DIP network to compute reflections and illumination. Benefiting from the convolutional structure, the runtime of the DIP model changes very little as the image resolution grows. Compared with RetinexDIP, the proposed method converges faster and requires less runtime due to the consideration of noise. Compared with Zero-DCE and KinD, our method can also save memory, since Zero-DCE and KinD are pixel-wise methods, while the proposed method is based on Retinex decomposition. The proposed method does not require the actual resolution of the image in the operation, and the memory will not increase significantly with an increase in the image resolution.

4. Conclusions

In this paper, we propose a novel low-light image enhancement method via Retinex decomposition of denoised Deep Image Prior. Noise is considered in the image decomposition using deep learning generative strategies. As a comparison, we also consider six other methods, i.e., LIME, NPE, SRIE, KinD, Zero-DCE, and RetinexDIP. Extensive experiments demonstrate that our method outperforms existing methods qualitatively and quantitatively. Unlike some other learning-based methods, the method proposed in this paper is a no-reference method, which means that only the input images are required without any extra data. Taking the reflection noise into consideration, our experiments show that the denoised Deep Image Prior can produce images with less noise.
In real scenes, noise always conforms to some scene-dependent distribution such as the Poisson distribution. In future work, other approaches such as normalizing flow will be considered to simulate a more realistic noise distribution than that of DIP.

Author Contributions

Conceptualization, X.G. and J.L.; methodology, X.G. and J.L.; validation, X.G., M.Z., and J.L.; resources, X.G.; writing—original draft preparation, X.G.; writing—review and editing, M.Z. and J.L.; supervision, M.Z. and J.L.; project administration, X.G.; funding acquisition, X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Innovation Fund Project of Shanxi Agricultural University (2021BQ10); the National Natural Science Foundation of China (12101378); the Scientific Research for Excellent Doctors Project, Shanxi Province, China (SXBYKY2021046); and the Shanxi Provincial Research Foundation for Basic Research, China (20210302124548).

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to the anonymous reviewers and the editors for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hu, G.; Yang, Y.; Yi, D.; Kittler, J.; Christmas, W.; Li, S.Z.; Hospedales, T. When face recognition meets with deep learning: An evaluation of convolutional neural networks for face recognition. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 142–150. [Google Scholar]
  2. Koch, C.; Georgieva, K.; Kasireddy, V.; Akinci, B.; Fieguth, P. A review on computer vision based defect detection and condition assessment of concrete and asphalt civil infrastructure. Adv. Eng. Inform. 2015, 29, 196–210. [Google Scholar] [CrossRef] [Green Version]
  3. Qayyum, A.; Anwar, S.M.; Awais, M.; Majid, M. Medical image retrieval using deep convolutional neural network. Neurocomputing 2017, 266, 8–20. [Google Scholar] [CrossRef] [Green Version]
  4. Buch, N.; Velastin, S.A.; Orwell, J. A review of computer vision techniques for the analysis of urban traffic. IEEE Trans. Intell. Transp. Syst. 2011, 12, 920–939. [Google Scholar] [CrossRef]
  5. Cheng, Z.; Bai, F.; Xu, Y.; Zheng, G.; Pu, S.; Zhou, S. Focusing attention: Towards accurate text recognition in natural images. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 5076–5084. [Google Scholar]
  6. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 81–88. [Google Scholar]
  7. Lyu, G.; Huang, H.; Yin, H.; Luo, S.; Jiang, X. A novel visual perception enhancement algorithm for high-speed railway in the low light condition. In Proceedings of the 2014 12th International Conference on Signal Processing (ICSP), Hangzhou, China, 19–23 October 2014; pp. 1022–1025. [Google Scholar]
  8. Cho, Y.; Jeong, J.; Kim, A. Model-assisted multiband fusion for single image enhancement and applications to robot vision. IEEE Robot. Autom. Lett. 2018, 3, 2822–2829. [Google Scholar]
  9. Ibrahim, H.; Kong, N.S.P. Brightness preserving dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  10. Abdullah-Al-Wadud, M.; Kabir, M.H.; Dewan, M.A.A.; Chae, O. A dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  11. Arici, T.; Dikbas, S.; Altunbasak, Y. A histogram modification framework and its application for image contrast enhancement. IEEE Trans. Image Process. 2009, 18, 1921–1935. [Google Scholar] [CrossRef]
  12. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vision, Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  13. Pisano, E.D.; Zong, S.; Hemminger, B.M.; DeLuca, M.; Johnston, R.E.; Muller, K.; Braeuning, M.P.; Pizer, S.M. Contrast limited adaptive histogram equalization image processing to improve the detection of simulated spiculations in dense mammograms. J. Digit. Imaging 1998, 11, 193–200. [Google Scholar] [CrossRef] [Green Version]
  14. Celik, T.; Tjahjadi, T. Contextual and variational contrast enhancement. IEEE Trans. Image Process. 2011, 20, 3431–3441. [Google Scholar] [CrossRef] [Green Version]
  15. Lee, C.; Lee, C.; Kim, C.S. Contrast enhancement based on layered difference representation of 2D histograms. IEEE Trans. Image Process. 2013, 22, 5372–5384. [Google Scholar] [CrossRef] [PubMed]
  16. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef] [PubMed]
  17. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef]
  18. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [Green Version]
  19. Wang, S.; Zheng, J.; Hu, H.M.; Li, B. Naturalness preserved enhancement algorithm for non-uniform illumination images. IEEE Trans. Image Process. 2013, 22, 3538–3548. [Google Scholar] [CrossRef] [PubMed]
  20. Fu, X.; Zeng, D.; Huang, Y.; Liao, Y.; Ding, X.; Paisley, J. A fusion-based enhancing method for weakly illuminated images. Signal Process. 2016, 129, 82–96. [Google Scholar] [CrossRef]
  21. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  22. Kimmel, R.; Elad, M.; Shaked, D.; Keshet, R.; Sobel, I. A variational framework for retinex. Int. J. Comput. Vis. 2003, 52, 7–23. [Google Scholar] [CrossRef]
  23. Fu, X.; Zeng, D.; Huang, Y.; Ding, X.; Zhang, X.P. A variational framework for single low light image enhancement using bright channel prior. In Proceedings of the 2013 IEEE Global Conference on Signal and Information Processing, Austin, TX, USA, 3–5 December 2013; pp. 1085–1088. [Google Scholar]
  24. Park, S.; Yu, S.; Moon, B.; Ko, S.; Paik, J. Low-light image enhancement using variational optimization-based retinex model. IEEE Trans. Consum. Electron. 2017, 63, 178–184. [Google Scholar] [CrossRef]
  25. Fu, G.; Duan, L.; Xiao, C. A hybrid L2-Lp variational model for single low-light image enhancement with bright channel prior. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1925–1929. [Google Scholar]
  26. Zhang, Y.; Di, X.; Zhang, B.; Wang, C. Self-supervised image enhancement network: Training with low light images only. arXiv 2020, arXiv:2002.11300. [Google Scholar]
  27. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef] [Green Version]
  28. Tao, L.; Zhu, C.; Xiang, G.; Li, Y.; Jia, H.; Xie, X. LLCNN: A convolutional neural network for low-light image enhancement. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar]
  29. Ignatov, A.; Kobyshev, N.; Timofte, R.; Vanhoey, K.; Van Gool, L. Dslr-quality photos on mobile devices with deep convolutional networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3277–3285. [Google Scholar]
  30. Shen, L.; Yue, Z.; Feng, F.; Chen, Q.; Liu, S.; Ma, J. Msr-net: Low-light image enhancement using deep convolutional network. arXiv 2017, arXiv:1711.02488. [Google Scholar]
  31. Gharbi, M.; Chen, J.; Barron, J.T.; Hasinoff, S.W.; Durand, F. Deep bilateral learning for real-time image enhancement. ACM Trans. Graph. 2017, 36, 1–12. [Google Scholar] [CrossRef]
  32. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  33. Wang, W.; Wei, C.; Yang, W.; Liu, J. Gladnet: Low-light enhancement network with global awareness. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 751–755. [Google Scholar]
  34. Chen, C.; Chen, Q.; Xu, J.; Koltun, V. Learning to see in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3291–3300. [Google Scholar]
  35. Chen, Y.S.; Wang, Y.C.; Kao, M.H.; Chuang, Y.Y. Deep photo enhancer: Unpaired learning for image enhancement from photographs with gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6306–6314. [Google Scholar]
  36. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  37. Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6849–6857. [Google Scholar]
  38. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  39. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From fidelity to perceptual quality: A semi-supervised approach for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3063–3072. [Google Scholar]
  40. Lv, F.; Liu, B.; Lu, F. Fast enhancement for non-uniform illumination images using light-weight CNNs. In Proceedings of the 28th ACM International Conference on Multimedia, Seattle, WA, USA, 12–16 October 2020; pp. 1450–1458. [Google Scholar]
  41. Wang, L.W.; Liu, Z.S.; Siu, W.C.; Lun, D.P. Lightening network for low-light image enhancement. IEEE Trans. Image Process. 2020, 29, 7984–7996. [Google Scholar] [CrossRef]
  42. Zhu, M.; Pan, P.; Chen, W.; Yang, Y. Eemefn: Low-light image enhancement via edge-enhanced multi-exposure fusion network. Proc. Aaai Conf. Artif. Intell. 2020, 34, 13106–13113. [Google Scholar] [CrossRef]
  43. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  44. Li, J.; Feng, X.; Hua, Z. Low-light image enhancement via progressive-recursive network. IEEE Trans. Circ. Syst. Video Technol. 2021, 31, 4227–4240. [Google Scholar] [CrossRef]
  45. Zhang, F.; Li, Y.; You, S.; Fu, Y. Learning temporal consistency for low light video enhancement from single images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4967–4976. [Google Scholar]
  46. Fu, Y.; Hong, Y.; Chen, L.; You, S. LE-GAN: Unsupervised low-light image enhancement network using attention module and identity invariant loss. Knowl. Based Syst. 2022, 240, 108010. [Google Scholar] [CrossRef]
  47. Zhao, Z.; Xiong, B.; Wang, L.; Ou, Q.; Yu, L.; Kuang, F. Retinexdip: A unified deep framework for low-light image enhancement. IEEE Trans. Circ. Syst. Video Technol. 2021, 32, 1076–1088. [Google Scholar] [CrossRef]
  48. Liu, S.; Long, W.; He, L.; Li, Y.; Ding, W. Retinex-based fast algorithm for low-light image enhancement. Entropy 2021, 23, 746. [Google Scholar] [CrossRef] [PubMed]
  49. Liang, H.; Yu, A.; Shao, M.; Tian, Y. Multi-Feature Guided Low-Light Image Enhancement. Appl. Sci. 2021, 11, 5055. [Google Scholar] [CrossRef]
  50. Li, Q.; Wu, H.; Xu, L.; Wang, L.; Lv, Y.; Kang, X. Low-light image enhancement based on deep symmetric encoder—decoder convolutional networks. Symmetry 2020, 12, 446. [Google Scholar] [CrossRef] [Green Version]
  51. Han, S.; Lee, T.B.; Heo, Y.S. Deep Image Prior for Super Resolution of Noisy Image. Electronics 2021, 10, 2014. [Google Scholar] [CrossRef]
  52. Ai, S.; Kwon, J. Extreme low-light image enhancement for surveillance cameras using attention U-Net. Sensors 2020, 20, 495. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  53. Zhao, B.; Gong, X.; Wang, J.; Zhao, L. Low-Light Image Enhancement Based on Multi-Path Interaction. Sensors 2021, 21, 4986. [Google Scholar] [CrossRef] [PubMed]
  54. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  55. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  56. Ulyanov, D.; Vedaldi, A.; Lempitsky, V. Deep image prior. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 9446–9454. [Google Scholar]
  57. Ng, M.K.; Wang, W. A total variation model for Retinex. SIAM J. Imaging Sci. 2011, 4, 345–365. [Google Scholar] [CrossRef]
  58. Yue, H.; Yang, J.; Sun, X.; Wu, F.; Hou, C. Contrast enhancement based on intrinsic image decomposition. IEEE Trans. Image Process. 2017, 26, 3981–3994. [Google Scholar] [CrossRef]
  59. Lee, C.; Lee, C.; Lee, Y.Y.; Kim, C.S. Power-constrained contrast enhancement for emissive displays based on histogram equalization. IEEE Trans. Image Process. 2011, 21, 80–93. [Google Scholar]
  60. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind" image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
  61. Gu, K.; Lin, W.; Zhai, G.; Yang, X.; Zhang, W.; Chen, C.W. No-reference quality metric of contrast-distorted images based on information maximization. IEEE T. Cybern. 2017, 47, 4559–4565. [Google Scholar] [CrossRef] [PubMed]
  62. Gu, K.; Tao, D.; Qiao, J.F.; Lin, W. Learning a no-reference quality assessment model of enhanced images with big data. IEEE Trans. Neural Netw. Learn. Syst. 2017, 29, 1301–1313. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2782–2790. [Google Scholar]
Figure 1. Overall framework of our method, where z 1 , 2 , 3 N ( 0 , σ 2 ) signify Gaussian noise. DIP1 is used to generate noise N, and DIP2 and DIP3 are used to generate the reflectance R and the latent illumination I. l r e c , l i c , l r e f , and l i s represent reconstruction loss, illumination-consistency loss, reflectance loss, and illumination smoothness loss, respectively. I 0 is the initial illumination, S 0 is the input image, and S is the enhanced image.
Figure 1. Overall framework of our method, where z 1 , 2 , 3 N ( 0 , σ 2 ) signify Gaussian noise. DIP1 is used to generate noise N, and DIP2 and DIP3 are used to generate the reflectance R and the latent illumination I. l r e c , l i c , l r e f , and l i s represent reconstruction loss, illumination-consistency loss, reflectance loss, and illumination smoothness loss, respectively. I 0 is the initial illumination, S 0 is the input image, and S is the enhanced image.
Sensors 22 05593 g001
Figure 2. Results of each step of our method: (a,f) are the input image and enhanced result; (be) represent illumination I, reflectance R, noise N, and R + N .
Figure 2. Results of each step of our method: (a,f) are the input image and enhanced result; (be) represent illumination I, reflectance R, noise N, and R + N .
Sensors 22 05593 g002
Figure 3. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method yields natural exposure and clear details.
Figure 3. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method yields natural exposure and clear details.
Sensors 22 05593 g003
Figure 4. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method enhances the image and the edges are clearly visible.
Figure 4. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method enhances the image and the edges are clearly visible.
Sensors 22 05593 g004
Figure 5. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method does not have the problems of overexposure and artifacts.
Figure 5. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method does not have the problems of overexposure and artifacts.
Sensors 22 05593 g005
Figure 6. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method improves the contrast effectively and maintains the natural color.
Figure 6. Comparisons of enhanced images. Red boxes indicate the obvious differences. Compared with other methods, our method improves the contrast effectively and maintains the natural color.
Sensors 22 05593 g006
Table 1. Comparison of average NIQE on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
Table 1. Comparison of average NIQE on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
MethodDICMFusionLIMEMEFNPEVVAverage
LIME3.53603.91834.14233.70224.26252.74753.5442
NPE3.45303.88833.90313.51553.95013.02903.4928
SRIE3.57683.97413.78683.47423.98833.13573.5668
KinD4.26914.10274.35254.13183.95893.42554.0752
Zero-DCE3.60914.24213.93543.40444.09443.22453.6332
RetinexDIP3.76124.23083.63553.27214.10122.48903.5363
Ours3.79114.06283.76153.23634.04262.46043.5294
Table 2. Comparison of average NIQMC on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
Table 2. Comparison of average NIQMC on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
MethodDICM FusionLIMEMEFNPEVVAverage
LIME5.33975.36865.49565.41685.44805.58055.4121
NPE5.08954.58024.61684.86105.17385.26555.0104
SRIE4.99904.35684.50324.70455.18485.30214.9246
KinD4.61554.52484.68414.67254.57664.81594.6511
Zero-DCE4.89844.73465.06785.05045.10685.36145.0062
RetinexDIP4.99124.44494.78305.01515.32225.39155.0126
Ours5.00934.52104.79965.07615.29315.41385.0398
Table 3. Comparison of average CPCQI on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
Table 3. Comparison of average CPCQI on six datasets. (The red, green, and blue scores represent the top three in the corresponding dataset, respectively).
MethodDICM FusionLIMEMEFNPEVVAverage
LIME0.89860.96421.08821.03850.98440.95550.9515
NPE0.91390.97051.08121.03721.02280.95570.9609
SRIE0.90561.00941.11211.09671.02580.96290.9721
KinD0.74590.81480.83360.78770.80070.74180.7670
Zero-DCE0.78180.88200.98030.94610.85780.83960.8415
RetinexDIP0.99991.06801.15951.10881.04111.05251.0436
Ours1.00381.07871.15851.09261.05241.04451.0437
Table 4. Runtime (RT) comparison (in seconds).
Table 4. Runtime (RT) comparison (in seconds).
Method( 640 × 480 )( 1280 × 960 )( 1920 × 1440 )( 2560 × 1920 )( 3200 × 2400 )( 3840 × 2880 )( 4480 × 3360 )( 5120 × 3840 )
LIME0.11330.41961.01481.57132.39013.33024.40585.7054
NPE5.886126.634058.5019104.8345163.9938235.7513326.0996427.1531
SRIE4.764333.6684121.5802343.9839726.5981386.7066544.0660865.0404
KinD0.15540.0464------
Zero-DCE0.125590.13900.25390.40510.83371---
RetinexDIP15.248215.594531.457552.1564102.5126139.4568182.4861212.1594
Ours15.365515.491030.813148.952794.5498122.6732154.7097189.1483
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, X.; Zhang, M.; Luo, J. Low-Light Image Enhancement via Retinex-Style Decomposition of Denoised Deep Image Prior. Sensors 2022, 22, 5593. https://doi.org/10.3390/s22155593

AMA Style

Gao X, Zhang M, Luo J. Low-Light Image Enhancement via Retinex-Style Decomposition of Denoised Deep Image Prior. Sensors. 2022; 22(15):5593. https://doi.org/10.3390/s22155593

Chicago/Turabian Style

Gao, Xianjie, Mingliang Zhang, and Jinming Luo. 2022. "Low-Light Image Enhancement via Retinex-Style Decomposition of Denoised Deep Image Prior" Sensors 22, no. 15: 5593. https://doi.org/10.3390/s22155593

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop