Next Article in Journal
Impact of Rainfall on the Detection Performance of Non-Contact Safety Sensors for UAVs/UGVs
Previous Article in Journal
Enhanced Lightweight YOLOX for Small Object Wildfire Detection in UAV Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Color and Luminance Separated Enhancement for Low-Light Images with Brightness Guidance

Key Laboratory of Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2711; https://doi.org/10.3390/s24092711
Submission received: 27 March 2024 / Revised: 18 April 2024 / Accepted: 22 April 2024 / Published: 24 April 2024

Abstract

:
Existing retinex-based low-light image enhancement strategies focus heavily on crafting complex networks for Retinex decomposition but often result in imprecise estimations. To overcome the limitations of previous methods, we introduce a straightforward yet effective strategy for Retinex decomposition, dividing images into colormaps and graymaps as new estimations for reflectance and illumination maps. The enhancement of these maps is separately conducted using a diffusion model for improved restoration. Furthermore, we address the dual challenge of perturbation removal and brightness adjustment in illumination maps by incorporating brightness guidance. This guidance aids in precisely adjusting the brightness while eliminating disturbances, ensuring a more effective enhancement process. Extensive quantitative and qualitative experimental analyses demonstrate that our proposed method improves the performance by approximately 4.4 % on the LOL dataset compared to other state-of-the-art diffusion-based methods, while also validating the model’s generalizability across multiple real-world datasets.

1. Introduction

Low-light enhancement is a crucial yet challenging aspect of image processing. Under low-light conditions, images are deteriorated by reduced brightness and a poor signal-to-noise ratio (SNR), obscuring details and compromising the quality. This issue, primarily due to sensor limitations in capturing weak light, results in noise that degrades the image and can be intensified during enhancement, causing color distortions. Therefore, it is crucial to develop methods to improve the visibility and quality.
Various convolutional neural networks (CNNs) [1,2,3,4,5,6,7,8] have been proposed in the field of low-light image enhancement, many of which are based on Retinex theory [9]. This theory guides the separation of images into reflectance and illumination maps. Diverse methods [10,11,12] leverage Transformer models to restore the decomposed maps. However, these approaches cannot be developed further due to the limited representation capabilities of CNN-based models.
Recently, diffusion models have become increasingly important in image enhancement tasks [13], including low-light enhancement [14,15,16,17,18,19,20], due to their excellent performance in modeling complex noise and artifact distributions in images. Ref. [14] employs early downsampling and a global corrector to accelerate processing and mitigate color shifts. Ref. [16] also uses downsampling and gradual upsampling, but with a unique normalization strategy. Ref. [20] incorporates prior knowledge, using color maps to represent image color information. Refs. [15,17] introduce novel components; Ref. [15] reduces the input size through a wavelet transform and maintains details with a high-frequency restoration module, while [17] enhances the effectiveness by feeding degradation features from a designed DGNET to a U-Net in the diffusion model. However, these methods are susceptible to color shifts, making the design of complex modules for color adjustment necessary.
Alternatively, some works [18,19] attempt to combine the diffusion model with Retinex theory, which mitigates the issue of color shifts. Ref. [18] uses a Transformer-based structure for decomposition and conditional diffusion models to restore reflectance and illumination maps. Ref. [19] follows a similar approach but improves the decomposition method and includes a refined module for enhanced color and detail restoration. Ref. [21], on the other hand, uses the diffusion model to extract auxiliary features during Retinex decomposition, integrating these into the RGformer network for enhancement. Ref. [20] indicates that using low-light images and normal-light images with mixed noise directly as inputs to conditional diffusion models yields subpar results, leading to the introduction of additional prior knowledge like color maps and illumination embedding. However, these methods utilize learning-based networks to predict estimates of the reflectance and illumination maps, and such methods are inaccurate while causing the loss of some image structure information.
In this paper, we present a simple but effective image decomposition strategy as a unique paradigm for Retinex decomposition. Specifically, we utilize the grayscaled image as the illumination map; then, we divide the RGB channels of the input image pixel by pixel by the grayscaled image to obtain the reflectance map, and we enhance the reflectance and illumination maps separately using a diffusion model. Furthermore, we introduce brightness guidance to guide the brightness adjustment module to better learn the illumination pixel distribution of the reference images. Extensive experiments demonstrate that our model performs favorably against existing state-of-the-art methods.
The main contributions of this work are as follows:
  • We propose a simple but effective image decomposition strategy, which can serve as a unique paradigm for Retinex decomposition;
  • We introduce brightness guidance to guide the brightness adjustment and remove the disturbance of the diffusion model;
  • We conduct extensive experiments on a benchmark dataset to demonstrate the feasibility of our proposed method.

2. Related Work

2.1. Traditional Methods

Traditional low-light image enhancement methods can be broadly categorized into three types: gamma correction (GC) [22], histogram equalization (HE) [23], and Retinex theory [9]. Bennet et al. [24] proposed employing bilateral filtering to decompose low-light observations, followed by applying gamma curve methods with different parameter settings to process the decomposed layers and then recombining them. Additionally, Yuan et al. [25] attempted to perform gamma curve operations on each sub-region generated through segmentation. Furthermore, Rahman et al. [26] introduced an adaptive gamma correction method, which dynamically determines the intensity transformation function based on the statistical characteristics of the image. Overall, the existing gamma correction-based methods still suffer from the largest problem of uneven exposure phenomena. Pizer et al. [27] proposed adaptive histogram equalization (AHE) to map the histograms of local regions to simple mathematical distributions. Building upon the principles of human visual locality, Pizer et al. [23] further introduced contrast-limited adaptive histogram equalization (CLAHE). Jobson et al. [28] made some initial attempts based on Retinex theory, estimating illumination through filter introduction, but obtained results that deviated from the distribution of real natural images, leading to unknown artifacts and color biases. With deeper exploration, a series of works [29,30,31] have focused on designing regularization terms for variables within the Retinex model to assist in estimating accurate target variables.

2.2. Homomorphic-Based Methods

Homomorphic filtering is a technique that operates in the frequency domain to separate illumination and reflectance components in images. This method is particularly advantageous in enhancing low-light images, where non-uniform lighting conditions prevail. By transforming the multiplicative relationships of these components into additive ones via a logarithmic domain, homomorphic filtering allows for the differential manipulation of illumination and reflectance, thus enhancing the visibility and details in dark regions while preserving the quality of well-lit sections. Sobbahi et al. [32] present a novel approach by embedding homomorphic filtering within a deep learning architecture. The model not only enhances low-light images but also tailors the enhancement process to improve subsequent image classification tasks. This integration demonstrates the dual benefit of image preprocessing for visual improvements and performance boosts in machine learning applications. Chavarín et al. [33] integrate cluster-chaotic optimization with homomorphic filtering. The chaotic optimization approach adjusts the filter parameters dynamically, optimizing the contrast enhancement process. The adaptation is guided by the peculiarities of the image content, leading to superior enhancement results compared to static parameter selection. While effective, homomorphic-based approaches face challenges such as noise sensitivity, computational complexity, parameter sensitivity, poor handling of non-uniform illumination, complex integration with advanced models, and reliance on specific illumination models.

2.3. Learning-Based Methods

Compared to other low-level vision tasks, the development of deep learning-based low-light image enhancement progressed relatively slowly until the advent of LL-Net in 2017 [34]. In 2018, Chen et al. [35] pioneered the development of a series of paired datasets with weak light input and normal exposure labels using long- and short-exposure shots, thereby propelling the advancement of deep network technologies for low-light image enhancement. Subsequently, methods based on deep learning gradually became mainstream.
The integration of Retinex theory with deep learning for low-light image enhancement was first proposed by Shen et al. [36]. They combined convolutional neural networks with Retinex theory, treating multiscale Retinex as a cascade of Gaussian convolutions with skip connections or in residual form, and designed a multiscale convolutional neural network, MSR-Net, based on paired data processed with Photoshop. The logarithmic transformation was used in the network to convert the Retinex model from a multiplication form to an addition form. However, this method tends to lose details due to the logarithmic transformation, which suppresses gradient changes in bright areas.
Subsequent works introduced Retinex theory into deep neural networks [11,18,35,37,38,39,40,41,42]. Among them, Retinex-Net [35] is the most inspiring method; it combines physical models and deep neural networks (DNNs). Following this, Zhang et al. proposed KinD [43] and KinD++ [38], offering more effective solutions. Unlike complex multi-stage training processes, Fu et al. [41] and Cai et al. [11] explored the possibility of end-to-end frameworks, achieving significant performance improvements.
According to the current literature statistics, almost one third of methods combine the design of deep networks with Retinex theory—for instance, designing different sub-networks to estimate the components of the Retinex model and estimating illumination maps to guide network learning. Although such combinations can integrate deep learning-based methods with traditional methods, their respective weaknesses may also be introduced into the final model: (1) the ideal assumption in Retinex-based low-light image enhancement methods, i.e., that reflectance is the final enhancement result, still affects the final outcome; (2) despite the use of Retinex theory, the risk of overfitting in deep networks still exists.
In cases where Retinex theory is not used, recent works have focused on directly sampling end-to-end methods [44,45,46,47,48,49,50,51] for low-light image enhancement. LLNet, proposed by Lore et al. [34], inspired the emergence of end-to-end methods, mainly showcasing the potential of supervised methods in enhancement. To mitigate color biases, some methods [52,53] employ three-dimensional look-up tables (3D-LUT) and histograms to maintain color consistency. In other methods [49,50], attempts have been made to use signal-to-noise ratio (SNR) perception priors and structure-aware features as guidance to produce realistic results. Recently, with the popularity of ultra-high-definition (UHD) images, methods such as LLformer, proposed by Wang et al. [12], and UHDFour, proposed by Li et al. [54], have been used to enhance UHD images, and related UHD datasets have been released to promote further research. Additionally, when training data are limited, semi-supervised [55,56], unsupervised [42,57], and zero-shot learning [58,59] methods are valuable research directions and important branches of deep learning-based low-light image enhancement.

3. Methodology

3.1. Separated Enhancement

The proposed framework seeks to enhance the brightness of low-light images, drawing upon the foundational principles of Retinex theory [9], as shown in Figure 1. Central to this theoretical framework is the premise that low-light images can be systematically decomposed into two components: reflectance maps and illumination maps. The decomposition process can be formulated as follows:
I = R L ,
where ∘ designates the element-wise product, I represents the input low-light images, and R and L stand for the reflectance and illumination maps.
As indicated by [11], while a clean image is free from corruption, applying Retinex decomposition to low-light images results in reflectance and illumination maps that are marred by noise and artifacts, further complicating their estimation. Mathematically, a degraded low-light image can be naturally modeled as follows:
I = ( R + R ^ ) ( L + L ^ ) ,
where R ^ and L ^ are the polluted terms that perturbate the reflectance and illumination maps, respectively.
After the initial decomposition phase, the process of enhancing low-light images involves the application of element-wise multiplication using a ‘light-up map’ L ¯ . This map is utilized to intensify the brightness of the input low-light image I. The mathematical formulation of this enhancement process can be expressed as follows:
I L ¯ = ( R + R ^ ) ( L + L ^ ) L ¯ = ( R + R ^ ) ( L L ¯ + L ^ L ¯ ) .
According to the above theory, decomposing the images in a Retinex manner consequently decouples the removal of the perturbation terms R ^ and L ^ L ¯ into two separate subspaces, allowing them to be better and more easily regularized/learned.

3.2. Retinex Decomposition Strategy

While we can recover low-light images by ensuring the better removal of perturbation terms in both the reflectance and luminance subspaces, how the Retinex is decomposed is still a fundamentally ill-posed problem, as indicated by a range of previous studies [28,35,43,60,61,62,63]. Some of them [28,61,62] have employed traditional methods to decompose low-light images into reflectance and illumination maps, which utilize channel-wise Gaussian blur to compute the illumination maps, subsequently obtaining the reflectance maps through the channel-wise division of the original image by the illumination maps. Meanwhile, learning-based methods typically involve the use of intricately designed deep neural networks (DNNs) to generate initial estimates of the reflectance and illumination maps, as seen in [35,38,43], often supplemented by hand-crafted constraints and priors [64,65,66,67,68,69], although these may be constrained by the model’s capacity.
Although these methods have marked significant advancements in the estimation of reflectance and illumination maps, achieving precise outcomes remains a challenge. Ref. [11] have demonstrated that while a clean image is free from corruption, the application of Retinex decomposition to low-light images results in reflectance and illumination maps that are marred by noise and artifacts, further complicating their estimation. Addressing this challenge, this study diverges from the pursuit of increasingly complex networks for more accurate initial estimations. Instead, we introduce a novel decomposition strategy that employs a fixed pattern, decomposing images into a graymap and colormap as the initial estimation of the reflectance and illumination maps, bearing similarity to [20,47], but with notable distinctions. This innovative strategy presents a unique paradigm for the problem of Retinex decomposition.
Graymap: We execute a weighted summation of the RGB channels, prioritizing the green (g) channel, followed by the red (r) and blue (b) channels, in descending order according to the weights set by the color space standards [70]. The formulation of this function is expressed as follows:
L = 0.299 × r + 0.587 × g + 0.114 × b .
For the graymap, as illustrated in Figure 2b,e, our strategy deviates from the method in [47], which calculates the mean across the RGB channels. Recognizing the human eye’s heightened sensitivity to green, we assign greater weight to this channel. Notably, this weighting ensures that brighter pixels in the original image retain their prominence in the weighted sum, thereby preserving the strong correlation between the illumination map and the original image’s brightness. This characteristic renders the map suitable for use as the illumination component in the Retinex model.
Colormap: With the acquisition of the graymap, we can obtain the corresponding colormap by performing the pixel-wise division of each RGB channel by the graymap as follows:
R = r L , g L , b L .
For the colormap, as illustrated in Figure 2c,f, the reflectance map exhibits a notable degree of consistency, even amidst varying luminance levels. This steadfast characteristic positions it as an apt candidate for the newly estimated reflectance map within the Retinex framework, leveraging its stability across different lighting conditions.

3.3. Conditional Diffusion Model

Conditional diffusion models are commonly used in image enhancement [13,71,72]. These models enhance images by inputting the degraded image as conditional information into a noise prediction network, guiding the diffusion model to generate an enhanced image corresponding to the degraded one.
Specifically, the forward process systematically introduces Gaussian noise into the clean image X 0 . As delineated in [73], this Gaussian diffusion forward process incrementally contaminates the clean data X 0 through a sequential T diffusion time step mechanism, thereby enabling the acquisition of the sampled intermediate state X t :
q ( X t | X t 1 ) = N ( X t ; α t X t 1 , ( 1 α t ) ϵ t ) , ϵ t N ( 0 , I ) .
where α t = 1 β t , β t represents the variance schedule, and X t 1 is the intermediate state of the previous sampling steps. Additionally, ϵ t , drawn from a normal distribution N ( 0 , 1 ) , possesses the same dimensionality as the input data X 0 .
Given that α ^ t = i = 1 t α i , the equation describing the aforementioned process can be simplified as follows:
X t = α ^ t X t 1 + 1 α ^ t ϵ t , ϵ t N ( 0 , I ) .
The reverse process in diffusion models constitutes a denoising procedure, wherein the model is trained to effectively reconstruct a clean original signal from noisy data. Specifically, sampling is conducted using the Gaussian transitions p θ ( X ^ t 1 | X ^ t , X ˜ ) , which are parameterized by learned parameters. This process initiates from X ^ T N ( 0 , I ) through the following mechanism:
X ^ t 1 = 1 α t X ^ t 1 α t 1 α ¯ t ϵ θ ( X ^ t , t , X ˜ ) ,
where X ^ t represents the sampled random Gaussian noise, X ^ t 1 denotes the intermediate result following one step of the denoising process, and X ˜ is the conditional guide parameter. In this study, we select the input low-light image to serve as the guide parameter.
Given its efficacy in addressing complex degradation patterns, this study utilizes a conditional diffusion model, specifically a typical patch-based conditional diffusion model [13], to effectively eliminate perturbation terms in the reflectance and illumination maps. The details of this framework are illustrated in Figure 1.

3.4. Brightness Adjustment Module

Given that the graymap effectively isolates color information, the primary objective is to enhance the brightness while concurrently eliminating perturbations. As depicted in Figure 2, the graymap of a low-light image typically exhibits a markedly low pixel intensity, leading to the significant loss of visible details. Consequently, it becomes imperative to augment the pixel values within the graymap. However, this enhancement process also tends to amplify the noise, thereby exacerbating the perturbations, and, consequently, the process of luminance enhancement can adversely impact the perturbation removal phase.
Previous research [20] has indicated that utilizing a low-light image directly as a conditional input does not produce optimal outcomes, underscoring the inherent challenges and inefficiencies in simultaneously achieving brightness enhancement and noise removal. To tackle this issue, we adopt a two-step approach, initially focusing on brightness adjustment, followed by the removal of perturbations. This methodology ensures that each aspect of image improvement is addressed effectively without adversely affecting the other.
Brightness Guidance: Recent studies [18,35,43] have often overlooked the complex interrelation between brightness enhancement and perturbation removal. Contrarily, our approach prioritizes brightness adjustment through brightness guidance. This strategy effectively elevates the brightness of low-light graymaps, albeit with a consequent increase in noise and artifacts. The subsequent use of a conditional diffusion model, therefore, concentrates exclusively on removing these noise and artifacts. This focused approach, by distinctly separating the tasks of brightness adjustment and disturbance removal, results in a more efficient enhancement process.
The utilization of brightness guidance is tailored differently for the training and inference phases. In the training phase, paired images enable the use of normal-light images for brightness reference. In contrast, the inference phase, lacking paired images, leverages a pre-trained network to adjust the low-light graymaps towards normal light, thus providing brightness guidance. The focus here is on matching the brightness levels, primarily using Gaussian-blurred reference graymaps, rather than preserving fine details. This approach ensures brightness alignment with the reference, making it an effective strategy for brightness guidance despite the potential limitations in detail retention.
The method for brightness adjustment comprises the following steps.
  • Mean Gray Value Calculation: Compute the mean gray value g 1 for blocks in low-light images and g 2 for blocks in normal light or as determined by the learned light model.
  • Brightening Coefficient Determination: Establish the brightening coefficient γ using the formula γ = g 2 g 1 . This coefficient represents the factor for the enhancement of the brightness of the low-light graymap to produce the final conditional image. This approach ensures that the conditional image matches the guidance image in terms of the brightness level.

4. Experiments

We employ the Adam optimizer [74] to train the proposed diffusion model, with the parameter configurations as follows: the initial learning rate is set to 2 × 10 5 , and no weight decay is applied. In the parameter updating process, exponential moving averages are utilized with a weight of 0.999 to promote more stable learning. For an RGB image, it is randomly cropped into image patches of size 64 × 64 . Xavier’s method [75] is employed for the random initialization of the parameters in each module of the network.
We employ the PyTorch [76] deep learning framework to implement the training and testing processes of the neural networks. Throughout the network training, the total number of iterations is 960k, with the learning rate remaining constant throughout. Additionally, to expedite the sampling phase, the DDIM [77] training method is adopted, where the final augmentation results are obtained after every 15 iterations.
We train and evaluate the proposed model on the LOL [35] dataset, LOLv2-real [78] dataset, and LOLv2-syn dataset [78]. The LOLv2-real dataset comprises 689 pairs of low-/normal-luminance paired images collected from real scenarios, including 689 training pairs and 100 testing pairs, with an image resolution of 600 × 400 . The LOLv2-syn dataset consists of 1000 pairs of synthetically generated low-/normal-luminance paired images, also with a resolution of 600 × 400 . The batch processing approach is employed to feed the training data into the neural network, with each batch containing 16 pairs of samples. The entire experiment is conducted on an NVIDIA 1080Ti GTX GPU, and the training of the proposed network model takes approximately two days to fully converge.

5. Results

5.1. Comparison of Results on Real Datasets

We conduct the testing of the proposed method on low-light images captured in real-world scenarios and visually compare its effectiveness with that of other algorithms. Figure 3 and Figure 4, respectively, depict the enhancement results of real-scene images captured from two different datasets, the LOL dataset and the LOLv2-real dataset. It is evident from the figures that the proposed method exhibits significant visual advantages over the state-of-the-art algorithms. Previous methods exhibit various shortcomings: the Retinexnet [35] method results in color distortion; the Retinexformer [11] method encounters difficulties in over-/underexposed regions and noise suppression; the DiffLL [15] and CLE [20] methods produce blurred areas; and the Kind++ [38] method introduces unnatural artifacts and edge distortions. In contrast, the proposed method has achieved significant success in enhancing the image clarity, as evidenced by the clearer visibility of the text in Figure 3. The sharpening of text edges and the enhancement in contrast have been effectively applied. In terms of brightness restoration, as shown in Figure 4, our method also demonstrates its advantages, effectively recovering the brightness information of the enhanced image to levels closer to the reference image. However, in terms of color restoration, although our method provides satisfactory results in most cases, it falls slightly short in terms of color accuracy and richness compared to the CLE method in certain specific scenes.
In order to better comprehend the effectiveness of the proposed method and other methods in enhancing real-scene images, various quantitative evaluation metrics, such as the peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and learned perceptual image patch similarity (LPIPS), were introduced. As depicted in Table 1 and Table 2, the quantitative metrics of the proposed method on both real datasets reached the state-of-the-art level.

5.2. Comparison of Results on Synthetic Datasets

We utilize the LOLv2-syn dataset [78] to evaluate the enhancement efficacy of the proposed method and compare it with existing state-of-the-art algorithms. In the quantitative comparison of the results, we calculate the peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and learned perceptual image patch similarity (LPIPS) between the enhanced images and reference normal luminance images to quantitatively assess the enhancement quality of the methods. As depicted in Table 3, the proposed method exhibits higher objective evaluation metrics on the simulated dataset compared to the current state-of-the-art low-light image enhancement algorithms. Figure 5 displays a comparison of the enhancement results on an outdoor scene low-light image among various methods in the dataset. The proposed method effectively recovers the color fidelity of the image, rendering it more akin to that of a reference image, exemplified by elements such as the wall of a house. Additionally, the brightness levels and sharpness of the enhanced image align closely with those observed in the reference image, demonstrating the method’s capacity to maintain consistency in key visual parameters.
It is noteworthy that the benchmark model trained by the proposed method is based on a diffusion model. The performance improvement of the proposed method on the simulated dataset compared to the current diffusion models reaches 1.73 dB. Furthermore, on the previously tested real datasets, the improvement is also significant, reaching 1.15 dB and 0.3 dB, respectively. This indicates that the prior brightness information proposed by us contributes significantly to the low-light image enhancement task, showcasing promising enhancement.

5.3. Generalization Ability to Real-World Images

This section aims to elucidate the generalization performance of the proposed method by conducting a comparative analysis with current state-of-the-art methods. Training is conducted using the LOL training dataset, while testing is carried out on a variety of real-world low-light datasets captured in diverse scenes.
We conducted an extensive experimental analysis utilizing two classical low-light image datasets: MEF [79], which comprises 17 test images, and VV, containing 24 test images. The efficacy of our proposed method is illustrated through three sets of detailed visual comparisons in Figure 6 and Figure 7, which provide an intuitive assessment of the visual effects. The results reveal that our method effectively enhances dark regions while preserving the color fidelity. The outcome is visually pleasing, devoid of significant noise and color casts. In contrast, wang2022low and Retinexformer do not adequately enhance the image brightness, resulting in inferior visualization outcomes. On the other hand, while DiffLL produces visually appealing results, it sometimes suffers from localized overexposure or underexposure. These observations demonstrate that our method possesses robust generalization capabilities, delivering more naturalistic image quality in real-world scenarios.
To further demonstrate the practical advantages of our method, we also performed experiments on a dataset designed specifically for object detection and recognition. For this purpose, we selected low-light images from the ExDark dataset [80] for testing. Figure 8 displays the comparative results, from which it is evident that Retinexnet is plagued by severe artifacts. wang2022low offers improved visualization, albeit with notable overexposure issues. DiffLL is marred by blurring effects in its results. Retinexformer, meanwhile, fails to adequately brighten the houses in the distance. Conversely, the results from our proposed method are visually superior, rendering the images more natural and clearer, particularly in areas such as the distant white houses and the sky.
The experimental results underscore not only the effectiveness of the proposed method but also the superior generalization performance of diffusion models compared to traditional convolutional neural networks (CNNs). As an emerging deep learning architecture, diffusion models excel in handling complex data distributions, offering significant improvements in areas like image processing. These findings highlight the potential of diffusion models to outperform traditional models in generalization capabilities, providing valuable insights for future research.

6. Ablation Study

This section will analyze and discuss the effectiveness of the novel Retinex decomposition strategy proposed in our method, followed by a discussion of the effectiveness of the prior brightness information proposed in our method.

6.1. Analysis of Differences in Image Decomposition Strategies

To rigorously evaluate our proposed Retinex decomposition strategy against traditional and contemporary methods, we conduct a series of comparative experiments. These experiments distinctly contrast our approach with the classical Single-Scale Retinex algorithm (SSR) [28] and the deep learning-based RetinexNet [35]. The aim is to validate our strategy’s effectiveness in enhancing the image quality and in detail preservation, highlighting its advancements over existing methods. The experimental conditions were standardized, except for the decomposition strategy, to ensure an accurate assessment of its efficacy.
The experimental results, as shown in Table 4, reveal the relative disadvantage of the traditional SSR method in the performance metrics. This disadvantage mainly stems from the inaccuracies of traditional decomposition methods in estimating the luminance component using Gaussian-blurred images. While Gaussian blur simplifies the representation of the image luminance, it often leads to the loss of important details, thereby affecting the naturalness and realism of the final image.
On the other hand, decomposition methods based on deep learning, although demonstrating comparable levels of learned perceptual image patch similarity (LPIPS, an index for the evaluation of the perceptual similarity between images) to our proposed approach, show a decline in performance in terms of the structural similarity index (SSIM, an index for the measurement of image quality) scores and peak signal-to-noise ratio (PSNR, an index reflecting the quality of image restoration). This outcome suggests that while deep learning-based methods are effective in handling some image issues, they may lead to the loss of image information in the encoder and decoder structures, particularly when attempting to separate delicate luminance information from the reflectance components, thereby revealing the limitations of such approaches.
To further demonstrate the effectiveness of the proposed image decomposition strategy, the contrasting enhancement results of different decomposition strategies are provided. As illustrated in Figure 9, the traditional SSR-based decomposition strategy exhibits severe image blurring issues, while the deep learning-based decomposition strategy, although capable of enhancing images reasonably well, performs comparatively poorer in terms of color and saturation compared to the proposed decomposition strategy. Therefore, the proposed decomposition strategy ensures better performance in enhancing low-light images compared to existing decomposition strategies.
Our proposed image decomposition strategy is based on straightforward operations that separate image data into luminance and chrominance components. This simplicity leads to predictable and consistent outputs, which are particularly advantageous in scenarios where interpretability and reproducibility are critical. In contrast, learning-based methods, often reliant on complex neural networks, introduce a level of opacity due to their ‘black box’ nature. Moreover, because of its non-parametric nature, this decomposition strategy does not suffer from overfitting, a common issue in learning-based methods that can detrimentally impact their generalizability to new, unseen data. In conclusion, while learning-based decomposition methods continue to evolve and offer compelling benefits in certain applications, the simplicity, efficiency, and robustness of our proposed decomposition strategy make it an effective and reliable choice in many practical scenarios.

6.2. Analysis of Effectiveness of Prior Brightness Information

The application of prior brightness information contributes to enhancing the visual quality of images, particularly in processes involving brightness adjustment and disturbance removal. Prior brightness information not only guides enhancement algorithms to adjust the image brightness more accurately but also helps to maintain the naturalness and continuity of images when removing noise and disturbances.
To deeply understand the role of prior brightness information within the framework of the method proposed in our method, a comparative experiment was designed. By excluding the prior brightness information from the enhancement process, the impact of this change on the enhancement effect was observed. The experimental results, as shown in Table 5, indicate a significant decrease in the enhancement performance when the prior brightness information is not utilized. Specifically, the enhanced images exhibit noticeable inaccuracies and discontinuities in brightness, significantly compromising the visual quality of the images.
Without prior brightness information, the tasks of brightness adjustment and disturbance removal have to be coupled together, making it difficult for the algorithm to balance the relationship between them, thereby affecting the final enhancement effect. As depicted in Figure 10, the brightness adjustment module struggles to accurately determine how to adjust the brightness of various regions in the image without guidance from prior brightness information, resulting in uneven brightness in the enhanced images. Similarly, disturbance removal becomes less effective due to the lack of prior brightness information, leading to the loss of image details or the generation of unnatural visual effects.
These comparative experimental results further emphasize the importance of prior brightness information in the image enhancement process. Prior brightness information not only helps to improve the accuracy of enhancement algorithms in brightness adjustment but also effectively guides disturbance removal, ensuring that the enhanced images maintain the natural brightness while enhancing the overall visual quality. Therefore, the guiding role of prior brightness information is crucial in ensuring fidelity and naturalness in the image enhancement process.

7. Conclusions

In this paper, we propose an adaptive brightness method. Utilizing a pretrained model, adaptive brightness information is extracted and mapped from low-light images, which is then enhanced through a brightness adjustment module. Subsequently, a conditional diffusion model is employed to mitigate the noise perturbations introduced by the brightness adjustment, thus separating the intertwined challenges of brightness enhancement and noise perturbation removal and reducing the complexity involved in enhancing the illumination component. Moreover, we introduce a simple yet effective image decomposition strategy that decomposes the image into graymaps and colormaps, serving as initial estimates similar to the illumination and reflectance components in Retinex decomposition.
To enhance the performance of low-light image enhancement, we employ diffusion models instead of convolutional neural networks to separately enhance the estimated illumination and reflectance components. For the reflectance component, due to its consistency under various lighting conditions, it is sufficient to directly use the conditional diffusion model to remove noise perturbations. In the comprehensive quantitative and qualitative analyses, our proposed method surpasses the current state-of-the-art across multiple datasets. Additionally, it demonstrates generalization capabilities through its performance on several real-world scenario datasets.

Author Contributions

Conceptualization, F.Z., X.L., C.G. and N.S.; methodology, F.Z. and X.L.; software, F.Z. and X.L.; validation, F.Z., X.L. and C.G.; formal analysis, F.Z.; investigation, F.Z.; resources, F.Z. and C.G.; data curation, F.Z. and N.S.; writing—original draft preparation, F.Z. and X.L.; writing—review and editing, C.G. and N.S.; visualization, F.Z. and X.L.; supervision, C.G. and N.S.; project administration, F.Z. and X.L.; funding acquisition, C.G. and N.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The publicly archived datasets can be found at https://daooshee.github.io/BMVC2018website/, accessed on 17 April 2024 (LOL dataset) and https://github.com/flyywh/CVPR-2020-Semi-Low-Light, accessed on 17 April 2024 (LOL-v2 dataset).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef] [PubMed]
  2. Zhao, M.; Yang, R.; Hu, M.; Liu, B. Deep Learning-Based Technique for Remote Sensing Image Enhancement Using Multiscale Feature Fusion. Sensors 2024, 24, 673. [Google Scholar] [CrossRef] [PubMed]
  3. Lu, Y.; Guo, Y.; Liu, R.W.; Ren, W. MTRBNet: Multi-branch topology residual block-based network for low-light enhancement. IEEE Signal Process. Lett. 2022, 29, 1127–1131. [Google Scholar] [CrossRef]
  4. Qiu, Y.; Chen, J.; Wang, Z.; Wang, X.; Lin, C.W. Spatio-spectral feature fusion for low-light image enhancement. IEEE Signal Process. Lett. 2021, 28, 2157–2161. [Google Scholar] [CrossRef]
  5. Lu, H.; Gong, J.; Liu, Z.; Lan, R.; Pan, X. FDMLNet: A Frequency-Division and Multiscale Learning Network for Enhancing Low-Light Image. Sensors 2022, 22, 8244. [Google Scholar] [CrossRef] [PubMed]
  6. Ma, S.; Ma, H.; Xu, Y.; Li, S.; Lv, C.; Zhu, M. A Low-Light Sensor Image Enhancement Algorithm Based on HSI Color Model. Sensors 2018, 18, 3583. [Google Scholar] [CrossRef] [PubMed]
  7. Shi, Y.; Wang, B.; Wu, X.; Zhu, M. Unsupervised low-light image enhancement by extracting structural similarity and color consistency. IEEE Signal Process. Lett. 2022, 29, 997–1001. [Google Scholar] [CrossRef]
  8. Ai, S.; Kwon, J. Extreme low-light image enhancement for surveillance cameras using attention U-Net. Sensors 2020, 20, 495. [Google Scholar] [CrossRef] [PubMed]
  9. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  10. Cui, Z.; Li, K.; Gu, L.; Su, S.; Gao, P.; Jiang, Z.; Qiao, Y.; Harada, T. You Only Need 90K Parameters to Adapt Light: A Light Weight Transformer for Image Enhancement and Exposure Correction. arXiv 2022, arXiv:2205.14871. [Google Scholar]
  11. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage Retinex-based Transformer for Low-light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–3 October 2023; pp. 12504–12513. [Google Scholar]
  12. Wang, T.; Zhang, K.; Shen, T.; Luo, W.; Stenger, B.; Lu, T. Ultra-high-definition low-light image enhancement: A benchmark and transformer-based method. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington DC, USA, 7–14 February 2023; Volume 37, pp. 2654–2662. [Google Scholar]
  13. Özdenizci, O.; Legenstein, R. Restoring vision in adverse weather conditions with patch-based denoising diffusion models. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 10346–10357. [Google Scholar] [CrossRef] [PubMed]
  14. Zhou, D.; Yang, Z.; Yang, Y. Pyramid Diffusion Models For Low-light Image Enhancement. arXiv 2023, arXiv:2305.10028. [Google Scholar]
  15. Jiang, H.; Luo, A.; Fan, H.; Han, S.; Liu, S. Low-light image enhancement with wavelet-based diffusion models. ACM Trans. Graph. (TOG) 2023, 42, 1–14. [Google Scholar] [CrossRef]
  16. Nguyen, C.M.; Chan, E.R.; Bergman, A.W.; Wetzstein, G. Diffusion in the dark: A diffusion model for low-light text recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 1–7 January 2024; pp. 4146–4157. [Google Scholar]
  17. Wang, T.; Zhang, K.; Shao, Z.; Luo, W.; Stenger, B.; Kim, T.K.; Liu, W.; Li, H. LLDiffusion: Learning degradation representations in diffusion models for low-light image enhancement. arXiv 2023, arXiv:2307.14659. [Google Scholar]
  18. Yi, X.; Xu, H.; Zhang, H.; Tang, L.; Ma, J. Diff-retinex: Rethinking low-light image enhancement with a generative diffusion model. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 12302–12311. [Google Scholar]
  19. Wu, Y.; Wang, G.; Wang, Z.; Yang, Y.; Li, T.; Wang, P.; Li, C.; Shen, H.T. ReCo-Diff: Explore Retinex-Based Condition Strategy in Diffusion Model for Low-Light Image Enhancement. arXiv 2023, arXiv:2312.12826. [Google Scholar]
  20. Yin, Y.; Xu, D.; Tan, C.; Liu, P.; Zhao, Y.; Wei, Y. CLE Diffusion: Controllable Light Enhancement Diffusion Model. In Proceedings of the 31st ACM International Conference on Multimedia, Ottawa, ON, Canada, 29 October–3 November 2023; pp. 8145–8156. [Google Scholar]
  21. He, C.; Fang, C.; Zhang, Y.; Li, K.; Tang, L.; You, C.; Xiao, F.; Guo, Z.; Li, X. Reti-Diff: Illumination Degradation Image Restoration with Retinex-based Latent Diffusion Model. arXiv 2023, arXiv:2311.11638. [Google Scholar]
  22. Farid, H. Blind inverse gamma correction. IEEE Trans. Image Process. 2001, 10, 1428–1433. [Google Scholar] [CrossRef]
  23. Pizer, S.M. Contrast-limited adaptive histogram equalization: Speed and effectiveness stephen m. pizer, r. eugene johnston, james p. ericksen, bonnie c. yankaskas, keith e. muller medical image display research group. In Proceedings of the First Conference on Visualization in Biomedical Computing, Atlanta, GA, USA, 22–25 May 1990; Volume 337. [Google Scholar]
  24. Bennett, E.P.; McMillan, L. Video enhancement using per-pixel virtual exposures. In ACM SIGGRAPH 2005 Papers; ACM, Inc.: New York, NY, USA, 2005; pp. 845–852. [Google Scholar]
  25. Yuan, L.; Sun, J. Automatic exposure correction of consumer photographs. In Proceedings of the Computer Vision–ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Proceedings, Part IV 12. Springer: Berlin/Heidelberg, Germany, 2012; pp. 771–785. [Google Scholar]
  26. Rahman, S.; Rahman, M.M.; Abdullah-Al-Wadud, M.; Al-Quaderi, G.D.; Shoyaib, M. An adaptive gamma correction for image enhancement. EURASIP J. Image Video Process. 2016, 2016, 35. [Google Scholar] [CrossRef]
  27. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vision, Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  28. Jobson, D.J.; Rahman, Z.U.; Woodell, G.A. Properties and performance of a center/surround retinex. IEEE Trans. Image Process. 1997, 6, 451–462. [Google Scholar] [CrossRef]
  29. Fu, X.; Liao, Y.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A probabilistic method for image enhancement with simultaneous illumination and reflectance estimation. IEEE Trans. Image Process. 2015, 24, 4965–4977. [Google Scholar] [CrossRef]
  30. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2782–2790. [Google Scholar]
  31. Cai, B.; Xu, X.; Guo, K.; Jia, K.; Hu, B.; Tao, D. A joint intrinsic-extrinsic prior model for retinex. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4000–4009. [Google Scholar]
  32. Al Sobbahi, R.; Tekli, J. Low-light homomorphic filtering network for integrating image enhancement and classification. Signal Process. Image Commun. 2022, 100, 116527. [Google Scholar] [CrossRef]
  33. Chavarín, Á.; Cuevas, E.; Avalos, O.; Gálvez, J.; Perez, M. Contrast Enhancement in Images by Homomorphic Filtering and Cluster-Chaotic Optimization. IEEE Access 2023, 11, 73803–73822. [Google Scholar] [CrossRef]
  34. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  35. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  36. Shen, L.; Yue, Z.; Feng, F.; Chen, Q.; Liu, S.; Ma, J. Msr-net: Low-light image enhancement using deep convolutional network. arXiv 2017, arXiv:1711.02488. [Google Scholar]
  37. Li, C.; Guo, J.; Porikli, F.; Pang, Y. LightenNet: A convolutional neural network for weakly illuminated image enhancement. Pattern Recognit. Lett. 2018, 104, 15–22. [Google Scholar] [CrossRef]
  38. Zhang, Y.; Guo, X.; Ma, J.; Liu, W.; Zhang, J. Beyond brightening low-light images. Int. J. Comput. Vis. 2021, 129, 1013–1037. [Google Scholar] [CrossRef]
  39. Li, J.; Feng, X.; Hua, Z. Low-light image enhancement via progressive-recursive network. IEEE Trans. Circuits Syst. Video Technol. 2021, 31, 4227–4240. [Google Scholar] [CrossRef]
  40. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Uretinex-net: Retinex-based deep unfolding network for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5901–5910. [Google Scholar]
  41. Fu, H.; Zheng, W.; Meng, X.; Wang, X.; Wang, C.; Ma, H. You Do Not Need Additional Priors or Regularizers in Retinex-Based Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18125–18134. [Google Scholar]
  42. Fu, Z.; Yang, Y.; Tu, X.; Huang, Y.; Ding, X.; Ma, K.K. Learning a Simple Low-Light Image Enhancer From Paired Low-Light Instances. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22252–22261. [Google Scholar]
  43. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the 27th ACM international conference on multimedia, Nice, France, 21–25 October 2019; pp. 1632–1640. [Google Scholar]
  44. Xu, K.; Yang, X.; Yin, B.; Lau, R.W. Learning to restore low-light images via decomposition-and-enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2281–2290. [Google Scholar]
  45. Lim, S.; Kim, W. DSLR: Deep stacked Laplacian restorer for low-light image enhancement. IEEE Trans. Multimed. 2020, 23, 4272–4284. [Google Scholar] [CrossRef]
  46. Zheng, C.; Shi, D.; Shi, W. Adaptive unfolding total variation network for low-light image enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 4439–4448. [Google Scholar]
  47. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.P.; Kot, A. Low-light image enhancement with normalizing flow. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual Event, 22 February–1 March 2022; Volume 36, pp. 2604–2612. [Google Scholar]
  48. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for fast image restoration and enhancement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 1934–1948. [Google Scholar] [CrossRef] [PubMed]
  49. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. SNR-aware low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17714–17724. [Google Scholar]
  50. Xu, X.; Wang, R.; Lu, J. Low-Light Image Enhancement via Structure Modeling and Guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9893–9903. [Google Scholar]
  51. Jin, X.; Han, L.H.; Li, Z.; Guo, C.L.; Chai, Z.; Li, C. DNF: Decouple and Feedback Network for Seeing in the Dark. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 18135–18144. [Google Scholar]
  52. Yang, C.; Jin, M.; Jia, X.; Xu, Y.; Chen, Y. AdaInt: Learning adaptive intervals for 3D lookup tables on real-time image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 17522–17531. [Google Scholar]
  53. Wu, Y.; Pan, C.; Wang, G.; Yang, Y.; Wei, J.; Li, C.; Shen, H.T. Learning Semantic-Aware Knowledge Guidance for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1662–1671. [Google Scholar]
  54. Li, C.; Guo, C.L.; Zhou, M.; Liang, Z.; Zhou, S.; Feng, R.; Loy, C.C. Embedding fourier for ultra-high-definition low-light image enhancement. arXiv 2023, arXiv:2302.11831. [Google Scholar]
  55. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. From fidelity to perceptual quality: A semi-supervised approach for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3063–3072. [Google Scholar]
  56. Malik, S.; Soundararajan, R. Semi-Supervised Learning for Low-light Image Restoration through Quality Assisted Pseudo-Labeling. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 4105–4114. [Google Scholar]
  57. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
  58. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 1780–1789. [Google Scholar]
  59. Zheng, S.; Gupta, G. Semantic-guided zero-shot learning for low-light image/video enhancement. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 581–590. [Google Scholar]
  60. Xu, J.; Hou, Y.; Ren, D.; Liu, L.; Zhu, F.; Yu, M.; Wang, H.; Shao, L. Star: A structure and texture aware retinex model. IEEE Trans. Image Process. 2020, 29, 5022–5037. [Google Scholar] [CrossRef] [PubMed]
  61. Jobson, D.J.; Rahman, Z.u.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [PubMed]
  62. Rahman, Z.u.; Jobson, D.J.; Woodell, G.A. Multi-scale retinex for color image enhancement. In Proceedings of the 3rd IEEE International Conference on Image Processing, Lausanne, The Switzerland, 19 September 1996; IEEE: Piscataway, NJ, USA, 1996; Volume 3, pp. 1003–1006. [Google Scholar]
  63. Kong, X.Y.; Liu, L.; Qian, Y.S. Low-light image enhancement via poisson noise aware retinex model. IEEE Signal Process. Lett. 2021, 28, 1540–1544. [Google Scholar] [CrossRef]
  64. Dong, X.; Pang, Y.; Wen, J. Fast efficient algorithm for enhancement of low lighting video. In ACM SIGGRAPH 2010 Posters; ACM: New York, NY, USA, 2010; p. 1-1. [Google Scholar]
  65. Lee, H.; Sohn, K.; Min, D. Unsupervised low-light image enhancement using bright channel prior. IEEE Signal Process. Lett. 2020, 27, 251–255. [Google Scholar] [CrossRef]
  66. Oh, J.; Hong, M.C. Low-light image enhancement using hybrid deep-learning and mixed-norm loss functions. Sensors 2022, 22, 6904. [Google Scholar] [CrossRef] [PubMed]
  67. Ko, S.; Park, J.; Chae, B.; Cho, D. Learning lightweight low-light enhancement network using pseudo well-exposed images. IEEE Signal Process. Lett. 2021, 29, 289–293. [Google Scholar] [CrossRef]
  68. Shen, L.; Ma, Z.; Er, M.J.; Fan, Y.; Yin, Q. Blind adaptive structure-preserving imaging enhancement for low-light condition. IEEE Signal Process. Lett. 2022, 29, 917–921. [Google Scholar] [CrossRef]
  69. Gao, X.; Zhang, M.; Luo, J. Low-light image enhancement via retinex-style decomposition of denoised deep image prior. Sensors 2022, 22, 5593. [Google Scholar] [CrossRef] [PubMed]
  70. Anderson, M.; Motta, R.; Chandrasekar, S.; Stokes, M. Proposal for a standard default color space for the internet—srgb. In Proceedings of the Color and Imaging Conference, Scottsdale, AZ, USA, 19–22 November 1996; Society of Imaging Science and Technology: Springfield, VA, USA, 1996; Volume 4, pp. 238–245. [Google Scholar]
  71. Whang, J.; Delbracio, M.; Talebi, H.; Saharia, C.; Dimakis, A.G.; Milanfar, P. Deblurring via stochastic refinement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 16293–16303. [Google Scholar]
  72. Saharia, C.; Ho, J.; Chan, W.; Salimans, T.; Fleet, D.J.; Norouzi, M. Image super-resolution via iterative refinement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 4713–4726. [Google Scholar] [CrossRef] [PubMed]
  73. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  74. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  75. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; JMLR Workshop and Conference Proceedings, 2010. pp. 249–256. [Google Scholar]
  76. Paszke, A.; Gross, S.; Chintala, S.; Chanan, G.; Yang, E.; DeVito, Z.; Lin, Z.; Desmaison, A.; Antiga, L.; Lerer, A. Automatic Differentiation in PyTorch. In Proceedings of the NIPS 2017 Workshop on Autodiff, Long Beach, CA, USA, 9 December 2017. [Google Scholar]
  77. Song, J.; Meng, C.; Ermon, S. Denoising Diffusion Implicit Models. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  78. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef] [PubMed]
  79. Lee, C.; Lee, C.; Lee, Y.Y.; Kim, C.S. Power-constrained contrast enhancement for emissive displays based on histogram equalization. IEEE Trans. Image Process. 2011, 21, 80–93. [Google Scholar]
  80. Loh, Y.P.; Chan, C.S. Getting to know low-light images with the exclusively dark dataset. Comput. Vis. Image Underst. 2019, 178, 30–42. [Google Scholar] [CrossRef]
Figure 1. An overview of the framework. It contains four modules: image decomposition, brightness adjustment, reflectance restoration, and illumination restoration.
Figure 1. An overview of the framework. It contains four modules: image decomposition, brightness adjustment, reflectance restoration, and illumination restoration.
Sensors 24 02711 g001
Figure 2. Examples of low-light and normal-light images, along with their corresponding colormaps and graymaps—which can be viewed as reflectance maps and illumination maps—are shown. The example image displayed is from the test set of the LOL dataset.
Figure 2. Examples of low-light and normal-light images, along with their corresponding colormaps and graymaps—which can be viewed as reflectance maps and illumination maps—are shown. The example image displayed is from the test set of the LOL dataset.
Sensors 24 02711 g002
Figure 3. Visual comparison with other state-of-the-art methods on the LOL real-world dataset.
Figure 3. Visual comparison with other state-of-the-art methods on the LOL real-world dataset.
Sensors 24 02711 g003
Figure 4. Visual comparison with other state-of-the-art methods on the LOLv2-real real-world dataset.
Figure 4. Visual comparison with other state-of-the-art methods on the LOLv2-real real-world dataset.
Sensors 24 02711 g004
Figure 5. Visual comparison with other state-of-the-art methods on the LOLv2-syn synthetic dataset.
Figure 5. Visual comparison with other state-of-the-art methods on the LOLv2-syn synthetic dataset.
Sensors 24 02711 g005
Figure 6. Visual comparison with other state-of-the-art methods on the MEF dataset.
Figure 6. Visual comparison with other state-of-the-art methods on the MEF dataset.
Sensors 24 02711 g006
Figure 7. Visual comparison with other state-of-the-art methods on the VV dataset.
Figure 7. Visual comparison with other state-of-the-art methods on the VV dataset.
Sensors 24 02711 g007
Figure 8. Visual comparison with other state-of-the-art methods on the ExDark dataset.
Figure 8. Visual comparison with other state-of-the-art methods on the ExDark dataset.
Sensors 24 02711 g008
Figure 9. Comparison of enhancement results from different image decomposition strategies.
Figure 9. Comparison of enhancement results from different image decomposition strategies.
Sensors 24 02711 g009
Figure 10. Comparison of enhancement results when removing prior brightness information.
Figure 10. Comparison of enhancement results when removing prior brightness information.
Sensors 24 02711 g010
Table 1. Quantitative comparison with other state-of-the-art methods on the LOL dataset.
Table 1. Quantitative comparison with other state-of-the-art methods on the LOL dataset.
Model TypeMethodPSNR↑SSIM↑LPIPS↓
CNNRetinexnet [35]16.770.5390.474
Kind++ [38]21.800.8760.158
TransformerRetinexformer [11]25.150.8970.131
Normalizing Flowwang2022low [47]25.190.9300.110
Diffusion ModelCLE [20]25.510.8880.164
DiffLL [15]26.320.8980.118
Ours27.470.9290.098
Table 2. Quantitative comparison with other state-of-the-art methods on the LOLv2-real dataset.
Table 2. Quantitative comparison with other state-of-the-art methods on the LOLv2-real dataset.
Model TypeMethodPSNR↑SSIM↑LPIPS↓
CNNRetinexnet [35]15.960.6740.390
Kind++ [38]17.660.7830.217
TransformerRetinexformer [11]22.790.8660.171
Normalizing Flowwang2022low [47]25.420.8920.157
Diffusion ModelCLE [20]20.720.8060.232
DiffLL [15]28.880.8960.100
Ours29.160.9140.119
Table 3. Quantitative comparison with other state-of-the-art methods on the LOLv2-syn dataset.
Table 3. Quantitative comparison with other state-of-the-art methods on the LOLv2-syn dataset.
Model TypeMethodPSNR↑SSIM↑LPIPS↓
CNNRetinexnet [35]19.390.8330.252
Kind++ [38]17.480.8130.232
TransformerRetinexformer [11]25.670.9520.059
Normalizing Flowwang2022low [47]26.060.9570.047
Diffusion ModelCLE [20]28.170.9410.078
DiffLL [15]22.460.8880.159
Ours29.900.9630.046
Table 4. Ablation study on the decomposition strategy.
Table 4. Ablation study on the decomposition strategy.
Decomposition StrategyPSNR↑SSIM↑LPIPS↓
Classical Decomposition24.580.8500.171
Learning-Based Decomposition23.690.9120.095
Ours27.470.9290.098
Table 5. Ablation study on brightness guidance.
Table 5. Ablation study on brightness guidance.
PriorPSNR↑SSIM↑LPIPS↓
w/o Brightness Guidance18.080.8580.098
Ours27.470.9290.098
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, F.; Liu, X.; Gao, C.; Sang, N. Color and Luminance Separated Enhancement for Low-Light Images with Brightness Guidance. Sensors 2024, 24, 2711. https://doi.org/10.3390/s24092711

AMA Style

Zhang F, Liu X, Gao C, Sang N. Color and Luminance Separated Enhancement for Low-Light Images with Brightness Guidance. Sensors. 2024; 24(9):2711. https://doi.org/10.3390/s24092711

Chicago/Turabian Style

Zhang, Feng, Xinran Liu, Changxin Gao, and Nong Sang. 2024. "Color and Luminance Separated Enhancement for Low-Light Images with Brightness Guidance" Sensors 24, no. 9: 2711. https://doi.org/10.3390/s24092711

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop