Next Article in Journal
Closed-Loop Control of Melt Pool Temperature during Laser Metal Deposition
Previous Article in Journal
The Effectiveness of eHMI Displays on Pedestrian–Autonomous Vehicle Interaction in Mixed-Traffic Environments
Previous Article in Special Issue
RCEAU-Net: Cascade Multi-Scale Convolution and Attention-Mechanism-Based Network for Laser Beam Target Image Segmentation with Complex Background in Coal Mine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fast, Zero-Reference Low-Light Image Enhancement with Camera Response Model

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(15), 5019; https://doi.org/10.3390/s24155019 (registering DOI)
Submission received: 5 July 2024 / Revised: 26 July 2024 / Accepted: 1 August 2024 / Published: 2 August 2024

Abstract

:
Low-light images are prevalent in intelligent monitoring and many other applications, with low brightness hindering further processing. Although low-light image enhancement can reduce the influence of such problems, current methods often involve a complex network structure or many iterations, which are not conducive to their efficiency. This paper proposes a Zero-Reference Camera Response Network using a camera response model to achieve efficient enhancement for arbitrary low-light images. A double-layer parameter-generating network with a streamlined structure is established to extract the exposure ratio K from the radiation map, which is obtained by inverting the input through a camera response function. Then, K is used as the parameter of a brightness transformation function for one transformation on the low-light image to realize enhancement. In addition, a contrast-preserving brightness loss and an edge-preserving smoothness loss are designed without the requirement for references from the dataset. Both can further retain some key information in the inputs to improve precision. The enhancement is simplified and can reach more than twice the speed of similar methods. Extensive experiments on several LLIE datasets and the DARK FACE face detection dataset fully demonstrate our method’s advantages, both subjectively and objectively.

1. Introduction

Due to insufficient light intensity or performance limitations of imaging equipment, low-light images inevitably exist in many machine vision applications, such as intelligent monitoring, automatic driving and mine monitoring. Especially in the field of space-based remote sensing, a large object distance makes it difficult for the sensor to receive enough energy to obtain images with high brightness. Although extending the exposure time can play a role in improvement, carriers such as satellites generally need to maintain a moving state, and it is difficult to meet the needs of both aspects at the same time. Low-light image enhancement (LLIE) helps alleviate the requirement of the movement mechanism to obtain high-quality image information. In addition to low brightness, low-light images often have problems such as poor contrast and strong noise, which hinder subsequent analysis and processing, so it is of great significance to study the enhancement of low-light images. So far, many methods have been proposed to realize LLIE. They can be broadly divided into two categories: traditional methods and deep learning-based methods.
Traditional low-light image enhancement methods are mainly based on histogram equalization [1,2,3], Retinex theory [4,5,6], wavelet transform [7], defogging or dehazing models [8,9,10] or multi-graph fusion [11]. The first two methods have been studied the most. Although the histogram equalization method has a simple principle and fast calculation, the enhancement effect is poor. Methods based on Retinex theory usually divide a low-light image into a reflection component and an illumination component based on some prior knowledge, among which the estimated reflection component is the enhancement result. However, the ideal assumption of the reflection component as the enhancement result is not always true, which may lead to degraded enhancement, with effects such as the loss of detail or color distortion. In addition, it is very difficult to find effective prior knowledge, and inaccurate prior knowledge may lead to artifacts or color deviations in the enhancement results. In addition to these mainstream research directions, LECARM [12] thoroughly studied an LLIE architecture based on the response characteristics of the camera. It analyzed the camera response model parameters and estimated the exposure ratio of each pixel using illumination estimation technology to realize the enhancement of low-light images. But, the exposure ratio estimation method used by LECARM is LIME [13] based on Retinex theory. Various parameters need to be designed manually, which is difficult to control.
Since KG Lore et al. proposed LLNet [14] in 2017, low-light image enhancement methods based on deep learning have attracted the attention of more and more researchers. Compared with traditional methods, deep learning-based solutions often have better accuracy. Such methods generally use a large set of labeled pairwise data to drive a pre-defined network to extract image features and realize enhancement. Some dehazing methods can also be used to improve brightness. SID-Net [15] integrates adversarial learning and contrastive learning into a unified net to efficiently enhance pictures. More can be found in [16]. However, in the field of LLIE, it is difficult to determine a fixed standard for the target image, and it is extremely difficult to obtain ideal paired data. Moreover, preset target images will drive the model to learn more unique features of the dataset and reduce the generalization ability. In this respect, zero-reference methods in the form of unsupervised learning can enhance low-light input images without requiring references from the dataset, which reduces the dependence on the dataset and is beneficial in expanding the adaptation scenarios of the model. Based on the idea of generation and confrontation, a zero-reference attention-guided U-Net network was designed in EnlightenGAN [17]. It uses a global discriminator, a local discriminator and self-feature preserving loss to restrain the training process. The ZeroDCE [18] proposed by Guo et al. uses a variety of images taken at different exposures to train a lightweight depth curve estimation network. A series of elaborately designed loss functions are used to realize training without references.
Although each method can achieve a certain degree of enhancement, there are still some problems. EnlightenGAN establishes the model based on a complete U-Net, which is relatively complex and is not conducive to real-time applications. The fixed-exposure loss function used by ZeroDCE ignores the brightness distribution in the original images, which will limit the contrast of the results. And, its enhancement process requires many iterations, which is contrary to high-speed processing. The subsequently proposed ZeroDCE++ [19] improves the speed by compressing the size of input images and reducing the number of convolution operations. But, it is easy to lose important information when the scale of targets of interest is small. Most other similar methods also use multiple iterations [18,19,20,21] or complicated model structures [17,22] to achieve enhancement, limiting the processing speed of the algorithm.
In order to solve the above problems, this paper proposes the Zero-Reference Camera Response Network (ZRCRN). Inspired by the CRM, the ZRCRN realizes an efficient framework. Without multiple iterations, it just uses a simple network structure to achieve high-quality enhancement, which effectively improves the processing speed. Two reference-free loss functions are also designed to restrict the training process to improve the generalization ability of the model.
The main contributions of this research can be summarized as follows:
  • A fast LLIE method is proposed, namely, the ZRCRN, which establishes a double-layer parameter-generating network to automatically extract the exposure ratio of a camera response model to realize enhancement. The process is simplified, and the speed can reach more than twice that of similar methods. In addition, the ZRCRN can still obtain an accuracy comparable to SOTA methods.
  • A contrast-preserving brightness loss is proposed to retain the brightness distribution in the original input and enhance the contrast of the final output. It converts the input image to a single-channel brightness diagram to avoid serious color deviations caused by subsequent operations and linearly stretches the brightness diagram to obtain the expected target. It can effectively improve the brightness and contrast without requiring references from the dataset, which can improve the generalization ability of the model.
  • An edge-preserving smoothness loss is proposed to remove noise and enhance details. The variation trend in the input image pixel values is selectively promoted or suppressed to reach the desired goal. While maintaining the advantage of zero references, it can also drive the model to achieve the effect of sharpening and noise reduction, further refining its performance.
The rest of this paper is organized as follows. The related works on low-light image enhancement are reviewed in Section 2. In Section 3, we introduce the details of our method. Numerous experiments on standard datasets have been implemented to evaluate the enhancement performance and are presented in Section 4. The discussion and conclusion are finally given in the last two sections.

2. Related Works

2.1. Camera Response Model

Images are mainly obtained by digital cameras, which collect the radiation energy from the real scene through the lens and photosensitive sensors, convert the optical signals into electrical signals, and then obtain the final images through a series of digital image processing methods and store them in the memory. The whole process, from the incident radiation at the lens to the pixel values of the final images, contains many linear or nonlinear transformations, which can be represented by a camera response function (CRF). It is convenient for analysis to condense the whole process into one function.
When the exposure setting of the camera changes, the radiation energy incident to the sensor will change accordingly, and these two changes should be linearly correlated. Meanwhile, the pixel values of the image output by the camera will also change, and this change can be expressed by the brightness transformation function (BTF). Denoting the CRF by f ( · ) and BTF by g ( · ) , then, as shown in Figure 1, these two functions have the following relationships:
P = f E ,    E 1 = k E 0 ,      P 1 = g P 0 , k ,
in which E represents the incident radiation matrix, E 0 and E 1 are two of its index values under two different exposure conditions, P 0 and P 1 are the corresponding image pixel value matrices, and k is the exposure ratio.
The CRF and BTF describe the characteristics of the imaging process of a camera. Together, they form a camera response model (CRM) [12]. A relationship also exists between the CRF and BTF:
g f E , k = f k E .
This equation, also known as a comparative equation, can be used to convert between the CRF and BTF [12].
The CRF can be obtained by experimentally measuring radiation intensity and image pixel values. In practice, it is generally assumed that the CRF and BTF are the same at each spatial position. DoRF [23] is a database that stores the response curves of 201 real cameras. Although different cameras should have different camera response models in theory, if the camera response model is regarded as a medium for transforming the incident radiation matrix E 0 to the perfect image P 1 , and E 0 obviously has nothing to do with the camera and the goal of P 1 is to match human perception, then considering that most human perception standards are similar, the CRM should also have a general solution that can meet most requirements. The study of Yurui Ren et al. [12] proved that, for most cameras, using the average curve of the DoRF database instead of its own three-channel camera response curve for LLIE will produce only limited distortions. They also summarized a general and fast CRM, whose CRF and BTF are as shown in the following equation:
f ( E ) = ( 1 + a ) E b E b + a , g ( P , k ) = k b P ( 1 + a ) ( k b 1 ) P + 1 + a , a = 0.6 , b = 0.9 ,
where E is the pixel value of the incident radiation metric, and P is the pixel value of the low-light image.
The CRF in this model has the form of a sigmoid function and was proposed by Gabriel Eilertsen et al. [24]. The ZRCRN establishes a network based on this model to realize LLIE.

2.2. Zero-Reference LLIE Loss Function

A loss function in the field of image enhancement often uses the mean absolute error or mean square error (MSE) between the enhancement result and the target image. The former is not derivable at 0. The latter is always expressed as Equation (4), with n denoting the count of the samples, y i denoting the true value of the i-th sample and y i ^ denoting its predicted value.
M S E = 1 n i = 1 n ( y i ^ y i ) 2 .
Although susceptible to abnormal data, the MSE is globally derivable and is more widely applied. When no target image is provided, there are two main schemes for the establishment of the loss function: building the target image based on the original image or setting the ideal target image directly. Although the final form of the loss function is always still the mean error, the target image is no longer limited to RGB pixel values but is also characterized by a brightness map [18,19,20,21,22,25,26] (usually with a fixed value of 0.6 for improving brightness), a hue diagram [25] (H channel in the HSV space to maintain color), the difference between any two color channels [18,19,21,26,27,28] (usually with a fixed value of 0 for holding color), the difference between neighborhood brightness levels [18,19,20,21,22,25,26,27,28] (usually with a fixed value of 0 or extracted from the original image, often used for noise suppression or maintaining the brightness distribution trend) and so on. The enhancement results used to calculate the error will also be replaced by the corresponding features. In addition, there are some loss functions designed according to proprietary model structures, such as generative adversarial loss [17,29], semantic loss [21], atmospheric light loss [20], the consistency loss of neighborhood parameter values [18,19], etc. In practice, the sum of multiple loss functions is often used to achieve the purpose of brightness enhancement, contrast enhancement and noise suppression at the same time.
When a combination of multiple loss functions is used, the weight of each component should be carefully set. Otherwise, one kind of loss function can easily dominate, and other loss functions are almost ineffective. At present, most weights are determined by manual testing. There are also some semi-automatic configuration methods. For example, A. Kendall et al. [30] set the weights by analyzing the homoscedastic uncertainty of each task, while O. Sener et al. [31] solved multi-task learning as a multi-objective optimization problem and proposed a multiple-gradient descent algorithm.
The main goal of LLIE is to improve the brightness and contrast of the image and suppress noise. The commonly used brightness loss function is meant to constrain the enhancement result to a target brightness map with a fixed value at each position. Although it can improve brightness, the brightness distribution may be abnormal. And, this loss is not conducive to contrast improvement. In addition, the loss function is less able to suppress noise. The main objective of increasing the consistency of neighborhood brightness differences between the result and the original image is to maintain the brightness distribution trend. If the brightness difference tends to zero, it will lead to global smoothing, the easy loss of details and lower contrast. This paper designs two loss functions that extract key information from the original image to establish the reference target for the performance improvement of the model.

3. Methods

The framework of the ZRCRN is shown in Figure 2. Based on the CRM, the original low-light input image P 0 is first inverted to the radiation map E 0 according to the sigmoid CRF to reduce complexity. A streamlined parameter-generating network (PGN) is then established to extract the pixel-wise exposure ratio K from the radiation map. Finally, the corresponding BTF calls K to perform one transformation on P 0 to achieve LLIE. The CRM, including the sigmoid CRF, is shown in Equation (3). The designs of the parameter-generating network and the loss functions are detailed below.

3.1. Parameter-Generating Network

The image quality in low-light scenarios can be improved by adjusting the camera exposure parameters. Extending the exposure time or increasing the aperture size can increase both the brightness and contrast of the final image. At the same time, according to the CRM, the adjustment of exposure parameters can be represented by the parameter k. As long as an appropriate k can be obtained, LLIE can be realized by the CRF or BTF.
In LLIE tasks, k represents the exposure ratio between the low-light image and the normal-light image. The ideal normal-light image should have a fixed average brightness, and the camera needs to be configured to different exposures for different scenes to obtain this target value. In practice, the scene and exposure of the low-light image are fixed, so the exposure ratio should be a function of the low-light input image. In addition, considering that low-light images often have the problem of uneven illumination, which also needs to be tackled in the enhancement, the optimal exposure ratio should vary at different coordinates. Accordingly, the exposure ratio should be a matrix matching P 0 pixel by pixel, and the objective function φ should meet the following relationship:
K = φ P 0 ,
in which K is the exposure ratio matrix. Although K can be obtained from the features extracted from P 0 directly, a deeper and more complex model may be needed to achieve a good fit since P 0 is obtained by the camera through a CRF transformation, and the CRF is a collection of many transformations. In addition, the main function of K is to raise the environmental radiation to the level of the normal illumination scene, which is independent of the complex CRF transformation inside the camera. Therefore, the exposure ratio parameter K required by the enhancement can be obtained by the following process: first, inversely transform P 0 by the CRF to obtain the radiation map in the low-light scene ( E 0 ), and then extract K from E 0 . This process can be expressed as the following equation:
K = ψ E 0 = ψ f 1 P 0 .
At this time, the required objective function ( ψ ) has weaker nonlinearity, which is conducive to achieving a better fit using a simpler model and improving the processing speed.
Convolution requires very few parameters to achieve the extraction of important information in the matrix. It has the characteristics of translation invariance and locality and is widely used in the field of image processing. The PGN is built using m convolution layers and m 1 activation layers, with each convolution layer followed by one activation layer, except for the last one. The kernel size of the convolution layer is 3 × 3 . The height and width of the output image remain unchanged by padding. The activation layer uses the LeakyReLU function with a negative slope of 0.01 as the activation function. The first convolution layer has 3 channels. The last has 1 channel. All output channels of the other convolution layers are n. The overall parameter-generating network structure is shown in Figure 2. A larger m or n can improve the representation ability of the model. Their best values were determined to be 2 and 2 through experiments described in the next section (Figure 3).

3.2. Loss Function

The loss function plays an important role in learning-based methods. It controls the direction of the training process. With it, a model can acquire specific abilities. A loss function should be derivable to implement backpropagation. The MSE is a commonly used loss function. It measures the distance between two instances and is globally derivable. But, it needs a reference, which is not always available in a dataset. Considering that some traditional methods have unique advantages in enhancing specific properties of an image, we propose two loss functions to generate references according to the low-light inputs.
Aiming at the three main problems of low brightness, low contrast and strong noise, two loss functions are designed to constrain the training process: contrast-preserving brightness loss and edge-preserving smoothness loss. Both are essentially MSE loss functions. Some attributes of the original image are separately extracted to construct the target image with restricted enhancement. Then, the corresponding properties are calculated from the output of the network. The deviation between these two elements is minimized so as to achieve the enhancement effect.

3.2.1. Contrast-Preserving Brightness Loss

The original low-light image ( P 0 ) has three RGB channels and low pixel values in each channel. The brightness map is extracted from P 0 , and linear stretching is performed to obtain the target brightness map. The specific process is as follows:
  • In order to reduce the impact on color information, the RGB channel is fused into a single channel to obtain the brightness map. The conversion formula uses the following, which is more suitable for human perception:
    Φ ( x i ) = 0.299 · r i + 0.587 · g i + 0.114 · b i ,
    where x i represents the value of the i-th pixel in the image. It is a vector consisting of a red channel value ( r i ), a green channel value ( g i ) and a blue channel value ( b i ).
  • After applying Φ ( · ) to each pixel of P 0 , the brightness map ( E brightness ) is dark. Then, it is linearly stretched to enhance the brightness and contrast. The following equation is used to extend the pixel values of E brightness linearly to the a b interval to obtain the expected brightness map ( E exptar ):
    E exptar = a + b a m a x ( E brightness ) m i n ( E brightness ) ( E brightness a ) .
    where m a x ( · ) and m i n ( · ) represent the maximum and minimum values of all elements in the argument matrix. Considering the purpose of increasing the brightness, b can be directly set as the theoretical maximum value of 1. Since the brightness distribution of the original image is concentrated close to the minimum value, the value of a should be more carefully determined. The larger a is, the smaller the width of the brightness distribution interval and the lower the contrast. If a is too small, the visual effect of the target image will obviously be darker, and the corresponding model will not effectively enhance the brightness of the original image. The best value of a was determined to be 0.2 through experiments described in the next section.
Once E exptar is obtained, the contrast-preserving brightness loss can be calculated by the following equation:
L e x p = 1 N i = 1 N ( Φ ( x i ) y i ) 2 , x i P 1 , y i E exptar ,
where N represents the total count of pixels in the image. x i denotes the i-th element (including three RGB channels) of the enhancement result ( P 1 ). y i is the i-th element of E exptar . The brightness distribution of a low-light image is generally concentrated in a small range close to zero. Linearly stretching it to a 1 can improve the brightness and contrast. Compared with the commonly used fixed target brightness of 0.6 [18,19,21,22], the contrast-preserving brightness loss is more conducive to retaining the spatial distribution information of the brightness in the original image and improving the contrast of the enhancement results.

3.2.2. Edge-Preserving Smoothness Loss

In addition to the brightness and contrast, noise is also a major problem in low-light images. Suppressing the variation trend in the input image pixel values can remove noise. The variation trends in the image pixel values are always represented as a gradient diagram. Noise appears as a large number of non-zero isolated points in the gradient diagram. To realize proper noise reduction, the gradient map and edge map are generated based on P 0 , the values in the non-edge region of the gradient map are set to zero and the gradient in the edge region is properly amplified to obtain the target gradient map. Specifically, it includes the following three steps:
  • Edge Detection. The main basis for generating the edge diagram is the Canny algorithm [32]. Firstly, horizontal and vertical Gaussian filters are applied to P 0 to suppress the influence of noise on edge detection. The size and standard deviation of the filter inherit the default parameters of 5 and 1. Then, the gradient information is extracted. The horizontal and vertical Sobel operators of size 3 × 3 are used to calculate the gradient diagram in two directions of each of the three RGB channels from the filtered result. And, the amplitude of the total gradient at each position is obtained by summing the gradient amplitudes in each channel. Next, the double-threshold algorithm is used to find the positions of the strong and weak edges. Finally, the edge diagram ( E edge ) is obtained by judging the connectivity and removing the isolated weak edge. The edge refinement operation with non-maximum suppression from the original Canny algorithm is skipped. So, there is no need to calculate the direction of the total gradient. The main purpose is to keep the gradient edge with a larger width in the original image and make the enhancement result more natural. The two thresholds α and β of the double-threshold algorithm have a great influence on the edge detection results. Their best values, as determined by the experiments described in the next section, are 0.707 and 1.414, respectively.
  • Gradient Extraction. The horizontal and vertical Sobel operators are applied directly to P 0 to obtain a total of six gradient maps in three channels and two directions. Then, the total gradient amplitude (the sum of the gradient amplitudes in all channels) is calculated as the pixel value in the final gradient map ( E grad ). The entire gradient extraction process can be expressed as the following equation:
    E grad = Ψ ( P 0 ) .
  • Selectively Scaling. The edge diagram ( E edge ) is used as a mask to multiply the gradient map ( E grad ) element by element, which can filter out the non-edge part of E grad . Then, the result is scaled by an amplification factor ( γ ) to obtain the expected target image ( E smoothtar ):
    E smoothtar = γ E grad E edge .
    where ⨀ indicates the multiplication of two metrics pixel by pixel. The gradient at the edge of the low-light image will be enhanced correspondingly after the brightness is enhanced. So, the target gradient map should also amplify this part. The best value of γ was determined to be 2 by the experiments described in the next section.
After obtaining E smoothtar , denoting the total count of elements in the image by N, the edge-preserving smoothness loss can be calculated by the following formula:
L s m o o t h = 1 N i = 1 N ( x i y i ) 2 , x i Ψ ( P 1 ) , y i E smoothtar .
Although each step in the calculation of the loss function is derivable, the total gradient amplitude is solved by the root operation, and the problem of dividing by 0 can easily occur. In practice, an additional minimum of 1 × 10−8 is added to the base number to ensure that the training process is continuous.
The gradient map of the low-light image often contains a large number of non-zero isolated points. Making them close to zero can achieve smoothness. Enlarging the gradients at the edge positions is beneficial in retaining the details of the original image. To the best of our knowledge, in zero-reference LLIE, the loss function is currently less able to suppress noise. Although some work tried to constrain the brightness difference in the neighborhood, the main objective of increasing the consistency of neighborhood brightness differences between the result and the original image [18,19] is to maintain the brightness distribution trend. If the brightness difference tends to zero [21,27,28], it will lead to global smoothing, the easy loss of details and lower contrast. The edge-preserving smoothness loss attempts to separate the edge region specifically, enhancing the variation in this area and suppressing others to build references, which can further improve precision.

3.2.3. Total Loss Function

Finally, according to the multi-loss combination strategy proposed by A. Kendall et al. [30], two learnable parameters are added to combine the two loss functions together to obtain the total loss function:
L t o t a l = e p L e x p + p + e q L s m o o t h + q .
where p and q are two learnable parameters with an initial value of zero. To a certain extent, they can release us from the adjustment of weights and improve precision.
In practice, the expected target images required for each loss can be calculated according to the images in the training dataset before formal training and directly loaded during training to enhance efficiency.

4. Experiments and Results

4.1. Implementation Details

The contrast-preserving brightness loss and the edge-preserving smoothness loss leave the training of the ZRCRN without the requirement for paired data from datasets. To address the common problem of non-uniform illumination in LLIE tasks, multi-exposure image sequences from the SICE dataset [33] were selected for training, the same as ZERO_DCE [18], to give the model the ability to process input images with diverse exposure states.
The model was implemented through PyTorch and trained on a single GPU with the batch size set to 10. At initialization, the weights of all convolution layers were sampled from a Gaussian distribution with a mean of 0 and a standard deviation of 0.02, with the bias set to 0. The initial values of the two learnable parameters α and β in the loss function were also set to 0. The ADAM optimizer [34] was used to update the parameters. In addition to the learning rate, the value of each parameter was left as the default. The learning rate was configured according to the warm start + cosine annealing strategy [35]. The basic learning rate was 0.05. The learning rate increased linearly from 1 % of the basic learning rate to 100 % in 100 batches. The maximum learning rate of cosine annealing was also the basic value. The period was kept at 200, and the minimum learning rate was set to one-tenth of the basic learning rate. The learning rate was updated at every batch.
The training loss and the validation SSIM during our training process are presented in Figure 4. After the initial rapid improvement and the following slow increase, the validation SSIM tended to converge to a value near 0.7. Although the cosine annealing strategy led to some fluctuations in training loss, it converged to a value near 1.0.

4.2. Ablation Study

To analyze the influence of key parameters and verify the role of inverse transformation and loss functions, five sets of ablation experiments were designed for comparison. The test set used for the quantitative comparison in each experiment was the test part of the real-captured set of LOL_v2 [36]. The SSIM [37], the PSNR and the average single-image processing time (denoted as TIME, in ms, calculated by dividing the processing time of all images in the dataset by the count of images contained in the dataset) were selected for performance analysis. In all tables, the optimal results in each column are shown in bold, the second-best values are underlined and the third-best results are in italics.

4.2.1. Influence of Key Parameters

Number of Layers and Channels in PGN. Using deeper and wider convolution structures gives the model stronger representation but leads to more computation, less efficiency and more difficulties in finding model parameters that meet the expected goals. m = 1 , 2 , 3 , 4 and n = 1, 2, 4 were used to generate binary groups for the quantitative evaluation. The results are shown in Table 1. It can be seen that the scheme with m = 1, which obtains single-channel enhancement parameters directly from the three-channel input image, only achieved an SSIM of 0.44. It is too simple a structure to extract enough features and cannot obtain sufficient accuracy. The combination of m = n = 2 achieved the highest SSIM in the test set and could process extremely fast. Structures with m > 2 and n > 2 failed to achieve higher accuracy with the current learning strategy. Although setting the additional convolution kernels to specific values should yield the same output result as the combination of m = n = 2 , a deeper and wider model structure has a larger parameter space and is more likely to fall into other local minimum points during the optimization, which leads to sub-optimal results. They also slowed down the enhancement. Considering accuracy and efficiency, a network structure with m = n = 2 was chosen.
Minimum of Linear Stretching Target Interval in Contrast-Preserving Brightness Loss. The brightness and contrast of low-light images can be improved by linear stretching. The brightness distribution of low-light images is concentrated near zero, so the selection of the lower limit of the target interval has a great influence on the enhancement effect. A larger value will brighten the image remarkably. Values of a = 0, 0.1, 0.2, 0.3, 0.4 and 0.5, combined b with a value of 1, were taken to obtain the linear stretching target interval [ a , 1 ] for comparison. The results are shown in Table 2. When a was too large (a = 0.5) or too small (a = 0.0), the SSIM between the output of the network and the target image was degraded significantly. Although the SSIMs of the other four sets of parameters were not much different, the PSNR achieved the highest value at a = 0.2. We think that with a = 0.2, a good reference can be built for most low-light images, and a = 0.2 is chosen as the minimum of the linear stretching target interval in contrast-preserving brightness loss.
In addition, we calculated the average SSIM between all generated references under five settings and the brightness of the normal-light images in our test set. The results are shown in Table 2 under the heading “SSIM-EXPTAR2BONL”. They show a similar trend to other SSIMs, which means that our model robustly learned the ability to brighten an image, and the better the generated reference is, the more accurate the model can be. Considering that a learning process still exists between these two processes, we identified some candidate parameters based on a rule of thumb to build the references and drive the model to learn. Then, the best result was found, and its parameter setting was used as the final choice.
Double Thresholds and Amplification Factor in Edge-preserving Smoothness Loss. The double thresholds ( α and β ) in the edge-preserving smoothness loss directly determine the final position of the detected edge. It is easy to lose details when the thresholds are too large, and a large amount of noise will be retained when they are too small. The amplification factor ( γ ) changes the amplitude of the gradient at the edge, and an inappropriate release may trigger additional color variations. Considering that a great change in any color channel can indicate the presence of an edge, alternative parameters were set based on the theoretical maximum of the single-channel gradient amplitude. The larger ( β ) of the double thresholds was taken as 0.304, 0.707, 1.414 and 2.828, which were halved to obtain the smaller threshold ( α ), combined with γ = 1, 2, 3 for comparison. The results are shown in Table 3. As shown in the table, both metrics at γ = 1 were poor. Although the SSIM was similar, the PSNR was better when γ = 2 compared to γ = 3 with the same β . And, the PSNR achieved the highest value at β = 1.414 . Thus, α = 0.707 , β = 1.414 and γ = 2 were chosen as the double thresholds and the amplification factor in the edge-preserving smoothness loss. These values may build good references for most low-light images.

4.2.2. Effect of Inverse Transformation

The main purpose of the inverse transformation is to reduce the nonlinearity of the subsequent parameter extraction process and to facilitate the use of a simple model to obtain excellent performance. Under the same PGN structure, the performance of the two schemes in extracting directly from the original low-light image (namely, “without inverse”) and from the inverse transformation result (namely, “with inverse”) is shown in Table 4. It can be seen that the inverse transformation led to improvements of 0.04 in the SSIM and 1.00 in the PSNR. The scheme of first inverting and then extracting obtained better performance. This shows that the inverse transformation is conducive to the simplification of the PGN.

4.2.3. Effect of Loss Functions

Contrast-preserving brightness loss and edge-preserving smoothness loss realize the extraction and retention of the original brightness information and edge information separately. A comparison was carried out by using the fixed-target-exposure loss (approximately equivalent to the contrast-preserving brightness loss with each element of E exptar as 0.6) and discarding the corresponding loss function. The results are shown in Figure 5 and Table 5.
As can be seen in Figure 5, the scheme of fixed-target-exposure loss plus edge-preserving smoothness loss showed a problem of abnormal color distribution. The numbers on the digital clock appeared black, contrary to the expectation. And, the specific numbers were hardly visible before amplification. Discarding the brightness loss directly led to the whole image being black. It only achieved a weak background brightness enhancement, and the original clock part with high brightness was suppressed. As a result of discarding the smoothness loss, the wall tiles under the digital clock appeared slightly more noisy. The edges of the digital parts were also more vague and had larger halos. The quantitative comparison in Table 5 also shows that the model output under our loss is closest to the expected target.

4.3. Experimental Configuration

Experiments were conducted on commonly used LLIE public datasets to compare the performance of the ZRCRN with several other SOTA methods: two traditional methods (LECARM [12], SDD [38]), two supervised learning methods (StableLLVE [39], URetinex_Net [40]) and three unsupervised learning methods (SCI [41], EnlightenGAN [17], Zero_DCE [18]). Each was implemented using the public code. The traditional methods ran on a CPU (Intel (R) Core (TM) i7-10870H). The learning-based methods, including our ZRCRN, loaded the specified checkpoints and ran on a GPU (NVIDIA RTX A6000). SCI had three available checkpoints, and the “difficult” version was selected here.
The experimental datasets include two full-reference (FR) datasets (LOL [42], LOL_v2 [36] (test part of real_captured)) and three no-reference (NR) datasets (DICM [43], LIME [13], MEF [44]). The performance of the nine algorithms was compared qualitatively and quantitatively. Then, the adaptability of these methods to advanced image tasks was evaluated by a night face detection experiment on the Dark Face [45] dataset.

4.4. Benchmark Evaluations

4.4.1. Qualitative Comparison

The qualitative comparisons of each method on each dataset are shown in Figure 6 and Figure 7, where the first row indicates the method to which each column belongs, and the first column indicates the dataset to which each row belongs. The parts in the red boxes are enlarged to facilitate comparison. It can be seen that each method could achieve a certain degree of enhancement. But, StableLLVE and URetinex_Net, using supervised learning strategies, showed significantly abnormal colors on the DICM and MEF datasets. Their application scenarios are severely limited. The enhancement results of the two traditional methods (LECARM and SDD) and the unsupervised method SCI on each dataset were somewhat dark. They could not achieve sufficient brightness improvement. Our ZRCRN resulted in good enhancement on all datasets and has the ability to suppress noise, which is most evident in the comparison on the LIME dataset. In the image of the LIME dataset enhanced by the ZRCRN, the wall on the right side of the bottle is very smooth, while the other three unsupervised methods retained relatively severe noise.

4.4.2. Quantitative Comparison

The quality of the images enhanced by each method was evaluated using two metrics on each type of dataset separately. The SSIM [37] and PSNR were used for the FR datasets. BRISQUE [46] and NIQE [47] were used for the NR datasets. The comparisons are shown in Table 6 and Table 7. Two supervised learning methods achieved the top two mean SSIMs on LOL and LOL_v2, but their metrics on the three NR datasets ranked almost in the last two, which shows great application limitations. Four learning methods that do not require references from the datasets achieved the optimal four mean NIQE values on the NR datasets and also achieved excellent SSIMs and PSNRs on the FR datasets, which reflects their splendid generalization ability. Our ZRCRN obtained the best mean BRISQUE, and the other three metrics also ranked second among the four unsupervised methods.
In addition, the efficiency of each method was evaluated using FLOPs, the count of model parameters (#Params) and the average single-image processing time (denoted as TIME, in ms, calculated by dividing the processing time of all images in the test part of real_captured in LOL_v2 by the count of images contained in the same dataset). The results are shown in Table 8 (FLOPs and #Params are only calculated for learning-based methods). All of the metrics of our ZRCRN ranked first. The speed is more than twice that of Zero_DCE and EnlightenGAN with similar quality. Although the speed is not significantly improved compared to SCI, the accuracy is effectively improved by 0.07 for the mean SSIM and 1.77 for the mean PSNR on the LOL and LOL_v2 datasets.

4.5. Night Face Detection

One of the main applications of high-speed enhancement methods is to extend the applicable scenarios for advanced image tasks such as target detection and target tracking. The improvement of the detection precision by each enhancement method was tested on the night face detection task. The Dual Shot Face Detector [48] (DSFD) with the official checkpoint (trained on the WIDER FACE dataset [49]) was used as the detector. One hundred low-light images with their localization labels were randomly selected from the training set of the DARK FACE dataset to form the testing set. The original low-light input image was first processed by each enhancement method and then underwent target detection. The evaluation tool (https://github.com/Ir1d/DARKFACE_eval_tools (accessed on 10 May 2024)) was used to draw the AP results at the IoU threshold of 0.5, and the results are shown in Figure 8 (label the scheme in which images are detected directly without enhancement as “ori”). As can be seen from the figure, the LLIE method can effectively improve the detection performance, and the effect of ours is similar to ZERO_DCE and is better than other methods. The total processing time is equivalent to the sum of the processing time of each enhancement method and the same DSFD detection time. Our ZRCRN is still the fastest.

5. Discussion

The ZRCRN proposed in this paper establishes a double-layer convolutional neural network based on the CRM to extract exposure ratio parameters to achieve LLIE. Combined with the two new loss functions, the ZRCRN can effectively improve efficiency without requiring references from the dataset. The sigmoid CRF and the corresponding CRM can be used in the learning models directly. Inversely transforming the input low-light image back to the incident radiation map helps to obtain high-quality enhancement with a more streamlined model. The contrast-preserving brightness loss and edge-preserving smoothness loss can also be used to constrain the training of other enhancement models. Appropriate enhancements of the input image features in the loss function combined with the excellent fitting ability of the learning models may bring about effective performance improvements. In addition, with the advantage of high efficiency, the ZRCRN might be an appropriate pre-processing method in object detection. But, the precision is still not very good. More diverse training data could help. Further work will be carried out to address this issue.

6. Conclusions

In this paper, we propose a high-speed zero-reference LLIE method, namely, the ZRCRN. The ZRCRN is established based on the CRM to achieve enhancement. High-quality outputs can be obtained with only one transformation using the BTF. The inverse CRF transformation reduces the need for a complex model structure in the extraction of enhancement parameters. Both can effectively improve efficiency. Contrast-preserving brightness loss and edge-preserving smoothness loss are designed to constrain the training. They do not need references from datasets and directly generate the expected targets from the features of low-light images, which can further retain the key information in the original input and is conducive to improving the quality of the enhancement. The experimental results on several LLIE datasets and the DARK FACE face detection dataset have revealed the advantages of our method when compared with several state-of-the-art methods. More work will be carried out to further improve precision and explore its application in object detection. Our source code is available on the project website.

Author Contributions

Conceptualization, X.W.; methodology, X.W.; software, X.W.; validation, T.N. and M.L.; writing—original draft preparation, X.W.; writing—review and editing, X.W., T.N., C.H., X.L., M.L. and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62105328).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The corresponding code is openly available at https://github.com/kazei7/ZRCRN (accessed on 10 May 2024). The LOL, LOL_v2, DICM, LIME and MEF datasets used in the research are available from the website https://github.com/Li-Chongyi/Lighting-the-Darkness-in-the-Deep-Learning-Era-Open (accessed on 10 May 2024). The DARK FACE dataset used in the research is available from the website https://flyywh.github.io/CVPRW2019LowLight/ (accessed on 10 May 2024).

Acknowledgments

The authors would like to thank C.W., Y.W., L.C., G.J., M.K., Y.W. and C.L. for providing the data.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ibrahim, H.; Kong, N.S.P. Brightness Preserving Dynamic Histogram Equalization for Image Contrast Enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  2. Chen, S.D.; Ramli, A.R. Minimum mean brightness error bi-histogram equalization in contrast enhancement. IEEE Trans. Consum. Electron. 2004, 49, 1310–1319. [Google Scholar] [CrossRef]
  3. Lu, L.; Zhou, Y.; Panetta, K.; Agaian, S. Comparative study of histogram equalization algorithms for image enhancement. Proc. SPIE Int. Soc. Opt. Eng. 2010, 7708, 770811-1. [Google Scholar]
  4. Ren, X.; Yang, W.; Cheng, W.H.; Liu, J. LR3M: Robust Low-Light Enhancement via Low-Rank Regularized Retinex Model. IEEE Trans. Image Process. 2020, 29, 5862–5876. [Google Scholar] [CrossRef] [PubMed]
  5. Li, M.; Liu, J.; Yang, W.; Sun, X.; Guo, Z. Structure-Revealing Low-Light Image Enhancement Via Robust Retinex Model. IEEE Trans. Image Process. 2018, 27, 2828–2841. [Google Scholar] [CrossRef]
  6. Seonhee, P.; Byeongho, M.; Seungyong, K.; Soohwan, Y.; Joonki, P. Low-light image enhancement using variational optimization-based Retinex model. In Proceedings of the 2017 IEEE International Conference on Consumer Electronics (ICCE), Berlin, Germany, 3–6 September 2017; pp. 70–71. [Google Scholar] [CrossRef]
  7. Loza, A.; Bull, D.R.; Hill, P.R.; Achim, A.M. Automatic contrast enhancement of low-light images based on local statistics of wavelet coefficients. Digit. Signal Process. 2013, 23, 1856–1866. [Google Scholar] [CrossRef]
  8. Malm, H.; Oskarsson, M.; Warrant, E.; Clarberg, P.; Hasselgren, J.; Lejdfors, C. Adaptive enhancement and noise reduction in very low light-level video. In Proceedings of the 2007 IEEE 11th International Conference on Computer Vision, Rio De Janeiro, Brazi, 14–21 October 2007; pp. 1–8. [Google Scholar] [CrossRef]
  9. Dong, X.; Wang, G.; Pang, Y.; Li, W.; Wen, J.; Meng, W.; Lu, Y. Fast efficient algorithm for enhancement of low lighting video. In Proceedings of the 2011 IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar] [CrossRef]
  10. Muhammad, N.; Khan, H.; Bibi, N.; Usman, M.; Ahmed, N.; Khan, S.N.; Mahmood, Z. Frequency component vectorisation for image dehazing. J. Exp. Theor. Artif. Intell. 2021, 33, 919–932. [Google Scholar] [CrossRef]
  11. Zhu, Y.; Jia, Z.; Yang, J.; Kasabov, N.K. Change detection in multitemporal monitoring images under low illumination. IEEE Access 2020, 126700–126712. [Google Scholar] [CrossRef]
  12. Ren, Y.R.; Ying, Z.Q.; Li, T.H.; Li, G. LECARM: Low-Light Image Enhancement Using the Camera Response Model. IEEE Trans. Circuits Syst. Video Technol. 2019, 29, 968–981. [Google Scholar] [CrossRef]
  13. Guo, X.J.; Li, Y.; Ling, H.B. LIME: Low-Light Image Enhancement via Illumination Map Estimation. IEEE Trans. Image Process. 2017, 26, 982–993. [Google Scholar] [CrossRef]
  14. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A Deep Autoencoder Approach to Natural Low-light Image Enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  15. Yi, W.; Dong, L.; Liu, M.; Hui, M.; Kong, L.; Zhao, Y. SID-Net: Single image dehazing network using adversarial and contrastive learning. Multimed. Tools Appl. 2024, 83, 71619–71638. [Google Scholar] [CrossRef]
  16. Khan, H.; Xiao, B.; Li, W.; Muhammad, N. Recent advancement in haze removal approaches. Multimed. Syst. 2022, 28, 687–710. [Google Scholar] [CrossRef]
  17. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. EnlightenGAN: Deep Light Enhancement Without Paired Supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
  18. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-Reference Deep Curve Estimation for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 1777–1786. [Google Scholar] [CrossRef]
  19. Li, C.; Guo, C.; Loy, C.C. Learning to Enhance Low-Light Image via Zero-Reference Deep Curve Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 4225–4238. [Google Scholar] [CrossRef]
  20. Xiang, S.; Wang, Y.; Deng, H.; Wu, J.; Yu, L. Zero-shot Learning for Low-light Image Enhancement Based on Dual Iteration. J. Electron. Inf. Technol. 2022, 44, 3379–3388. [Google Scholar] [CrossRef]
  21. Zheng, S.; Gupta, G. Semantic-Guided Zero-Shot Learning for Low-Light Image/Video Enhancement. In Proceedings of the 22nd IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2022; pp. 581–590. [Google Scholar] [CrossRef]
  22. Xia, Y.; Xu, F.; Zheng, Q. Zero-shot Adaptive Low Light Enhancement with Retinex Decomposition and Hybrid Curve Estimation. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Gold Coast, Australia, 18–23 June 2023. [Google Scholar] [CrossRef]
  23. Grossberg, M.D.; Nayar, S.K. Modeling the space of camera response functions. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 1272–1282. [Google Scholar] [CrossRef] [PubMed]
  24. Eilertsen, G.; Kronander, J.; Denes, G.; Mantiuk, R.K.; Unger, J. HDR image reconstruction from a single exposure using deep CNNs. ACM Trans. Graph. 2017, 36. [Google Scholar] [CrossRef]
  25. Wu, Y.; Liu, F. Zero-shot contrast enhancement and denoising network for low-light images. Multimed. Tools Appl. 2023, 83, 4037–4064. [Google Scholar] [CrossRef]
  26. Tian, J.; Zhang, J. A Zero-Shot Low Light Image Enhancement Method Integrating Gating Mechanism. Sensors 2023, 23, 7306. [Google Scholar] [CrossRef] [PubMed]
  27. Kar, A.; Dhara, S.K.; Sen, D.; Biswas, P.K. Zero-shot Single Image Restoration through Controlled Perturbation of Koschmieder’s Model. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16200–16210. [Google Scholar] [CrossRef]
  28. Xie, C.; Tang, H.; Fei, L.; Zhu, H.; Hu, Y. IRNet: An Improved Zero-Shot Retinex Network for Low-Light Image Enhancement. Electronics 2023, 12, 3162. [Google Scholar] [CrossRef]
  29. Zhang, Q.; Zou, C.; Shao, M.; Liang, H. A Single-Stage Unsupervised Denoising Low-Illumination Enhancement Network Based on Swin-Transformer. IEEE Access 2023, 11, 75696–75706. [Google Scholar] [CrossRef]
  30. Kendall, A.; Gal, Y.; Cipolla, R. Multi-Task Learning Using Uncertainty to Weigh Losses for Scene Geometry and Semantics. In Proceedings of the 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar] [CrossRef]
  31. Sener, O.; Koltun, V. Multi-Task Learning as Multi-Objective Optimization. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS), Montréal, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  32. Canny, J. A Computational approach to edge-detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 679–698. [Google Scholar] [CrossRef] [PubMed]
  33. Cai, J.; Gu, S.; Zhang, L. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef] [PubMed]
  34. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. 2017. Available online: https://xxx.lanl.gov/abs/1412.6980 (accessed on 24 June 2024).
  35. Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Restarts. 2016. Available online: http://xxx.lanl.gov/abs/1608.03983 (accessed on 24 June 2024).
  36. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse Gradient Regularized Deep Retinex Network for Robust Low-Light Image Enhancement. IEEE Trans. Image Process. 2021, 30, 2072–2086. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  38. Hao, S.J.; Han, X.; Guo, Y.R.; Xu, X.; Wang, M. Low-Light Image Enhancement With Semi-Decoupled Decomposition. IEEE Trans. Multimed. 2020, 22, 3025–3038. [Google Scholar] [CrossRef]
  39. Zhang, F.; Li, Y.; You, S.D.; Fu, Y. Learning Temporal Consistency for Low Light Video Enhancement from Single Images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 4965–4974. [Google Scholar] [CrossRef]
  40. Wu, W.H.; Weng, J.; Zhang, P.P.; Wang, X.; Yang, W.H.; Jiang, J.M.; Ieee Comp, S.O.C. URetinex-Net: Retinex-based Deep Unfolding Network for Low-light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5891–5900. [Google Scholar] [CrossRef]
  41. Ma, L.; Ma, T.Y.; Liu, R.S.; Fan, X.; Luo, Z.X.; Ieee Comp, S.O.C. Toward Fast, Flexible, and Robust Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 5627–5636. [Google Scholar] [CrossRef]
  42. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. 2018. Available online: http://xxx.lanl.gov/abs/1808.04560 (accessed on 24 June 2024).
  43. Lee, C.; Lee, C.; Kim, C.S. Contrast enhancement based on layered difference representation. In Proceedings of the 19th IEEE International Conference on Image Processing (ICIP), Orlando, FL, USA, 30 September–3 October 2012; pp. 965–968. [Google Scholar]
  44. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 24, 3345–3356. [Google Scholar] [CrossRef] [PubMed]
  45. Yang, W.; Yuan, Y.; Ren, W.; Liu, J.; Scheirer, W.J.; Wang, Z.; Zhang, T.; Zhong, Q.; Xie, D.; Pu, S.; et al. Advancing Image Understanding in Poor Visibility Environments: A Collective Benchmark Study. IEEE Trans. Image Process. 2020, 29, 5737–5752. [Google Scholar] [CrossRef] [PubMed]
  46. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef] [PubMed]
  47. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  48. Li, J.; Wang, Y.; Wang, C.; Tai, Y.; Qian, J.; Yang, J.; Wang, C.; Li, J.; Huang, F. DSFD: Dual Shot Face Detector. In Proceedings of the 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 5055–5064. [Google Scholar] [CrossRef]
  49. Yang, S.; Luo, P.; Loy, C.C.; Tanal, X. WIDER FACE: A Face Detection Benchmark. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 5525–5533. [Google Scholar] [CrossRef]
Figure 1. The relationship between the pixel values of the image and the incident radiation in the same scene under two different exposure conditions.
Figure 1. The relationship between the pixel values of the image and the incident radiation in the same scene under two different exposure conditions.
Sensors 24 05019 g001
Figure 2. The framework of the ZRCRN. f is the CRF, and g is the BTF. The parameter-generating network includes inverse CRF transformation and convolutional model extraction with two main operations. It extracts the parameter K from the input low-light image P 0 . The BTF calls K to transform P 0 to the enhancement result P 1 .
Figure 2. The framework of the ZRCRN. f is the CRF, and g is the BTF. The parameter-generating network includes inverse CRF transformation and convolutional model extraction with two main operations. It extracts the parameter K from the input low-light image P 0 . The BTF calls K to transform P 0 to the enhancement result P 1 .
Sensors 24 05019 g002
Figure 3. The structure of the PGN. The number at the bottom represents the number of channels of the corresponding image or convolution layer. f 1 indicating the inverse CRF transformation. The low-light image ( P 0 ) acquired by the camera first passes through inverse CRF transformation to obtain the low-light incident radiation map ( E 0 ) and then passes through a series of layers to obtain the enhancement-required single-channel pixel-wise parameter diagram ( K ).
Figure 3. The structure of the PGN. The number at the bottom represents the number of channels of the corresponding image or convolution layer. f 1 indicating the inverse CRF transformation. The low-light image ( P 0 ) acquired by the camera first passes through inverse CRF transformation to obtain the low-light incident radiation map ( E 0 ) and then passes through a series of layers to obtain the enhancement-required single-channel pixel-wise parameter diagram ( K ).
Sensors 24 05019 g003
Figure 4. The training loss and the validation SSIM during training.
Figure 4. The training loss and the validation SSIM during training.
Sensors 24 05019 g004
Figure 5. A comparison of the ablation of loss functions. The left side of “-” in each title indicates the state of brightness loss (Our: contrast-preserving brightness loss; Const: fixed-target-exposure loss; No: not use), and the right side indicates the state of smoothness loss.
Figure 5. A comparison of the ablation of loss functions. The left side of “-” in each title indicates the state of brightness loss (Our: contrast-preserving brightness loss; Const: fixed-target-exposure loss; No: not use), and the right side indicates the state of smoothness loss.
Sensors 24 05019 g005
Figure 6. The qualitative comparison of eight methods on FR datasets with the input and ground truth (GT).
Figure 6. The qualitative comparison of eight methods on FR datasets with the input and ground truth (GT).
Sensors 24 05019 g006
Figure 7. The qualitative comparison of eight methods on NR datasets with the input.
Figure 7. The qualitative comparison of eight methods on NR datasets with the input.
Sensors 24 05019 g007
Figure 8. A comparison of enhancement methods on the DARK FACE dataset. “ori” indicates no enhancement. Targets were detected by DSFD from the outputs of each enhancement method to obtain the metric.
Figure 8. A comparison of enhancement methods on the DARK FACE dataset. “ori” indicates no enhancement. Targets were detected by DSFD from the outputs of each enhancement method to obtain the metric.
Sensors 24 05019 g008
Table 1. Comparison for parameter extraction network.
Table 1. Comparison for parameter extraction network.
SSIMPSNRTIME/ms
m10.4412.320.41
m2-n10.6917.990.50
m2-n20.6918.200.51
m2-n40.6817.740.61
m3-n10.5415.800.56
m3-n20.6116.980.59
m3-n40.6116.980.73
m4-n10.5917.280.64
m4-n20.6017.480.64
m4-n40.6817.670.87
Table 2. Comparison for contrast-preserving brightness loss.
Table 2. Comparison for contrast-preserving brightness loss.
SSIMPSNRSSIM-EXPTAR2BONL
a = 0.00.6717.130.4
a = 0.10.6917.730.65
a = 0.20.6918.200.73
a = 0.30.6918.090.73
a = 0.40.6817.840.69
a = 0.50.6416.940.65
Table 3. Comparison for edge-preserving smoothness loss.
Table 3. Comparison for edge-preserving smoothness loss.
SSIMPSNR
γ = 1 , β = 0.304 0.6917.78
γ = 1 , β = 0.707 0.6817.33
γ = 1 , β = 1.414 0.6716.85
γ = 1 , β = 2.828 0.4311.67
γ = 2 , β = 0.304 0.6917.98
γ = 2 , β = 0.707 0.6918.14
γ = 2 , β = 1.414 0.6918.20
γ = 2 , β = 2.828 0.6918.01
γ = 3 , β = 0.304 0.6917.85
γ = 3 , β = 0.707 0.6917.94
γ = 3 , β = 1.414 0.6918.05
γ = 3 , β = 2.828 0.6917.87
Table 4. Results of ablation of inverse transformation.
Table 4. Results of ablation of inverse transformation.
SSIMPSNR
With inverse0.6918.20
Without inverse0.6517.2
Table 5. Results of ablation of loss functions.
Table 5. Results of ablation of loss functions.
SSIMPSNR
OurExp-OurSmooth0.6918.20
OurExp-NoSmooth0.6918.00
ConstExp-OurSmooth0.6817.93
NoExp-OurSmooth0.2810.02
Table 6. Quantitative comparison for quality on two FR datasets.
Table 6. Quantitative comparison for quality on two FR datasets.
LOLLOL_v2Mean
SSIMPSNRSSIMPSNRSSIMPSNR
LECARM0.6714.420.6717.660.6716.04
SDD0.6713.250.716.510.6914.88
StableLLVE0.7817.250.7819.80.7818.52
URetinex_Net0.8619.550.8520.660.8520.11
SCI0.6313.80.6317.30.6315.55
EnlightenGAN0.7417.440.7418.620.7418.03
Zero_DCE0.714.840.6918.130.6916.49
Ours0.7216.440.6918.20.7117.32
Table 7. Quantitative comparison for quality on three NR datasets.
Table 7. Quantitative comparison for quality on three NR datasets.
DICMLIMEMEFMean
BRISQUENIQEBRISQUENIQEBRISQUENIQEBRISQUENIQE
LECARM26.704.2420.973.9919.743.0022.473.74
SDD30.743.9427.094.0129.733.8629.193.94
StableLLVE33.734.2429.124.2034.114.3032.324.25
URetinex_Net28.044.0726.384.4125.143.7326.524.07
SCI23.043.6720.684.1719.243.1620.993.67
EnlightenGAN23.803.5019.623.5421.063.0721.493.37
Zero_DCE25.043.5722.173.9021.093.0322.773.50
Ours21.833.4022.393.8718.253.2320.823.50
Table 8. Quantitative comparison for efficiency.
Table 8. Quantitative comparison for efficiency.
FLOPs/G#Params/kTIME/ms
LECARM--71.310
SDD--5952.140
StableLLVE165.3164316.2593.306
URetinex_Net939.912340.10589.750
SCI0.2720.2580.676
EnlightenGAN275.0208636.67519.390
Zero_DCE85.76979.4161.300
Ours0.0810.0750.518
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Huang, L.; Li, M.; Han, C.; Liu, X.; Nie, T. Fast, Zero-Reference Low-Light Image Enhancement with Camera Response Model. Sensors 2024, 24, 5019. https://doi.org/10.3390/s24155019

AMA Style

Wang X, Huang L, Li M, Han C, Liu X, Nie T. Fast, Zero-Reference Low-Light Image Enhancement with Camera Response Model. Sensors. 2024; 24(15):5019. https://doi.org/10.3390/s24155019

Chicago/Turabian Style

Wang, Xiaofeng, Liang Huang, Mingxuan Li, Chengshan Han, Xin Liu, and Ting Nie. 2024. "Fast, Zero-Reference Low-Light Image Enhancement with Camera Response Model" Sensors 24, no. 15: 5019. https://doi.org/10.3390/s24155019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop