Next Article in Journal
Measurement and Estimation of Spectral Sensitivity Functions for Mobile Phone Cameras
Previous Article in Journal
Dipodal Tetraamide Derivatives of 1,10-Diaza-18-Crown-6 and Alkylmalonic Acids—Synthesis and Use as Ionophores in Ion Selective Membrane Electrodes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Low-Light Image Enhancement Based on Multi-Path Interaction

1
School of Microelectronics, Tianjin University, Tianjin 300072, China
2
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
3
National Ocean Technology Center, Tianjin 300112, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(15), 4986; https://doi.org/10.3390/s21154986
Submission received: 22 June 2021 / Revised: 16 July 2021 / Accepted: 19 July 2021 / Published: 22 July 2021
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Due to the non-uniform illumination conditions, images captured by sensors often suffer from uneven brightness, low contrast and noise. In order to improve the quality of the image, in this paper, a multi-path interaction network is proposed to enhance the R, G, B channels, and then the three channels are combined into the color image and further adjusted in detail. In the multi-path interaction network, the feature maps in several encoding–decoding subnetworks are used to exchange information across paths, while a high-resolution path is retained to enrich the feature representation. Meanwhile, in order to avoid the possible unnatural results caused by the separation of the R, G, B channels, the output of the multi-path interaction network is corrected in detail to obtain the final enhancement results. Experimental results show that the proposed method can effectively improve the visual quality of low-light images, and the performance is better than the state-of-the-art methods.

1. Introduction

With the development of computer technology and camera sensors, computer vision has been applied in various engineering fields—for example, object detection in autonomous vehicles [1] and harvesting robots [2], detection and monitoring in the field of civil engineering [3,4], video surveillance [5], 3D reconstruction [6] and so on. Since vision tasks play an important role in a wide range of fields, reliable working performance is required. However, the tasks rely on scene illumination, and the performance of any camera-sensor-based perception tasks is highly degraded in poor illumination conditions such as low-light scenes [7]. In low-light scenes, when a camera cannot receive sufficient light or the camera sensor is not sufficiently sensitive, the captured images may have problems such as poor visualization and low image quality, making the valid information of the image disturbed and limiting the use of the image in computer vision tasks [8,9]. As we know, the degradation of the low-light images captured in a non-uniform illumination environment results in severe object information loss and makes the object detection more challenging [10]. The camera’s night mode sometimes suppresses this degradation; however, a slight shake may introduce other problems such as blurring. Improving the illumination of the environment or updating the camera sensor is not feasible in some conditions [11]. Therefore, low-light image enhancement methods at the software end are needed.
At present, a large number of image enhancement methods have been proposed. Histogram equalization (HE)-based methods [12,13] redistribute pixel values according to the cumulative distribution function of the input image to expand the dynamic range. For example, Ibrahim et al. [14] smoothed the input histogram with a one-dimensional Gaussian filter, and then partitioned the smoothed histogram based on its local maximums. After each partition was assigned to a new dynamic range, the histogram equalization process was applied independently to these partitions. The last step in this method was to normalize the output image to the input mean brightness. Ying et al. [15] and Ren et al. [16] utilized the input image and camera response model to adjust the pixel values. The methods based on Retinex theory [17] adaptively adjusted the illuminance and reflectance components of the image, where the reflectance component was considered as an inherent attribute of the scene and was unchangeable in different lighting conditions [18,19]. Jobson et al. [20] extended a previously designed single-scale center/surround Retinex to a multiscale version that achieved simultaneous dynamic range compression/color consistency/lightness rendition. In order to correct the deficiency present in the extension, a method of color restoration, at the cost of a modest dilution in color consistency, was defined. Fu et al. [21] derived two inputs that represented luminance-improved and contrast-enhanced versions of the decomposed illumination using the sigmoid function and adaptive histogram equalization, and then fused the derived inputs with the corresponding weights in a multiscale fashion to adjust illumination. The method combined the advantages of sigmoid function and histogram equalization, and the final enhanced image was obtained by compensating the adjusted illumination back to the reflectance. Dong et al. [22] noticed that the inverted low-light images intuitively resembled images acquired in hazy lighting conditions; thus, low-lighting image enhancement has much in common with video haze removal. Therefore, they applied image de-hazing algorithm to the inverted image to enhance the image. These methods are simple and effective. However, the results may have undesirable illumination and amplified noise.
In recent years, with the improvement of computer performance and the establishment of publicly available datasets, image enhancement methods based on convolutional neural networks (CNNs) have been actively researched. The CNN-based method is one of the data-driven methods and uses paired images for end-to-end learning. Wei et al. [23] proposed a Retinex-Net learned on a real dataset, which includes a Decom-Net to decompose low-light images into illumination and reflectance components and an Enhance-Net to adjust the illumination component. Xu et al. [24] observed that noise exhibits different levels of contrast in different frequency layers, and it is much easier to detect noise in the low-frequency layer than in the high-frequency one. Therefore, they proposed a network that learns to recover image objects in the low-frequency layer and then enhances high-frequency details based on the recovered image objects. Chen et al. [25] used an exposure prediction network to generate under-/overexposure images and then fused them with the input image to obtain the enhanced image. Lv et al. [26] proposed a multi-branch network to extract rich features of different levels and then fused the multi-branch outputs to produce the output image. Wang et al. [27] considered the low-light image enhancement as a residual learning problem. They proposed a deep lightening network, which consists of several lightening back-projection blocks that perform lightening and darkening processes iteratively to learn the residual for normal-light estimations. Moreover, a feature aggregation block that adaptively fuses the results of different lightening back-projection blocks was designed to effectively utilize the local and global features. Ma et al. [11] transformed the original low-light image from the RGB to HSI color space and used the segmentation exponential method to process the saturation (S) while applying a specially designed deep convolutional neural network to enhance the intensity component (I). The final improved image could be obtained by going back to the original RGB space. Lore et al. [28] used a class of deep neural networks, a stacked sparse denoising autoencoder (SSDA), to enhance natural low-light images. They explored two types of deep architecture, including learning contrast-enhancement and denoising simultaneously, and learning contrast-enhancement and denoising sequentially. CNN-based methods are effective in preserving details and denoising. Nevertheless, existing methods may not perform well on color.
Another data-driven method is the generative adversarial network (GAN)-based method. Different from CNN-based methods, GAN-based methods do not require strictly paired images. They usually require careful selection of unpaired training data. Each GAN contains a generator to output enhanced images and a discriminator to determine whether the output produced by the generator is satisfactory. Jiang et al. [29] proposed to regularize the unpaired training using the information extracted from the input itself, and used a global-local discriminator structure to handle spatially varying light conditions in the input image, while adding the idea of self-regularization, which is implemented by both the self feature preserving loss and the self-regularized attention mechanism. Chen et al. [30] augmented the U-Net with global features and improved Wasserstein GAN (WGAN) with an adaptive weighting scheme, then used individual batch normalization layers for generators in two-way GANs to help generators better adapt to their own input distributions. The design improves the stability of GAN training for the application. Liu et al. [31] proposed a perceptual-details GAN utilizing ZeroDCE to initially recover illumination and combined a residual dense-block encoder–decoder structure to suppress noise while finely adjusting the illumination. In addition, the details were enhanced by using fractional differential gradient masks integrated into the discriminator. However, the generator may collapse due to the fact that the discriminator fails to discriminate its output, and it is difficult to obtain the desired output from two models with opposite objectives trained simultaneously [27].
In order to effectively enhance the brightness of low-light images while restoring the color and details, we propose an end-to-end learning method. The method consists of two cascaded subnetworks that first enhance the color channels and then adjust the details to obtain enhanced images with good color restoration. The enhanced images are expected to display improved visual quality and enhanced performance in computer vision tasks such as object detection and instance segmentation [32,33], and an example of text recognition is shown in Section 3.4. Overall, our contributions are as follows:
(1) The low-light image enhancement task is simplified into three steps. The first step is the enhancement of R, G, and B channels; then, the reconstruction of the color image is performed, and the last step is the adjustment of details.
(2) We design a multi-path interaction network (MPI-net) to enhance the R, G, and B channels. Then, through the interaction across the parallel paths, the feature maps are potentially more accurate.
(3) With the help of exposure amplification loss in the detail correction network (DC-net) and other losses, the final enhanced images are more natural. The experimental results demonstrate that our method outperforms several state-of-the-art enhancement methods.

2. Proposed Method

Networks inspired by U-Net [34] are often a single path from high-resolution to low-resolution for encoding, and low-resolution to high-resolution for decoding, where usually only the skip connections directly concatenate the feature maps in the downsampling layer to its corresponding upsampling layer according to space resolution to increase the amount of information in the upsampling steps [35]. In order to increase the information representation of feature maps in the network, we design a multi-path interaction network (MPI-net), which extends the network structure based on the U-Net idea to further enhance the information interaction between feature maps of different resolutions, while retaining a high-resolution path and enhancing the utilization of information in the network. Retaining the high-resolution path in the network, rather than upsampling to high-resolution from low-resolution, potentially leads to more accurate feature maps [36].
We consider the image enhancement as the enhancement of three channels. At first, the R, G, and B channels of low-light images are trained separately to obtain the enhanced R, G, and B channels and recombine them into a color image. Since the three channels are trained separately, the correlation between the color channels is ignored and the obtained images may have unnatural colors and overexposure. Therefore, a detail correction network (DC-net) is used after the multi-path interaction network (MPI-net) to further adjust the color images generated by the output of MPI-net. The DC-net consists of several convolutions, and the last layer of the convolution is a residual map. The enhanced images can be obtained by subtracting the residual maps from the color images generated by the output of MPI-net. The overall architecture of the proposed method is shown in Figure 1.

2.1. Multi-Path Interaction Network

2.1.1. Network

The first path of the multi-path interaction network (MPI-net) is a high-to-low and low-to-high resolution network (HL-net), and the number of HL-Net paths is gradually increased one by one to form more paths until the last path contains only high-resolution feature maps. The paths are connected in parallel, and the feature maps for the parallel paths of a later stage consist of the feature maps from the previous stage and an extra lower one. Meanwhile, there is a high-resolution path in the network. The architecture of MPI-net is shown in Figure 2. The MPI-net connects multiple paths to form a richer feature representation while retaining the ability of U-Net. At the same time, the existence of a high-resolution path and the interaction of information between feature maps of the same or different resolutions in different paths make the feature representation potentially more accurate [36]. In one path, each downsampling step is a convolution with stride 2. Each upsampling step contains a bilinear interpolation to expand the size of the feature map to twice the original. Moreover, three cascaded convolutional layers are included between two operations with different spatial resolutions. Each convolutional layer consists of a 3 × 3 convolution operation with padding, followed by a rectified linear unit (ReLU) activation function. In addition, skip connections directly concatenate the feature maps in the downsampling layer to the corresponding upsampling layer to increase the amount of information in the upsampling steps. The number of channels of feature maps with different resolutions in the first path is 32, 64, 128 and 256, respectively. Moreover, other paths are consistent with the first path in the number of feature map channels.
The exchange of information between feature maps of different resolutions leads to rich resolution representations [36]. Therefore, the exchange units are introduced across parallel paths in the MPI-net. An example is shown in Figure 3. Since the paths are connected in parallel, each path repeatedly receives the information from the other parallel paths. The feature maps for exchanging information are at the same depth in the network and usually have different resolutions. In the exchange unit, for different paths, they are transformed to the same resolution and concatenated on the path to complete the information exchange. Both upsampling and downsampling are used only once in one exchange unit.

2.1.2. Loss Function

The loss function of MPI-net consists of two components, the mean square error loss L m s e and the structural similarity loss L s s i m , expressed as follows:
L m p i n e t = L m s e m p i n e t + λ 1 L s s i m m p i n e t
where λ is used to control the image structure.
The mean squared error (MSE) is the average of the squared sum of the corresponding pixel errors between the enhanced channel and the reference channel, and is used to evaluate the overall difference between two channels. A smaller MSE means a better result. Therefore, the mean square error loss L m s e m p i n e t is defined as:
L m s e m p i n e t = 1 H × W c { R , G , B } | | I m p i _ c I r e f _ c | | 2 2
where I m p i _ c is the enhanced c channel, I r e f _ c is the c channel of the reference image, | | · | | 2 means L 2 norm, H and W are the height and width of the image.
The structural similarity (SSIM; [37]) is used to evaluate the similarity of two channels in terms of luminance, contrast and structure. The value of SSIM ranges from 0 to 1, and a larger value indicates better similarity. The definition of SSIM is as follows:
S S I M ( m p i _ c , r e f _ c ) = ( 2 μ m p i _ c μ r e f _ c + C 1 ) ( 2 σ m p i _ c , r e f _ c + C 2 ) ( μ m p i _ c 2 μ r e f _ c 2 + C 1 ) ( σ m p i _ c 2 + σ r e f _ c 2 + C 2 )
where the parameters mpi_c and ref_c are simple representations of the enhanced c channel and the reference c channel, μ m p i _ c is the mean of the I m p i _ c , μ r e f _ c is the mean of the I r e f _ c , σ m p i _ c is the variance of the I m p i _ c , σ r e f _ c is the variance of the I r e f _ c , σ m p i _ c , r e f _ c is the covariance of the I m i d _ c and the I r e f _ c , C 1 and C 2 are constants and take the default values. In order to improve the structural distortion problems that usually exist in low-light images [26], we introduce structural similarity loss L s s i m m p i n e t :
L s s i m m p i n e t = c { R , G , B } ( 1 S S I M ( m p i _ c , r e f _ c ) )

2.2. Detail Correction Network

2.2.1. Network

The enhanced R, G, B channels are concatenated to generate the preliminary enhanced image ( I m p i ). In order to avoid the loss of details caused by enhancing the color channels separately, the preliminary enhanced image and the low-light image are concatenated as the input of the detail correction network (DC-net) to adjust the details. As shown in Figure 4, the DC-net contains six 64-channel convolution layers, and the 3-channel feature map obtained from the last layer of convolution is a residual map. The final enhanced image is obtained by subtracting the residual map from the preliminary enhanced image. The activation function of the last convolutional layers in MPI-net and DC-net is none.

2.2.2. Loss Function

We introduce DC-net and design extra exposure amplification loss L e a and smoothing loss L s m o o t h to suppress overexposure and make the enhanced images more natural. The total loss function of DC-net is expressed as:
L d c n e t = L m s e d c n e t + L s s i m d c n e t + λ 2 L e a + λ 3 L s m o o t h
where λ 2 and λ 3 are used to control the degree of overexposure suppression and smoothing, respectively.
The L m s e d c n e t is expressed as:
L m s e d c n e t = 1 H × W | | I e n h I r e f | | 2 2
where I e n h is the final enhanced image and obtained by subtracting the residual map I r e s of DC-net from the preliminary enhanced image I m p i , and I r e f is the reference image.
The L s s i m d c n e t is expressed as:
L s s i m d c n e t = 1 S S I M ( e n h , r e f )
where the parameters enh and ref are simple representations of the I e n h and the I r e f .
Through the gamma transformation, the pixel difference between the bright areas of the enhanced image and the reference image could be greatly increased, while the pixel difference between the dark areas is slightly increased. Therefore, using the average of the pixel difference between the gamma-transformed enhanced image and the reference image as the loss function helps to place more emphasis on the bright area and suppress overexposure. The exposure amplification loss L e a is defined as follows:
L e a = 1 H × W | | ( I e n h ) γ ( I r e f ) γ | | 1
where | | · | | 1 means L 1 norm, and γ is used to control the increase in the relative difference.
To smooth the enhanced image and make it more natural, we introduce the smoothing loss L s m o o t h to minimize the difference between the horizontal and vertical gradients of the enhanced image and the reference image in the color channels. The definition of smoothing loss L s m o o t h is shown below:
L s m o o t h = 1 H × W c { R , G , B } ( | | I e n h x c I r e f x c | | 2 2 + | | I e n h y c I r e f y c | | 2 2 )
where I e n h x c is the horizontal gradient of the enhanced image I e n h in channel c, I r e f x c is the horizontal gradient of the reference image I r e f in channel c, I e n h y c is the vertical gradient of the enhanced image I e n h in channel c, I r e f y c is the vertical gradient of the reference image I r e f in channel c.
Figure 5 shows an example of the images used and generated in the proposed method, including the input images and its R, G, and B channels, the R, G, and B channels enhanced by the MPI-net and the preliminary enhanced images I m p i , the residual maps I r e s generated by the DC-net, and the obtained enhanced images I e n h .

3. Experiments

3.1. Training Details and Dataset

The experiments were carried out using Tensorflow 1.14.0, on a workstation with Intel(R) Xeon(R) CPU E5-2186 @ 3.80 GHz, Nvidia GeForce GTX 2080TI and 64 G RAM. The parameters λ 1 , λ 2 , λ 3 , and γ were set to 3, 1.3, 5, and 5 experimentally. The training images were normalized to [0, 1] and randomly cropped to patches with size of 48 × 48 . The Adam optimizer was used with default parameters and the training epochs for the two subnetworks were both set to 30. For the learning rate, we first initialized it to 0.001 and reduced it by 10 times every 10 epochs. The training could be completed within 5 min.
The training dataset was from the LOL dataset, which is a real dataset containing a training dataset with 485 image pairs and a testing dataset with 15 image pairs. The scenes in the dataset are rich, and the image resolution is 600 × 400 . We selected 234 image pairs of different scenes from the training dataset of the LOL-dataset as the new training dataset, and used the 15 images from the testing dataset of the LOL dataset and another 8 images from the LOL dataset (outside our training dataset) and SICE [38] as the new testing dataset. In addition, images from the LIME [19,39] were also selected to further demonstrate the effectiveness of the proposed model.

3.2. Comparison with State-of-the-Art Methods

We compared the proposed method with state-of-the-art methods: BIMEF [15], LECA RM [16], MSRCR [20], LIME [19], MF [21], and Retinex-Net [23]. In addition to analyzing the visual quality of the experimental results, we also adopted peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and natural image quality evaluator (NIQE) [40] to evaluate the image quality.

3.2.1. Visual Quality Comparison

We performed experiments on images with different lighting conditions. The images were from the LOL dataset (outside our training dataset), LIME, and SICE [38,39], and the results are shown in Figure 6 and Figure 7. As shown in Figure 6, although some methods such as LIME have good brightness in local areas, they amplify the noise at the same time. More details can be seen in the last row of Figure 6, where the enhanced images of the proposed method are smooth and noise-free. The brightness of the images enhanced by BIMEF and LECARM is insufficient. The MSRCR has unsatisfactory performance in the contrast of images. In Figure 7, the results of LIME are unnatural in bright areas. Over-enhancement of the input image distorts the color of MF and Retinex-Net. Relatively speaking, the proposed method effectively enhances the brightness and the enhanced images are the most natural.

3.2.2. Evaluation

For a fair comparison, we use the representative metrics PSNR (peak signal-to-noise ratio), SSIM (structural similarity), and NIQE (natural image quality evaluator) to evaluate the image quality of the enhanced images. The PSNR can detect whether an image is distorted. The SSIM measures image similarity from the three aspects of luminance, contrast, and structure. The NIQE is a non-reference image quality evaluation method. We use the average of the images in the testing dataset as the test value and the results are shown in Table 1. The larger the value of PSNR and SSIM, and the smaller the value of NIQE, the better the results. The best results of PSNR, SSIM, and NIQE are bolded in this paper. It can be seen that the proposed method outperforms the other methods in all three metrics. Compared to the best results of the state-of-the-art methods, the proposed method offers a 12.682% improvement in PSNR (compared to LIME), 22.930% improvement in SSIM (compared to BIMEF), and 27.701% improvement in NIQE (compared to BIMEF). This means that the images obtained by the proposed method have the best quality.

3.3. Ablation Study

To demonstrate the effectiveness of MPI-net, DC-net, and loss function L e a , we conducted an ablation study and analyzed the experimental results. Specifically, we designed experiments: (a) removing the loss function L e a ; (b) comparing with the preliminary enhanced image I m p i . The visual comparison results are presented in Figure 8. As can be seen, without the loss function L e a , the bright area is easily overexposed, resulting in a loss of content. Moreover, the preliminary enhanced images I m p i have unsatisfactory performance in bright areas and details. However, the proposed method enhances the dark areas while suppressing overexposure, and the details are natural.
Table 2 illustrates the comparison results in PSNR, SSIM, and NIQE values. We can find that the use of loss function L e a could effectively improve the quality of the enhanced image and the DC-net is necessary. As the color channels are first enhanced separately, sometimes, the DC-net may not perform well and bring color distortion. Nevertheless, the proposed method has satisfactory performance in most scenarios.

3.4. Application

To further illustrate the effectiveness of the proposed method in improving the accuracy of the computer vision task, we tested our output on Google Vision API (https://cloud.google.com/vision/, accessed on 12 June 2021). The results are shown in Figure 9. As can be seen, the Google Vision API could accurately recognize the text from the enhanced image while recognizing errors when using low-light images. The original image is from SICE [38].

4. Conclusions

In this paper, a multi-path interaction network (MPI-net) is designed to enhance the R, G, and B channels separately, and then a detail correction network (DC-net) and corresponding loss functions are used to adjust the details. Thanks to the information interaction between different paths in the MPI-net, the feature maps are potentially more accurate. Moreover, the enhanced images are more natural after the adjustment of DC-net. We compare our method with the state-of-the-art methods, and the experimental results show that the proposed method has better performance. The evaluation metrics of the proposed method are also superior to the state-of-the-art methods. With the wide application of computer vision, it is becoming increasingly important to improve the performance of computer vision tasks in low-light conditions. Our future work will focus on improving the generalization ability of the enhancement model and the enhancement effect in extreme environments, as well as building a complete enhancement and object detection system for nighttime autonomous driving and video surveillance.

Author Contributions

Conceptualization, X.G. and J.W.; methodology, B.Z., X.G. and J.W.; software, B.Z. and L.Z.; formal analysis, X.G. and J.W.; investigation, L.Z. and B.Z.; resources, X.G. and J.W.; data curation, B.Z.; writing—original draft preparation, B.Z.; writing—review and editing, X.G., J.W., B.Z. and L.Z.; funding acquisition, X.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by funding from the Tianjin Intelligent Security Industry Chain Technology Adaptation and Application Project (Grant No. 18ZXZNGX00320).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhao, X.; Sun, P.; Xu, Z.; Min, H.; Yu, H. Fusion of 3D LIDAR and camera data for object detection in autonomous vehicle applications. IEEE Sens. J. 2020, 20, 4901–4913. [Google Scholar] [CrossRef] [Green Version]
  2. Li, J.; Tang, Y.; Zou, X.; Lin, G.; Wang, H. Detection of fruit-bearing branches and localization of litchi clusters for vision-based harvesting robots. IEEE Access 2020, 8, 117746–117758. [Google Scholar] [CrossRef]
  3. Tang, Y.; Li, L.; Wang, C.; Chen, M.; Feng, W.; Zou, X.; Huang, K. Real-time detection of surface deformation and strain in recycled aggregate concrete-filled steel tubular columns via four-ocular vision. Robot. Comput. Integr. Manuf. 2019, 59, 36–46. [Google Scholar] [CrossRef]
  4. Tang, Y.; Chen, M.; Lin, Y.; Huang, X.; Huang, K.; He, Y.; Li, L. Vision-based three-dimensional reconstruction and monitoring of large-scale steel tubular structures. Adv. Civ. Eng. 2020, 2020, 1236021. [Google Scholar] [CrossRef]
  5. Zhang, T.; Chowdhery, A.; Bahl, P.; Jamieson, K.; Banerjee, S. The design and implementation of a wireless video surveillance system. In Proceedings of the Annual International Conference on Mobile Computing and Networking, Paris, France, 7–11 September 2015; pp. 426–438. [Google Scholar]
  6. Chen, M.; Tang, Y.; Zou, X.; Huang, Z.; Zhou, H.; Chen, S. 3D global mapping of large-scale unstructured orchard integrating eye-in-hand stereo vision and SLAM. Comput. Electron. Agric. 2021, 187, 106237. [Google Scholar] [CrossRef]
  7. Rashed, H.; Ramzy, M.; Vaquero, V.; El Sallab, A.; Sistu, G.; Yogamani, S. FuseMODNet: Real-Time Camera and LiDAR Based Moving Object Detection for Robust Low-Light Autonomous Driving. In Proceedings of the IEEE International Conference on Computer Vision Workshop, Seoul, Korea, 27–28 October 2019; pp. 2393–2402. [Google Scholar]
  8. Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed photo enhancement using deep illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6842–6850. [Google Scholar]
  9. Ai, S.; Kwon, J. Extreme low-light image enhancement for surveillance cameras using attention U-Net. Sensors 2020, 20, 495. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Xu, X.; Wang, S.; Wang, Z.; Zhang, X.; Hu, R. Exploring Image Enhancement for Salient Object Detection in Low Light Images. ACM Trans. Multimed. Comput. Commun. Appl. 2021, 17, 1–19. [Google Scholar]
  11. Ma, S.; Ma, H.; Xu, Y.; Li, S.; Lv, C.; Zhu, M. A low-light sensor image enhancement algorithm based on HSI color model. Sensors 2018, 18, 3583. [Google Scholar] [CrossRef] [Green Version]
  12. Stark, J.A. Adaptive image contrast enhancement using generalizations of histogram equalization. IEEE Trans. Image Process. 2000, 9, 889–896. [Google Scholar] [CrossRef] [Green Version]
  13. Abdullah-Al-Wadud, M.; Kabir, M.H.; Dewan, M.A.A.; Chae, O. A dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 593–600. [Google Scholar] [CrossRef]
  14. Ibrahim, H.; Kong, N.S.P. Brightness preserving dynamic histogram equalization for image contrast enhancement. IEEE Trans. Consum. Electron. 2007, 53, 1752–1758. [Google Scholar] [CrossRef]
  15. Ying, Z.; Li, G.; Gao, W. A bio-inspired multi-exposure fusion framework for low-light image enhancement. arXiv 2017, arXiv:1711.00591. [Google Scholar]
  16. Ren, Y.; Ying, Z.; Li, T.H.; Li, G. LECARM: Low-light image enhancement using the camera response model. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 968–981. [Google Scholar] [CrossRef]
  17. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef]
  18. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2782–2790. [Google Scholar]
  19. Guo, X.; Li, Y.; Ling, H. LIME: Low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 2016, 26, 982–993. [Google Scholar] [CrossRef]
  20. Jobson, D.J.; Rahman, Z.; Woodell, G.A. A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [Google Scholar] [CrossRef] [Green Version]
  21. Fu, X.; Zeng, D.; Huang, Y.; Liao, Y.; Ding, X.; Paisley, J. A fusion-based enhancing method for weakly illuminated images. Signal Process. 2016, 129, 82–96. [Google Scholar] [CrossRef]
  22. Dong, X.; Wang, G.; Pang, Y.; Li, W.; Wen, J.; Meng, W.; Lu, Y. Fast efficient algorithm for enhancement of low lighting video. In Proceedings of the IEEE International Conference on Multimedia and Expo, Barcelona, Spain, 11–15 July 2011; pp. 1–6. [Google Scholar]
  23. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. In Proceedings of the British Machine Vision Conference, Newcastle, UK, 3–6 September 2018; pp. 1–12. [Google Scholar]
  24. Xu, K.; Yang, X.; Yin, B.; Lau, R.W.H. Learning to restore low-light images via decomposition-and-enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2281–2290. [Google Scholar]
  25. Chen, Y.; Yu, M.; Jiang, G.; Peng, Z.; Chen, F. End-to-end single image enhancement based on a dual network cascade model. J. Vis. Commun. Image Represent. 2019, 61, 284–295. [Google Scholar] [CrossRef]
  26. Lv, F.; Lu, F.; Wu, J.; Lim, C. MBLLEN: Low-Light Image/Video Enhancement Using CNNs. In Proceedings of the British Machine Vision Conference, Newcastle, UK, 3–6 September 2018; p. 220. [Google Scholar]
  27. Wang, L.W.; Liu, Z.S.; Siu, W.C.; Lun, D.P.K. Lightening network for low-light image enhancement. IEEE Trans. Image Process. 2020, 29, 7984–7996. [Google Scholar] [CrossRef]
  28. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef] [Green Version]
  29. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. IEEE Trans. Image Process. 2021, 30, 2340–2349. [Google Scholar] [CrossRef]
  30. Chen, Y.S.; Wang, Y.C.; Kao, M.H.; Chuang, Y.Y. Deep photo enhancer: Unpaired learning for image enhancement from photographs with gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2018; pp. 6306–6314. [Google Scholar]
  31. Liu, Y.; Wang, Z.; Zeng, Y.; Zeng, H.; Zhao, D. PD-GAN: Perceptual-Details GAN for Extremely Noisy Low Light Image Enhancement. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Toronto, ON, Canada, 13 May 2021; pp. 1840–1844. [Google Scholar]
  32. Guo, C.; Li, C.; Guo, J.; Loy, C.C.; Hou, J.; Kwong, S.; Cong, R. Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1780–1789. [Google Scholar]
  33. Lv, F.; Li, Y.; Lu, F. Attention guided low-light image enhancement with a large scale low-light simulation dataset. Int. J. Comput. Vis. 2021, 129, 2175–2193. [Google Scholar] [CrossRef]
  34. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  35. Jiang, Z.; Li, H.; Liu, L.; Men, A.; Wang, H. A Switched View of Retinex: Deep Self-Regularized Low-Light Image Enhancement. arXiv 2021, arXiv:2101.00603. [Google Scholar]
  36. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Cai, J.; Gu, S.; Zhang, L. Learning a deep single image contrast enhancer from multi-exposure images. IEEE Trans. Image Process. 2018, 27, 2049–2062. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, W.; Wei, C.; Yang, W.; Liu, J. GLADNet: Low-light enhancement network with global awareness. In Proceedings of the 13th IEEE International Conference on Automatic Face & Gesture Recognition, Xi’an, China, 15–19 May 2018; pp. 751–755. [Google Scholar]
  40. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 2012, 20, 209–212. [Google Scholar] [CrossRef]
Figure 1. The architecture of the proposed method. The enhancement process is divided into three steps: color channel enhancement, reconstruction, detail adjustment. In the color channel enhancement step, a subnetwork MPI-net enhances the R, G, B channels. In the reconstruction step, the enhanced R, G, B channels are concatenated in order to generate the preliminary enhanced image I m p i . In the detail adjustment step, we concatenate the I m p i and the input image as the input of DC-net, and the output of DC-net is a residual map. The final enhanced image I e n h is obtained by subtracting the residual map from the I m p i .
Figure 1. The architecture of the proposed method. The enhancement process is divided into three steps: color channel enhancement, reconstruction, detail adjustment. In the color channel enhancement step, a subnetwork MPI-net enhances the R, G, B channels. In the reconstruction step, the enhanced R, G, B channels are concatenated in order to generate the preliminary enhanced image I m p i . In the detail adjustment step, we concatenate the I m p i and the input image as the input of DC-net, and the output of DC-net is a residual map. The final enhanced image I e n h is obtained by subtracting the residual map from the I m p i .
Sensors 21 04986 g001
Figure 2. The architecture of MPI-net.
Figure 2. The architecture of MPI-net.
Sensors 21 04986 g002
Figure 3. Illustrating how the exchange unit aggregates the information for different paths. Feature maps from different paths are transformed to the same resolution and then concatenated.
Figure 3. Illustrating how the exchange unit aggregates the information for different paths. Feature maps from different paths are transformed to the same resolution and then concatenated.
Sensors 21 04986 g003
Figure 4. The architecture of the DC-net. The input is the concatenation of the L m p i and the low-light image, and the output is a residual map with detailed information.
Figure 4. The architecture of the DC-net. The input is the concatenation of the L m p i and the low-light image, and the output is a residual map with detailed information.
Sensors 21 04986 g004
Figure 5. The example of the images used and generated in the proposed method. (a): the input images; (bd): R, G, and B channels of the input images; (eg): the enhanced R, G, and B channels; (h): the preliminary enhanced images I m p i ; (i): the residual maps I r e s ; (j) the final enhanced images I e n h .
Figure 5. The example of the images used and generated in the proposed method. (a): the input images; (bd): R, G, and B channels of the input images; (eg): the enhanced R, G, and B channels; (h): the preliminary enhanced images I m p i ; (i): the residual maps I r e s ; (j) the final enhanced images I e n h .
Sensors 21 04986 g005
Figure 6. Visual comparison of the proposed method and the state-of-the-art methods on images from LOL dataset but outside our training dataset.
Figure 6. Visual comparison of the proposed method and the state-of-the-art methods on images from LOL dataset but outside our training dataset.
Sensors 21 04986 g006
Figure 7. Visual comparison of the proposed method and the state-of-the-art methods on images from LIME [19,39] and SICE [38].
Figure 7. Visual comparison of the proposed method and the state-of-the-art methods on images from LIME [19,39] and SICE [38].
Sensors 21 04986 g007
Figure 8. Comparison results of ablation study.
Figure 8. Comparison results of ablation study.
Sensors 21 04986 g008
Figure 9. Results of Google Cloud Vision API. (a) Recognition result of low-light image; (b) Recognition result of our enhanced image.
Figure 9. Results of Google Cloud Vision API. (a) Recognition result of low-light image; (b) Recognition result of our enhanced image.
Sensors 21 04986 g009
Table 1. Comparison of BIMEF, LECARM, MSRCR, MF, LIME, Retinex-Net and ours in PSNR, SSIM, and NIQE.
Table 1. Comparison of BIMEF, LECARM, MSRCR, MF, LIME, Retinex-Net and ours in PSNR, SSIM, and NIQE.
MethodsBIMEFLECARMMSRCRMFLIMERetinex-NetOurs
PSNR14.05015.35415.69217.10317.40217.17719.609
SSIM0.6280.5980.5960.5410.5750.5080.772
NIQE12.012712.875813.018614.012713.272015.83248.6851
Table 2. Comparison of performance metrics for ablation study.
Table 2. Comparison of performance metrics for ablation study.
MethodsWithout L ea I mpi Ours
PSNR19.14519.32119.609
SSIM0.7680.7700.772
NIQE9.13928.81368.6851
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, B.; Gong, X.; Wang, J.; Zhao, L. Low-Light Image Enhancement Based on Multi-Path Interaction. Sensors 2021, 21, 4986. https://doi.org/10.3390/s21154986

AMA Style

Zhao B, Gong X, Wang J, Zhao L. Low-Light Image Enhancement Based on Multi-Path Interaction. Sensors. 2021; 21(15):4986. https://doi.org/10.3390/s21154986

Chicago/Turabian Style

Zhao, Bai, Xiaolin Gong, Jian Wang, and Lingchao Zhao. 2021. "Low-Light Image Enhancement Based on Multi-Path Interaction" Sensors 21, no. 15: 4986. https://doi.org/10.3390/s21154986

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop