Next Article in Journal
A Surface-Enhanced Raman Scattering Substrate with Tunable Localized Surface Plasmon Resonance Absorption Based on AgNPs
Previous Article in Journal
Selection of X-ray Tube Settings for Relative Bone Density Quantification in the Knee Joint of Cats Using Computed Digital Absorptiometry
Previous Article in Special Issue
Redefining Accuracy: Underwater Depth Estimation for Irregular Illumination Scenes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Enhancement Based on Luminance Reconstruction by Multi-Resolution Fusion of RGB Channels

National Key Laboratory of Transient Physics, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5776; https://doi.org/10.3390/s24175776
Submission received: 26 July 2024 / Revised: 31 August 2024 / Accepted: 3 September 2024 / Published: 5 September 2024
(This article belongs to the Special Issue Underwater Vision Sensing System)

Abstract

:
Underwater image enhancement technology is crucial for the human exploration and exploitation of marine resources. The visibility of underwater images is affected by visible light attenuation. This paper proposes an image reconstruction method based on the decomposition–fusion of multi-channel luminance data to enhance the visibility of underwater images. The proposed method is a single-image approach to cope with the condition that underwater paired images are difficult to obtain. The original image is first divided into its three RGB channels. To reduce artifacts and inconsistencies in the fused images, a multi-resolution fusion process based on the Laplace–Gaussian pyramid guided by a weight map is employed. Image saliency analysis and mask sharpening methods are also introduced to color-correct the fused images. The results indicate that the method presented in this paper effectively enhances the visibility of dark regions in the original image and globally improves its color, contrast, and sharpness compared to current state-of-the-art methods. Our method can enhance underwater images in engineering practice, laying the foundation for in-depth research on underwater images.

1. Introduction

The oceans are rich in natural resources, and technological advancements have accelerated their exploration [1,2,3,4,5]. Underwater imaging is crucial for exploring and exploiting marine resources [2]. It has various applications, including detecting cracks in underwater dams [6], tracking fish [1,7], studying marine biodiversity studies [8,9,10], and conducting archeology [11].
Compared to atmospheric imaging, underwater images are less clear due to the presence of suspended particles in the water, which causes light to scatter more [12,13,14,15]. Additionally, water is a ‘sticky’ substance that absorbs the energy of light, causing more refraction and altering the direction of light propagation in underwater imaging. Underwater, different wavelengths of electromagnetic waves attenuate differently [16,17]. For instance, blue and green light can reach farther regions than red light, resulting in the hue of underwater images being biased towards blue and green [12,18]. The imaging depth of underwater images is reduced by the complexity of the underwater situation, resulting in a foggy appearance and low contrast colors due to a large color bias [19]. This is illustrated in Figure 1.
To optimize the quality of images captured underwater, two principal methodologies have been employed: improving the imaging hardware and implementing software techniques [20] for image processing. Hardware techniques [21] are typically employed in the initial research phase, which includes enhancing the underwater imaging environment, utilizing polarization lenses [22], modifying sensors, and so on. Image-processing techniques have strong practical significance in underwater image recovery and enhancement due to their simplicity, cost-effectiveness, and ability to overcome environmental constraints in contrast to hardware methods which are more complicated, power-consuming, expensive, and vulnerable to such constraints [23,24].
Given the high cost of hardware and the difficulty of acquiring underwater image pairs, our objective is to develop a straightforward and computationally inexpensive method that can perform relatively fast operations on common hardware to enhance the visibility of underwater images. This paper proposes an image enhancement method for underwater images based on single optical image fusion. Unlike previous methods, our approach does not require complex transformations, as shown in Figure 2.
Contributions. This paper introduces the following main contributions:
(1)
The hidden information of the image is excavated by considering the three RGB channels of the image as three images of different bands. These are used as the three inputs to reconstruct the image luminance by multi-resolution fusion. This method preserves the details of each input and enhances the visibility of the image.
(2)
Enhanced white balance is employed for the initial color correction and sharpening of the image. We have improved the original white balance method of restoring the red channel; in addition, the blue or green channel is color-corrected according to the appearance of the image.
(3)
Saliency detection techniques are applied to improve the hazy appearance of images. Saliency detection simulates human visual characteristics to make objects in the image stand out from the background, giving the underwater image a clearer visual effect.
The rest of the paper is structured as follows: Section 2 will present an overview of the current state-of-the-art development of key technologies for underwater image processing. Section 3 presents the method for reconstructing image luminance by fusing information from the three RGB channels. This includes generating the multi-resolution fusion weight map and adjusting the image color after luminance reconstruction. The following section presents a qualitative and quantitative comparison of the underwater image enhancement method proposed in this paper with the existing common underwater image enhancement methods.

2. Related Work

Various categorizations exist for underwater image enhancement methods. They can be classified into multi-image processing and single-image processing methods [25] based on the number of input images.
Multi-image processing methods include polarization imaging-based and multi-image scene recovery-based methods. The polarization imaging method aims to restore the original image by using multiple images acquired by rotating the camera’s polarizer at specific angles [26]. On the other hand, the method for recovering multi-image scenes involves constructing a digital map using multiple images captured from identical positions [27]. It should be noted that these methods do not apply to video acquisition. Additionally, the multi-image scene recovery methods are highly complex and not practical for general users.
Hence, the single-image-based underwater image-processing method has emerged as the main focus of the current research in the field of underwater image enhancement technology [12,17,25,28]. Researchers initially applied image dehazing and cloud occlusion reduction techniques from atmospheric image processing to enhance underwater images. This is because clouds and haze in the atmosphere refract, scatter, and attenuate visible light, thereby influencing the processing of underwater images. Expanding on these initial studies, methods for enhancing single underwater images have been developed to some extent. These methods can be broadly classified into three categories: physical model-based, non-physical model-based, and deep learning-based [29].

2.1. Physical Model-Based Methods

The physical model-based methods stem from an optical modeling perspective, considering underwater imaging as an inverse problem of optical imaging. These methods emphasize the utilization of models and prior knowledge to estimate the underwater image. The main workflow of these methods involves establishing a degradation physical model, parameter estimation, and model inversion.
According to Jaffe–McGlamery’s underwater imaging model [30], as illustrated in Figure 3, the underwater optical signal received by the camera contains three components: direct component, forward scattering component, and backward scattering component.
The Jaffe–McGlamery model is currently the most widely used in underwater imaging [15]. Trucco et al. [32] proposed a self-adaptive underwater image restoration filter based on a simplified Jaffe–McGlamery model in 2006. This filter aimed to reduce the effect of light scattering in underwater images. In 2012, CHIANG et al. [17] introduced the Wavelength Compensation and Image Dehazing (WCID) method, which improved the luminance of images and mitigated the haze phenomenon.
Building upon the DCP model [33] for atmospheric dehazing, Drews et al. [34] proposed in 2013 an underwater dark channel prior (UDCP) specifically designed for underwater environments. This approach allowed for a more accurate estimation of the original image in underwater environments. In 2015, Galdran et al. [35] presented a method based on the red channel dark before addressing color attenuation issues in the red spectral band. In 2018, Tang et al. [36] proposed an improved dark channel prior image enhancement method. In this method, the luminance of the transmission map was enhanced using an adaptive scale selection strategy, and further color correction was performed to eliminate color deviations. In the same year, Peng et al. [37] improved the DCP-based method (GDCP), which can be applied to the enhancement and restoration of images affected by fog, haze, sandstorms, and underwater environments, including both well-lit and dimly lit images.
In 2020, Hou et al. [38] established an underwater total variation (UTV) model relying on underwater dark channel prior. The proposed method can achieve a good performance on dehazing, contrast enhancement, edge preservation, and noise suppression. In 2021, Xie et al. [39] introduced a red channel before the guided variational framework for underwater image dehazing and deblurring. The proposed UNTV method is based on the complete underwater image formation model including the forward scattering component. Li et al. [40] introduced a novel dual high-order regularization-guided variational framework (UDHTV) for underwater image restoration in 2024. In contrast to the majority of the extant model-based methodologies, which fail to account for the local illumination variations present in a single image, this approach constructs an extended underwater image model that incorporates these differences.

2.2. Non-Physical Model-Based Methods

Non-physical model-based methods for underwater image enhancement do not require the consideration of the underwater image model. Non-physical model-based methods for underwater image enhancement include histogram equalization, Retinex-based approaches, and image fusion techniques.
Histogram-based methods enhance images by stretching pixel values in one or multiple color spaces using various techniques. In 2010, Iqbal et al. [41] improved the color and contrast of underwater images by extending the dynamic pixel range in both the RGB and HSV color spaces. In the same year, Ghani et al. [42,43,44] further improved Iqbal’s method by implementing dynamic grayscale instead of dynamic pixels for image enhancement.
In 2017, Fu et al. [45] proposed a simple yet effective two-step enhancing strategy that requires only a single underwater image to be enhanced. Each step deals with one of the two major issues of color distortion and low contrast. For the color distortion problem, they proposed a color-correcting strategy based on a piecewise linear transformation.
The Retinex theory aims to restore illumination information in an image based on the characteristics of the human visual system. This approach involves multi-scale filtering to extract lightness and reflectance components at different scales, which are then adjusted to enhance the image. Retinex-based methods are effective in improving the luminance, contrast, and color of underwater images. Fu et al. [46] proposed a Bayesian Retinex approach that uses multi-scale gradient priors of reflectance and illumination to enhance single underwater images. Zhang et al. [47] also utilized an extended multi-scale Retinex method for enhancing underwater images. Additionally, Tang et al. [48] applied Retinex to enhance underwater videos.
For the enhancement of underwater images, fusion techniques can be used to merge images from different sensors or with different exposure or color characteristics. This improves image quality and visual perception. Ancuti et al. [49] developed a fusion-based model for underwater image enhancement. The model uses the results of histogram equalization to weigh the white balance color improvement and bilateral filtering output.
In 2022, Li et al. [50] propose a novel scheme by constructing an adaptive color and contrast enhancement, and denoising (ACCE-D) framework. In the proposed framework, the Difference of Gaussian (DoG) filter and the bilateral filter are employed to decompose the high-frequency and low-frequency components, respectively. In 2023, Zhang et al. [51] proposed a Piecewise Color Correction and Dual Prior Optimized Contrast Enhancement, called PCDE. They first considered the color cast of the underwater image. Afterward, they used different enhancement strategies to enhance the base and detail layers of the V channel in the HSV space decomposed using the spatial prior and texture prior.
Underwater image methods based on non-physical models aim to address various issues related to image quality degradation through a step-by-step approach. The enhancement effects of image quality are sequentially stacked in these methods, which have the advantages of high flexibility and the ability to be stacked. However, they often rely too heavily on pixel-level processing and neglect the principles of underwater optical imaging. This can result in enhanced images that do not accurately reflect the laws of underwater imaging.

2.3. Deep Learning-Based Methods

Deep learning technology has advanced significantly in recent decades and has been widely applied in computer vision and image-processing domains. Notably, the field of underwater image enhancement has made significant progress through the adoption of deep learning techniques [15,52,53]. As a result, many researchers have used deep learning approaches to improve the quality of underwater images [54].
In the early stages of research, convolutional neural networks (CNNs) emerged as a pioneering deep learning model and became one of the leading methods in the field [55]. When integrating CNNs with physical models for underwater image enhancement, the CNN model takes an underwater image as input and produces transmission maps and background light in the output layer. On the other hand, combining CNNs with non-physical models requires paired underwater images and their corresponding ground truth images. Importantly, the CNN model does not rely on prior knowledge, ensuring that the output images are unaffected by subjective factors.
Cai et al. [56] proposed a dehazing model based on the conventional convolutional neural network (CNN) architecture, which employed bilateral rectified linear units as activation functions. In 2018, Hou et al. [57] introduced residual learning into underwater image enhancement to enhance learning effectiveness. In 2020, Li et al. [58] constructed a benchmark dataset for underwater image enhancement using a large-scale collection of real underwater images. They further proposed an underwater image enhancement CNN trained on this dataset.
In recent years, there has been a growing trend toward the integration of generative adversarial networks (GANs) into machine learning methods [53,59,60,61,62]. Combining GANs with a physical model can effectively guide the synthesis of underwater images and facilitate the subsequent authenticity assessment through the discrimination process. This iterative procedure continues until the restored underwater image is obtained. Conversely, if GANs are combined with a non-physical model, the resulting underwater images are not limited by the constraints imposed by a physical model.
Fabbri et al. [63] proposed a method to enhance the visual quality of underwater images based on GANs, called the underwater generative adversarial network (UGAN). Wen Peizhi et al. [64] also developed a method for underwater image enhancement using GANs and improved convolutional neural networks (CNNs) to enhance the color correction capability of the models.
Current research faces a challenge in providing a substantial number of paired images for the effective training of deep learning-based methods for underwater image enhancement.
A variety of techniques have been investigated to improve the quality of underwater images. However, each approach has inherent limitations. The challenge for current researchers is to develop enhanced image-processing methods that require less computational power.

3. Proposed Algorithm

Figure 2 shows our proposed image enhancement method, which operates independently of the optical model of underwater images. It comprises two main stages: luminance reconstruction and color correction. In the luminance reconstruction stage, the initial image is decomposed into three separate single-channel images (red, green, and blue) using a multi-resolution fusion method to retain intricate details within the scene. Subsequently, the white balance color correction is improved by using a saliency map to compensate for color attenuation in underwater environments, resulting in overall image enhancement.

3.1. Decomposition of Image Color Space

In the dehazing methods for atmospheric images, researchers have incorporated near-infrared images to compensate for the attenuation caused by clouds and haze in the visible light spectrum. The near-infrared images are capable of penetrating clouds and haze, thereby enhancing the visibility of the atmospheric image [65]. This fusion technique has served as an inspiration for enhancing underwater images. However, it is important to note that not all images have paired visible light and near-infrared counterparts, as is the case with atmospheric images. Therefore, this study aims to explore the possibility of extracting relevant information directly from the original image that can produce a similar effect to fusing visible light and near-infrared images.
Different bands of visible light experience varying levels of attenuation underwater. The green band of visible light is relatively well preserved and retains a higher amount of image information. Underwater images are initially decomposed into the RGB color space, treating the R, G, and B channels as distinct channels representing different wavelengths of visible light. Figure 4 shows the original image of an underwater cave with noticeable red attenuation. The red channel provides more reliable information in regions of higher luminance, such as the upper part of the cave entrance. However, it fails to provide effective information in areas near the rocks due to the lower luminance levels. However, the original image loses some information near the cave entrance in the green and blue channels due to excessive luminance. In contrast, more details can be retained near the rocks.
This work aims to improve the quality of underwater images while preserving relevant information in each channel. The focus is on retaining more details, particularly related to luminance. To achieve this, we propose using the HSV color space, which effectively separates luminance from color and offers better color saturation preservation than other color spaces.
Expanding on the previous breakdown of color using the RGB color space, our next step involves rebuilding the brightness information within the HSV color space using the decomposed three-channel data obtained from the RGB color space. Subsequently, we apply color correction techniques to the rebuilt image to achieve the desired enhancement effect.

3.2. Multi-Resolution Fusion-Based Luminance Reconstruction

This section presents a new method for luminance reconstruction. First, we extract individual RGB channel images. Then, we use local entropy, local contrast, and local visibility measurements to create a weight map. This weight map guides the subsequent luminance reconstruction process, which involves the multi-resolution fusion of the single-channel images.
In image fusion, the main objective is to selectively retain the clearer and more detailed portions from each channel image. For instance, in Figure 4, we aim to preserve the red channel image information near the cave entrance and prioritize the blue and green channel information near the rocks. Suppose a simple weighted average fusion technique is used for all three channels. In that case, it may retain image information indiscriminately, regardless of whether there are significant details present in that particular area. As a result, this approach could blur crucial information and dilute prominent details.
To overcome these limitations, this research paper proposes using local entropy, local contrast, and local visibility measures to create a weight map. The weight map must meet the following criteria: weight values should range from 0 to 1, and the sum of the weight values across the three channels must be 1 for each pixel. This weight map enables a more precise selection of information during the fusion process, improving the overall quality and preserving the salient features in the final fused image.
The entropy of an image is a statistical metric that quantifies the information content of the image. It measures the average amount of information conveyed by the image. The entropy of a two-dimensional image is closely related to the amount of information contained in the distribution of gray levels within the image. This measure offers valuable information about the image’s spatial characteristics and clustering properties.
The proposal is to include a weighted measure, denoted as H , which quantifies the normalized local entropy. This measure is computed within the vicinity of a specific pixel ( x , y ) using the neighboring region N .
H ( x , y ) = ( i , j ) N p f , i , j l o g 2 p ( f , ( i , j ) )
The probability of observing the f th gray level is denoted by the symbol p f . In this implementation, a predetermined neighborhood with a size of 7 × 7 is employed. Therefore, to compute p f , the number of pixels with the same value as the pixel located at position i , j within the 7 × 7 neighborhood centered around it, including the pixel itself, is tallied. The desired probability is obtained by dividing the count by the total number of pixels within the neighborhood.
To guarantee the accuracy of H in practical calculations, it is essential to apply the subsequent transformations to the H formula:
H x , y = i , j N p f , i , j l o g 2   n u m f , i , j + ( i , j ) N p f , i , j l o g 2 n  
The equation defines n u m ( f , ( i , j ) as the count of pixels in the 7 × 7 neighborhood centered at pixel i , j that share the same value as the pixel i , j , including the pixel i , j itself. Here, n represents the total number of pixels in the local neighborhood N . This approach ensures computational accuracy is not lost due to excessively small values of p f , i , j .
Local contrast is the difference in intensity between the target and the background within a given image or region of interest. Its main purpose is to enhance fine details while reducing the impact of the background. The local contrast coefficient ( C ) is calculated within a local neighborhood ( N ) centered around pixel coordinates x , y . It is defined mathematically as follows:
      C ( x , y ) = 1 n ( i , j ) N ( I i , j μ x , y ) 2
μ x , y is defined as follows:
μ x , y = ( i , j ) N I i , j
where I i , j represents the pixel value of the input image. Higher weights and contrasts are indicative of more intricate details within the region if the neighboring pixels exhibit greater deviations from the local mean. This paper uses a fixed neighborhood with dimensions of 7 × 7.
Local visibility is evaluated by assessing the channel’s ability to capture specific details within a given medium, such as water. A fuzzy estimation approach is employed using a Gaussian function with a predetermined standard deviation to induce blurring in the image, denoted as I b l u r . The square root of the difference between the original image and its blurred counterpart is then used as the quantitative indicator of visibility, referred to as V .
I b l u r ( x , y ) = I ( x , y ) g M ( x , y ,   σ 1 )  
V ( x , y ) = ( I x , y I b l u r ( x , y ) ) 2 g M ( x , y ,   σ 2 )
The function   g M ( x , y ,   σ )   represents a two-dimensional circular symmetric Gaussian weighted function with an M × M window and a standard deviation of σ 2 . The operator denotes a two-dimensional convolution. In this implementation, M is set to 7 and both σ 1 and σ 2 are set to 2.
The local entropy, local contrast, and local visibility weights are normalized and then combined through multiplication to derive a resultant weight. This study postulates that each of the three weights should have an equal influence on the final fusion outcome. Therefore, the weight assigned to each pixel can be represented as follows:
W i ( x , y ) = H i ( x , y ) C i ( x , y ) V i ( x , y )
where i R , G , B , denoting the three channels R, G, and B. Subsequently, we normalize the weights to ensure their summation is equal to 1 at each pixel location.
Figure 5 shows the resulting weight map generated for each channel of the underwater image shown in Figure 4. In this plot, a lighter weight map indicates a higher weight assigned to the corresponding point. As can be seen, the red channel, which is better preserved near the entrance to the cave, has a higher weight. Conversely, the blue and green channels have relatively higher weights near the rocks.
After obtaining the normalized weight maps, a multi-resolution fusion approach is applied to perform image fusion. In this section, we propose a technique that utilizes Laplacian pyramid decomposition for three single-channel images, Gaussian pyramid decomposition for the weight maps, and subsequently conducts fusion using Laplacian pyramid.
The l t h level Gaussian pyramid decomposition of the weight map can be represented as follows:
G { W ( x , y ) } l = [ g ( x , y ) G { W ( x , y ) } l 1 ] 2
where g ( x , y ) represents the Gaussian kernel, G { W ( x , y ) } l   denotes the approximation from the previous level, and ↓2 signifies the downsampling operation that reduces the area of the newly obtained image to one-fourth of the lower level. The Gaussian function exhibits separability, enabling us to simplify the algorithm implementation through the utilization of a 7 × 7 Gaussian filter.
The decomposition of the l-th level of the image using the Laplacian pyramid can be represented as follows:
L { I ( x , y ) } l = G { I ( x , y ) } l 1 R x , y [ G { I ( x , y ) } l ] 2
where R x , y signifies the interpolation filter, while the notation 2 denotes the upsampling operation by 2.
To construct the Laplacian pyramid for the given set of three input images, multiply each image’s corresponding level with the respective weight map and then sum them together. The reconstruction process follows the established Laplacian pyramid method, starting from the highest level and progressively adding it to the subsequent lower level until reaching the lowest level of the pyramid. The process can be defined as follows:
L I F x , y l = G W R x , y l L I R x , y l + G W G x , y l L I G x , y l + G { W B ( x , y ) } l L { I B ( x , y ) } l
Finally, the reconstructed image I F is a matrix of identical dimensions to the source image. The pixel values of the I F image are then linearly scaled from the range of [ 0 , m a x ( I F ) ] to the normalized range of [ 0 ,   1 ] . This normalization process effectively represents the reconstructed luminance component:
V n e w x , y = g a m m a I F x , y , 1
w h e r e   g a m m a I F x , y , 1 is a correction function used for gamma adjustment. It applies a correction factor of 1.

3.3. Color Correction

After integrating luminance reconstruction, we apply color correction to the fused image using a method that combines the RGB and HSV color spaces. We begin with an enhanced white balance and saliency detection technique for preliminary color correction and image sharpening. Then, we rectify the hazy appearance of the image within the HSV color space, leveraging the specific properties of underwater images.
Following the luminance reconstruction, the image is restored from the HSV space to the RGB space. This process alters the overall color of the image due to the change in luminance. To correct the color, preliminary color correction is performed for each component (R, G, and B) of the restored image as outlined below:
I T , R G B ( x , y ) = I F , R G B ( x , y ) V n e w x , y α I R x , y + I G x , y + I B x , y 3
where I T , R G B ( x , y )   represents the preliminary color-corrected image, I F , R G B x , y represents the fused RGB image, and α is the control parameter set to 1.5 as specified in this paper.
Next, we will perform color correction on underwater images. Our analysis of numerous underwater images has shown that the green and blue components remain intact, while the red component undergoes significant attenuation. Therefore, we have developed an enhanced white balance methodology based on the established criteria:
Compensating the red channel based on the blue and green channels. Compensation levels are determined by the pixel value of the red channel and the information conveyed by the blue and green channels. Lower compensation levels are required for pixels with higher values in the red channel or those with limited information in the blue and green channels. Conversely, Higher compensation levels are required for pixels with lower values in the red channel or those with abundant information in the blue and green channels.
In underwater images, when the blue component is excessively attenuated, compensation for the reduced blue channel is necessary by utilizing the green channel. This compensation method follows a similar principle as the approach used for the red channel. However, in this scenario, it is essential to artificially attenuate the green channel to achieve optimal results. Otherwise, if the blue component is high, artificial attenuation is applied to the blue channel.
Following the prescribed criteria, we will first establish the compensation formula for the red channel. It is important to note that before starting the calculation process, the pixel values of the RGB image must be normalized to fall within the range of [0, 1].
I C , R = I T , R + 1 2 arg I T , G arg I T , R arg I T , R + arg I T , G + arg I T , B 1 I T , R I T , G + 1 2 arg I T , B arg I T , R arg I T , R + arg I T , G + arg I T , B 1 I T , R I T , B
where I C , R represents the corrected red channel. arg I represents the average pixel intensity of the image.
When arg I T , G > arg I T , B , it can be inferred that the attenuation of the blue component is higher. The compensation formula for the blue channel and the artificial attenuation formula for the green channel are provided below:
I C , B = I T , B + arg I T , G arg I T , B arg I T , R + arg I T , G + arg I T , B 1 I T , B I T , G
I C , G = I T , G arg I T , G arg I T , B arg I T , R + arg I T , G + arg I T , B I T , G
The formula for artificial attenuation of the blue component is as follows:
I C , B = I T , B arg I T , B arg I T , G arg I T , R + arg I T , G + arg I T , B I T , B
After compensating and attenuating pixels using artificial techniques, we integrated the saliency detection method proposed by Hou et al. [66] into our study. This method allows us to identify the most important areas within the image scene by analyzing the spectral residue of the image in the spectral domain. Our objective is to enhance the visually sensitive regions of underwater images by integrating the saliency detection method. The improved white balance image is represented based on the saliency map as follows:
I w b = I C + s a l i e n c y M a p I C
where s a l i e n c y M a p refers to the saliency map. In our study, we only preserve the regions that exceed the automatically determined threshold.
We then use a non-masked sharpening technique to enhance the white-balanced image:
I s h a r p = I w b + 1 4 N o r I w b g I w b
where N o r I denotes the linear normalization of an image, while g represents a Gaussian filter.
Underwater images often appear hazy due to similar color values and reduced saturation in certain areas. To address this issue, we use histogram equalization to improve image sharpness and enhance local contrast. This technique results in a clearer and more visually appealing image. However, it is worth noting that histogram equalization can cause significant changes in color for certain images and may reduce the level of detail in specific regions. Therefore, we chose to convert the histogram-equalized and sharpened images to the HSV color space, integrate them with the initial reconstructed luminance, and then apply gamma correction to achieve the desired enhancement in the final image:
H e h = H s h a r p
S e h = g a m m a λ 1 S s h a r p + 1 λ 1 S h i s t e q   , α
V e h = g a m m a λ 2 V s h a r p + λ 3 V h i s t e q + 1 λ 2 λ 3 I F   ,     β
In the HSV color space, H s h a r p , S s h a r p , and V s h a r p correspond to the three elements of the sharpened image. S h i s t e q and V h i s t e q represent the saturation and luminance components of the sharpened image following the histogram equalization. The correction coefficients are denoted by α and β , while λ 1 , λ 2 , and λ 3 signify the weighting factors. The values of λ 1 , λ 2 , and λ 3 must be within the range [0, 1], and the sum of λ 2 , and λ 3 should also be within this range. Finally, the image is converted to the RGB color space, resulting in I e h .
Finally, the image is processed based on the gray world principle [67]. To improve performance under underwater conditions, we employ the following adapted gray world method:
f = 1.2 m i n ( m a x ( I e h , R ¯ , I e h , G ¯ , I e h , B ¯ ) , 0.375 )
I o u t , S = f I e h , S I e h , S ¯
where S [ R , G , B ] , I e h , R ¯ , I e h , G ¯ , I e h , B ¯ represent the average values of the red, green, and blue channels, respectively, in the I e h image, and I o u t refers to the resultant final image.

4. Results and Discussion

In this section, we validated the methodologies presented in Section 3 and conducted a comparative analysis between our method and the existing methods for underwater enhancement. The experimental results in this paper are based on the underwater dataset UIEB proposed by Li et al. [58], and UCCS and UIQS proposed by Liu et al. [68]. UIEB includes 950 authentic underwater images, of which 890 have corresponding reference images, named standard dataset. Li et al. [58]’s method fails to produce satisfactory outcomes for the remaining 60 challenging dataset images. UCCS includes 300 images; this set aims to evaluate the ability of correcting color casts for algorithms. UIQS includes 3630 images; this subset is used to test algorithms for the improvement of image visibility.
To evaluate the effectiveness of the methods, this research paper will conduct a thorough analysis and comparison of the generated images under different environmental conditions using both objective evaluations and marked subjective evaluations. Firstly, we will assess the performance of various methods on the standard and challenging datasets separately while analyzing the outcomes of each method objectively.

4.1. Luminance Reconstruction Evaluation

Luminance reconstruction aims to preserve the most important information within each channel. This section conducts a qualitative analysis of the reconstructed luminance, focusing on its perceptual quality and fidelity. Figure 6 shows some examples of the luminance reconstruction results for part of the images from the image set.
As shown in Figure 6a, the original image displays excessive luminance in the zoomed-in area, leading to a loss of fine details near the coral. By reconstructing the luminance, the texture of the uneven ground surface becomes more prominent. In Figure 6b, several white objects underwater are affected by overexposure in the original luminance map, making it different to distinguish their specific contours. Luminance reconstruction highlights the corroded and rusted texture on the surface of the two cylindrical objects on the left, as well as the intersection lines between the faces of the polyhedron on the right.
Regarding Figure 6c, the original image has excessively high green channel values, resulting in an overall brighter luminance map. The reconstructed luminance enhances the contrast between the ground and the porcelain fragments, resulting in sharper edges of the fragments. In Figure 6d, the reconstructed luminance reveals the texture on the diver’s fins clearly, which was barely discernible in the previous image compared to the original luminance.
In Figure 6e, the background’s luminance near the original image’s water surface decreases significantly with depth, making it difficult to observe the bubbles exhaled by the diver. However, the reconstructed luminance map maintains a relatively consistent luminance across different underwater depths, making the bubbles stand out against the background.
The experimental results for Figure 6f are consistent with the hypothesis in section B of the third chapter: the reconstructed luminance map preserves more blue and green channel information near the rocks compared to the original luminance map while adding more red channel information near the cave entrance. As a result, the overall visibility of the image is improved.

4.2. Qualitative Evaluation

This section presents a comparative analysis between the method proposed in this paper and several established classical methods for underwater image enhancement. The parameters used for our method are specified as follows: λ 1 = 0.6 ,   λ 2 = 0.2 ,   a n d   λ 3 = 0.5 ;   α = 0.85   a n d   β = 1.5 .
The contrast methods under consideration for comparison consist of He et al.’s dark channel prior method (GDCP) [33], Peng et al.’s GDCP method [37], Fu et al.’s two-step method [45], Ancuti et al.’s fusion-based method [49], Hou et al.’s UVT method [38], Xie et al.’s UNVT method [39], and Zhang et al.’s PCDE method [51].
Figure 7 shows some results of the different methods on underwater images.
Figure 7 shows that DCP [33], GDCP [37], and UTV [38] methods are not particularly effective at handling blue and green colors in the background of images. The results produced by GDCP are much better than DCP’s, but GDCP [37] may lead to overexposure in areas of the original image with high luminance levels. The two-step [45] method is incapable of processing the blue background present in the image, resulting in unstable performance. The UNTV [39] and PCDE [51] methods are superior to the methods mentioned above, nevertheless, the objects in the processed image still exhibit hues characteristic of blue or green. The images obtained by the PCDE method conceal a considerable amount of detail, both in bright and dark areas. Ancuti et al. [49]’s fusion-based method and the method proposed in this paper perform exemplary jobs of correcting the image’s color. However, the fusion-based method is conducive to a certain degree of haze-like appearance. The method proposed in this paper facilitates the restoration of image color in the presence of turbid water and non-uniform lighting conditions, while also enhancing image details. This is in contrast to several methods that have been mentioned above. Moreover, as shown in the first and ninth rows of Figure 7a, it can enhance foreground images while maintaining the background.
Figure 8 illustrates the processing outcomes achieved by the methods on underwater images acquired from different perspectives of the identical scene. Consistent color representation should ideally be observed upon processing the same object. However, as depicted in Figure 8, the results using DCP [33], GDCP [37], two-step [45], UTV [38], UNTV [39], and PCDE [51] do not exhibit uniform colors for the restored sculptures. Conversely, the fusion-based [49] method and the method presented in this paper effectively restore objects in underwater images while maintaining similar color schemes. Additionally, the background colors produced by our method also demonstrate a remarkable level of cohesion.

4.3. Quantitative Evaluation

The method’s image quality was comprehensively evaluated using three metrics: UIQM [69] (Underwater Image Quality Measures), UCIQE [70] (Underwater Color Image Quality Evaluation Metric), FDUM [71] (Frequency Domain-based UIQA Method), CSN [72] (Contrast Sharpness and Naturalness), and image entropy.
Panetta et al. [69] proposed UIQM, a non-reference measure designed to assess the quality of underwater images. UIQM includes three attribute measurements inspired by the characteristics of the human visual system: color, sharpness, and contrast. the UIQM of a given image is calculated as follows:
U I Q M = c 1 × U I C M + c 2 × U I S M + c 3 × U I C o n M ,
where U I C M denotes the colorfulness, U I S M denotes sharpness, and U I C o n M denotes the contrast; c 1 , c 2 , and c 3   are three weighted coefficients.
UCIQE, on the other hand, evaluates the quality of underwater color images by considering a weighted combination of chroma, saturation, and sharpness in the CIELab color space. The resulting UCIQE index ranges from 0 to 1, with higher values indicating better balance performance among chroma, saturation, and sharpness. A higher UCIQE score corresponds to a higher-quality underwater image. The definition of UCIQE is given by the following:
U C I Q E = c 1 × σ C + c 2 × c o n L + c 3 × μ S
where σ C , c o n L , and μ S indicate the standard deviation of chroma, the contrast of luminance, and the average of saturation, respectively; c 1 , c 2 , and c 3   are three weighted coefficients.
FDUM is obtained by fusing the colorfulness, contrast, and sharpness measures. The formula is as follows:
F D U M = ω 1 × C o l o r f u l n e s s + ω 2 × C o n t r a s t + ω 3 × S h a r p n e s s
where ω 1 , ω 2 and ω 3 are the weighted coefficients corresponding to the colorfulness, contrast, and sharpness measurements, respectively.
CSN is a no-reference underwater IQA method which is designed by combining the contrast index, the sharpness index, and the naturalness statistical model.
Additionally, image entropy indicates the information content in an image. The image entropy metric measures the randomness or uncertainty present in the image data. Higher levels of entropy indicate a greater amount of information, while lower levels suggest a more predictable and less complex image.
Therefore, the UIQM, UCIQE, PDUM, and CSN evaluation metrics, along with image entropy, provide a comprehensive and professional assessment of the image quality of the method, taking into account various aspects such as color, sharpness, contrast, and information content.
Table 1 presents the evaluation metrics, namely UIQM, UCIQE, PDUM, CSN, and image entropy, for the methods applied to the UIEB standard set, UIEB challenging set, UCCS, and UIQS. These metrics reflect the average evaluation values across all the images within each dataset.
Our method achieves the top 2 scores for both UIQM and entropy on the four datasets, and both UIQM and entropy scored three top 1. In terms of the CSN metrics, our method achieved the highest score in two sets, came second in one, and was placed third in another.
It is interesting to note that the method introduced by He et al. [33] achieves the highest average UIQM score on the UIEB challenging set, which is intriguing. We conducted a thorough analysis of this matter and found that a majority of the 60 images included in the challenging dataset exhibit a hazy appearance. It should be noted that He et al.’s method [33] performs exceptionally well in haze removal, which is supported by both our analysis and the subjective evaluation of He et al.’s [33] defogging capabilities.
The UCIQE metric shows that Peng et al.’s GDCP method [37] and He et al.’s method [33] both achieved the top 2 UCIQE values on the four sets. However, our method’s performance ranked only in the middle, which contradicts our previous analysis. To further investigate this discrepancy, we conducted a series of experiments and made an interesting observation that the UCIQE method exhibits a bias toward image saturation. It was observed that increasing the saturation of certain images with pronounced color deviation, such as multiple images in the second and fifth rows of Figure 7a, resulted in a dominant green or blue appearance in the entire image. This led to a significant decrease in visibility, but an increase in the UCIQE values. Some of the evaluated methods with UCIQE values greater than 0.5 increased image saturation without implementing the proper color correction techniques. As a result, these images experienced color deviation and degraded visual effects, but still achieved higher UCIQE values.
To ensure a fair comparison of the UCIQE values, the methods lacking proper color correction should be excluded. After excluding such methods, our proposed method demonstrated an excellent UCIQE performance in this study.
It seems that our method did not achieve the top two positions in PDUM in Table 1. However, our method is currently ranked third, with a relatively narrow margin between it and the top two. Same as UCIQE, our method demonstrates favorable PDUM performance, with the exception of those algorithms that are unable to achieve superior color correction.
In addition, it can be seen that, except for the UIEB challenge set, which is an artificially selected image with particularly harsh underwater environments, the UIQM, UCIQE, and entropy results of this paper’s method on the UIEB standard set, the UCCS, and the UIQS, are virtually unchanged, which proves the stability of the method in this paper.
In light of the aforementioned methods, our method demonstrates superior efficacy in background color correction when compared to DCP [33], GDCP [37], and UTV [38]. Our method is efficacious in the removal of blue and green hues from the background. Compared to two-step [45], fusion-based [49], UNTV [39], and PCDE [51] methods, the method proposed in this paper is a more consistent means of removing the blue and green color from both the background and the objects therein. Furthermore, it maintains the vividness of the image’s color while retaining more detailed information and achieving favorable results in quantitative analysis.

5. Conclusions

This paper proposes a new method for enhancing underwater images using multi-resolution fusion for luminance reconstruction. The method aims to extract latent information from individual images by exploiting the RGB channels of the original image and subsequently performing color restoration. Extensive experimentation has shown that the method greatly improves image sharpness by emphasizing the importance of details within each channel. Furthermore, our color correction method effectively addresses color deviations caused by uneven underwater lighting conditions. As a result, our proposed method has significant potential for practical applications in engineering fields related to underwater image enhancement.
Our method also has several limitations. The multi-resolution fusion and Laplace–Gaussian pyramid approach may entail a considerable computational burden, rendering them unsuitable for real-time processing or deployment on lower-end devices. In future research, we will attempt to implement algorithmic lightweight. Furthermore, the technique of underwater image enhancement will be applied to other areas of research, including object recognition.

Author Contributions

Conceptualization, Z.C.; Methodology, Y.W.; Data curation, B.H.; Writing—original draft, Y.W.; Writing—review & editing, G.Y. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Marini, S.; Fanelli, E.; Sbragaglia, V.; Azzurro, E.; Del Rio Fernandez, J.; Aguzzi, J. Tracking Fish Abundance by Underwater Image Recognition. Sci. Rep. 2018, 8, 13748. [Google Scholar] [CrossRef] [PubMed]
  2. Lin, S.; Chi, K.-C.; Li, W.-T.; Tang, Y.-D. Underwater Optical Image Enhancement Based on Dominant Feature Image Fusion. Acta Photonica Sin. 2020, 49, 13. [Google Scholar] [CrossRef]
  3. Deluxni, N.; Sudhakaran, P.; Kitmo; Ndiaye, M.F. A Review on Image Enhancement and Restoration Techniques for Underwater Optical Imaging Applications. IEEE Access 2023, 11, 111715–111737. [Google Scholar] [CrossRef]
  4. Wang, N.; Zheng, H.; Zheng, B. Underwater Image Restoration via Maximum Attenuation Identification. IEEE Access 2017, 5, 18941–18952. [Google Scholar] [CrossRef]
  5. Wang, Y.; Liu, H.; Chau, L.P. Single Underwater Image Restoration Using Adaptive Attenuation-Curve Prior. IEEE Trans. Circuits Syst. I Regul. Pap. 2018, 65, 992–1002. [Google Scholar] [CrossRef]
  6. Ma, J.; Fan, X.; Wu, Z.; Zhang, X.; Shi, P.; Gengren, W. Underwater dam crack image enhancement algorithm based on improved dark channel prior. J. Image Graph. 2016, 21, 1574–1584. [Google Scholar] [CrossRef]
  7. Francescangeli, M.; Marini, S.; Martínez, E.; Del Río, J.; Toma, D.M.; Nogueras, M.; Aguzzi, J. Image dataset for benchmarking automated fish detection and classification algorithms. Sci. Data 2023, 10, 5. [Google Scholar] [CrossRef] [PubMed]
  8. Uemura, T.; Lu, H.; Kim, H. Marine Organisms Tracking and Recognizing Using YOLO; Springer: Cham, Switzerland, 2020; pp. 53–58. [Google Scholar]
  9. Rova, A.; Mori, G.; Dill, L. One Fish, Two Fish, Butterfish, Trumpeter: Recognizing Fish in Underwater Video. In Proceedings of the APR Conference on Machine Vision Applications, IAPR MVA 2007, Tokyo, Japan, 16–18 May 2007; pp. 404–407. [Google Scholar]
  10. Lebart, K.; Smith, C.; Trucco, E.; Lane, D.M. Automatic indexing of underwater survey video: Algorithm and benchmarking method. IEEE J. Ocean. Eng. 2003, 28, 673–686. [Google Scholar] [CrossRef]
  11. Kahanov, Y.A.; Royal, J.G. Analysis of hull remains of the Dor D Vessel, Tantura Lagoon, Israel. Int. J. Naut. Archaeol. 2001, 30, 257–265. [Google Scholar] [CrossRef]
  12. Peng, Y.T.; Cosman, P.C. Underwater Image Restoration Based on Image Blurriness and Light Absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef]
  13. Lu, H.; Li, Y.; Xu, X.; Li, J.; Liu, Z.; Li, X.; Yang, J.; Serikawa, S. Underwater image enhancement method using weighted guided trigonometric filtering and artificial light correction. J. Vis. Commun. Image Represent. 2016, 38, 504–516. [Google Scholar] [CrossRef]
  14. Chu, X.; Fu, Z.; Yu, S.; Tu, X.; Huang, Y.; Ding, X. Underwater Image Enhancement and Super-Resolution Using Implicit Neural Networks. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 1295–1299. [Google Scholar]
  15. Singh, N.; Bhat, A. A systematic review of the methodologies for the processing and enhancement of the underwater images. Multimed. Tools Appl. 2023, 82, 38371–38396. [Google Scholar] [CrossRef]
  16. Han, M.; Lyu, Z.; Qiu, T.; Xu, M. A Review on Intelligence Dehazing and Color Restoration for Underwater Images. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 1820–1832. [Google Scholar] [CrossRef]
  17. Chiang, J.Y.; Chen, Y.C. Underwater Image Enhancement by Wavelength Compensation and Dehazing. IEEE Trans. Image Process. 2012, 21, 1756–1769. [Google Scholar] [CrossRef] [PubMed]
  18. Zhou, J.; Zhang, D.; Zhang, W. A multifeature fusion method for the color distortion and low contrast of underwater images. Multimed. Tools Appl. 2021, 80, 17515–17541. [Google Scholar] [CrossRef]
  19. Deng, X.; Wang, H.; Liu, X. Underwater Image Enhancement Based on Removing Light Source Color and Dehazing. IEEE Access 2019, 7, 114297–114309. [Google Scholar] [CrossRef]
  20. Zhang, W.; Dong, L.; Pan, X.; Zhou, J.; Qin, L.; Xu, W. Single Image Defogging Based on Multi-Channel Convolutional MSRCR. IEEE Access 2019, 7, 72492–72504. [Google Scholar] [CrossRef]
  21. Zhou, J.; Hao, M.; Zhang, D.; Zou, P.; Zhang, W. Fusion PSPnet Image Segmentation Based Method for Multi-Focus Image Fusion. IEEE Photonics J. 2019, 11, 1–12. [Google Scholar] [CrossRef]
  22. Schechner, Y.Y.; Averbuch, Y. Regularized Image Recovery in Scattering Media. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1655–1660. [Google Scholar] [CrossRef]
  23. Drews, P.L.J.; Nascimento, E.R.; Botelho, S.S.C.; Campos, M.F.M. Underwater Depth Estimation and Image Restoration Based on Single Images. IEEE Comput. Graph. Appl. 2016, 36, 24–35. [Google Scholar] [CrossRef]
  24. Li, C.Y.; Guo, J.C.; Cong, R.M.; Pang, Y.W.; Wang, B. Underwater Image Enhancement by Dehazing With Minimum Information Loss and Histogram Distribution Prior. IEEE Trans. Image Process. 2016, 25, 5664–5677. [Google Scholar] [CrossRef] [PubMed]
  25. Carlevaris-Bianco, N.; Mohan, A.; Eustice, R.M. Initial results in underwater single image dehazing. In Proceedings of the OCEANS 2010 MTS/IEEE SEATTLE, Seattle, WA, USA, 20–23 September 2010; pp. 1–8. [Google Scholar]
  26. Schechner, Y.Y.; Karpel, N. Clear underwater vision. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; pp. 1–9. [Google Scholar]
  27. Uplavikar, P.; Wu, Z.; Wang, Z. All-In-One Underwater Image Enhancement using Domain-Adversarial Learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, CVPR 2019, Long Beach, CA, USA, 16–20 June 2019; pp. 1–8. [Google Scholar]
  28. Emberton, S.; Chittka, L.; Cavallaro, A. Underwater image and video dehazing with pure haze region segmentation. Comput. Vis. Image Underst. 2018, 168, 145–156. [Google Scholar] [CrossRef]
  29. Zhang, W.; Dong, L.; Pan, X.; Zou, P.; Qin, L.; Xu, W. A Survey of Restoration and Enhancement for Underwater Images. IEEE Access 2019, 7, 182259–182279. [Google Scholar] [CrossRef]
  30. McGlamery, B. A Computer Model for Underwater Camera Systems; SPIE: Bellingham, WA, USA, 1980; Volume 0208. [Google Scholar]
  31. Xie, K.; Pan, W.; Xu, S. An Underwater Image Enhancement Algorithm for Environment Recognition and Robot Navigation. Robotics 2018, 7, 14. [Google Scholar] [CrossRef]
  32. Trucco, E.; Olmos-Antillon, A.T. Self-Tuning Underwater Image Restoration. IEEE J. Ocean. Eng. 2006, 31, 511–519. [Google Scholar] [CrossRef]
  33. Kaiming, H.; Jian, S.; Xiaoou, T. Single image haze removal using dark channel prior. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1956–1963. [Google Scholar]
  34. Drews, P., Jr.; Nascimento, E.d.; Moraes, F.; Botelho, S.; Campos, M. Transmission Estimation in Underwater Single Images. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar]
  35. Galdran, A.; Pardo, D.; Picón, A.; Alvarez-Gila, A. Automatic Red-Channel underwater image restoration. J. Vis. Commun. Image Represent. 2015, 26, 132–145. [Google Scholar] [CrossRef]
  36. Tang, Z.; Zhou, B.; Dai, X.; Gu, H. Underwater Robot Visual Enhancements Based on the Improved DCP Algorithm. Jiqiren/Robot 2018, 40, 222–230. [Google Scholar] [CrossRef]
  37. Peng, Y.T.; Cao, K.; Cosman, P.C. Generalization of the Dark Channel Prior for Single Image Restoration. IEEE Trans. Image Process. 2018, 27, 2856–2868. [Google Scholar] [CrossRef]
  38. Hou, G.; Li, J.; Wang, G.; Yang, H.; Huang, B.; Pan, Z. A novel dark channel prior guided variational framework for underwater image restoration. J. Vis. Commun. Image Represent. 2020, 66, 102732. [Google Scholar] [CrossRef]
  39. Xie, J.; Hou, G.; Wang, G.; Pan, Z. A Variational Framework for Underwater Image Dehazing and Deblurring. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 3514–3526. [Google Scholar] [CrossRef]
  40. Li, Y.; Hou, G.; Zhuang, P.; Pan, Z. Dual High-Order Total Variation Model for Underwater Image Restoration. arXiv 2024, arXiv:2407.14868. [Google Scholar]
  41. Iqbal, K.; Odetayo, M.; James, A.; Rosalina Abdul, S.; Abdullah Zawawi Hj, T. Enhancing the low quality images using Unsupervised Colour Correction Method. In Proceedings of the 2010 IEEE International Conference on Systems, Man and Cybernetics, Istanbul, Turkey, 10–13 October 2010; pp. 1703–1709. [Google Scholar]
  42. Abdul Ghani, A.S.; Mat Isa, N.A. Underwater image quality enhancement through integrated color model with Rayleigh distribution. Appl. Soft Comput. 2015, 27, 219–230. [Google Scholar] [CrossRef]
  43. Abdul Ghani, A.S.; Mat Isa, N.A. Enhancement of low quality underwater image through integrated global and local contrast correction. Appl. Soft Comput. 2015, 37, 332–344. [Google Scholar] [CrossRef]
  44. Abdul Ghani, A.S. Image contrast enhancement using an integration of recursive-overlapped contrast limited adaptive histogram specification and dual-image wavelet fusion for the high visibility of deep underwater image. Ocean Eng. 2018, 162, 224–238. [Google Scholar] [CrossRef]
  45. Fu, X.; Fan, Z.; Ling, M.; Huang, Y.; Ding, X. Two-step approach for single underwater image enhancement. In Proceedings of the 2017 International Symposium on Intelligent Signal Processing and Communication Systems (ISPACS), Xiamen, China, 6–9 November 2017; pp. 789–794. [Google Scholar]
  46. Fu, X.; Zhuang, P.; Huang, Y.; Liao, Y.; Zhang, X.P.; Ding, X. A retinex-based enhancing approach for single underwater image. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 4572–4576. [Google Scholar]
  47. Zhang, S.; Wang, T.; Dong, J.; Yu, H. Underwater image enhancement via extended multi-scale Retinex. Neurocomputing 2017, 245, 1–9. [Google Scholar] [CrossRef]
  48. Tang, C.; von Lukas, U.F.; Vahl, M.; Wang, S.; Wang, Y.; Tan, M. Efficient underwater image and video enhancement based on Retinex. Signal Image Video Process. 2019, 13, 1011–1018. [Google Scholar] [CrossRef]
  49. Ancuti, C.O.; Ancuti, C.; Vleeschouwer, C.D.; Bekaert, P. Color Balance and Fusion for Underwater Image Enhancement. IEEE Trans. Image Process. 2018, 27, 379–393. [Google Scholar] [CrossRef]
  50. Li, X.; Hou, G.; Li, K.; Pan, Z. Enhancing underwater image via adaptive color and contrast enhancement, and denoising. Eng. Appl. Artif. Intell. 2022, 111, 104759. [Google Scholar] [CrossRef]
  51. Zhang, W.; Jin, S.; Zhuang, P.; Liang, Z.; Li, C. Underwater Image Enhancement via Piecewise Color Correction and Dual Prior Optimized Contrast Enhancement. IEEE Signal Process. Lett. 2023, 30, 229–233. [Google Scholar] [CrossRef]
  52. Li, C.; Anwar, S.; Hou, J.; Cong, R.; Guo, C.; Ren, W. Underwater Image Enhancement via Medium Transmission-Guided Multi-Color Space Embedding. IEEE Trans. Image Process. 2021, 30, 4985–5000. [Google Scholar] [CrossRef]
  53. Pham, T.T.; Mai, T.T.N.; Lee, C. Deep Unfolding Network with Physics-Based Priors for Underwater Image Enhancement. In Proceedings of the 2023 IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–11 October 2023; pp. 46–50. [Google Scholar]
  54. Anwar, S.; Li, C. Diving deeper into underwater image enhancement: A survey. Signal Process. Image Commun. 2020, 89, 115978. [Google Scholar] [CrossRef]
  55. Roska, T.; Chua, L.O. The CNN universal machine: An analogic array computer. IEEE Trans. Circuits Syst. II Analog Digit. Signal Process. 1993, 40, 163–173. [Google Scholar] [CrossRef]
  56. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. DehazeNet: An End-to-End System for Single Image Haze Removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef]
  57. Hou, M.; Liu, R.; Fan, X.; Luo, Z. Joint Residual Learning for Underwater Image Enhancement. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 4043–4047. [Google Scholar]
  58. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2020, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed]
  59. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  60. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised Generative Network to Enable Real-Time Color Correction of Monocular Underwater Images. IEEE Robot. Autom. Lett. 2018, 3, 387–394. [Google Scholar] [CrossRef]
  61. Espinosa, A.R.; McIntosh, D.; Albu, A.B. An Efficient Approach for Underwater Image Improvement: Deblurring, Dehazing, and Color Correction. In Proceedings of the 2023 IEEE/CVF Winter Conference on Applications of Computer Vision Workshops (WACVW), Waikoloa, HI, USA, 3–7 January 2023; pp. 206–215. [Google Scholar]
  62. Zhou, W.-H.; Zhu, D.-M.; Shi, M.; Li, Z.-X.; Duan, M.; Wang, Z.-Q.; Zhao, G.-L.; Zheng, C.-D. Deep images enhancement for turbid underwater images based on unsupervised learning. Comput. Electron. Agric. 2022, 202, 107372. [Google Scholar] [CrossRef]
  63. Fabbri, C.; Islam, M.J.; Sattar, J. Enhancing Underwater Imagery Using Generative Adversarial Networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 7159–7165. [Google Scholar]
  64. Wen, P.-Z.; Chen, J.-M.; Xiao, Y.-N.; Wen, Y.-Y.; Huang, W.-M. Underwater image enhancement algorithm based on GAN and multi-level wavelet CNN. J. Zhejiang Univ. (Eng. Sci.) 2022, 56, 213–224. [Google Scholar]
  65. Vanmali, A.V.; Gadre, V.M.J.S. Visible and NIR image fusion using weight-map-guided Laplacian–Gaussian pyramid for improving scene visibility. Sadhana 2017, 42, 1063–1082. [Google Scholar] [CrossRef]
  66. Hou, X.; Zhang, L. Saliency Detection: A Spectral Residual Approach. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  67. Ebner, M. Color Constancy; Wiley Publishing: Hoboken, NJ, USA, 2007. [Google Scholar]
  68. Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-World Underwater Enhancement: Challenges, Benchmarks, and Solutions Under Natural Light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  69. Panetta, K.; Gao, C.; Agaian, S. Human-Visual-System-Inspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  70. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  71. Yang, N.; Zhong, Q.; Li, K.; Cong, R.; Zhao, Y.; Kwong, S. A reference-free underwater image quality assessment metric in frequency domain. Signal Process. Image Commun. 2021, 94, 116218. [Google Scholar] [CrossRef]
  72. Hou, G.; Zhang, S.; Lu, T.; Li, Y.; Pan, Z.; Huang, B. No-reference quality assessment for underwater images. Comput. Electr. Eng. 2024, 118, 109293. [Google Scholar] [CrossRef]
Figure 1. Examples of underwater images.
Figure 1. Examples of underwater images.
Sensors 24 05776 g001
Figure 2. Workflow of our method.
Figure 2. Workflow of our method.
Sensors 24 05776 g002
Figure 3. Schematic of the underwater optical imaging model. Reprinted from Xie et al. [31].
Figure 3. Schematic of the underwater optical imaging model. Reprinted from Xie et al. [31].
Sensors 24 05776 g003
Figure 4. RGB spatial decomposition of the underwater image. From left to right: the original underwater image, the red channel image, the green channel image, and the blue channel image.
Figure 4. RGB spatial decomposition of the underwater image. From left to right: the original underwater image, the red channel image, the green channel image, and the blue channel image.
Sensors 24 05776 g004
Figure 5. RGB channels of the underwater images and their weight maps. First row from left to right: image of the red channel, image of the green channel, and image of the blue channel; second row from left to right: weight maps generated for the red channel, green channel, and blue channel.
Figure 5. RGB channels of the underwater images and their weight maps. First row from left to right: image of the red channel, image of the green channel, and image of the blue channel; second row from left to right: weight maps generated for the red channel, green channel, and blue channel.
Sensors 24 05776 g005
Figure 6. Example of luminance reconstruction results for part of the images from the image set. From left to right are six underwater images, from top to bottom: the original underwater images, the RGB components of the original images, the original luminance components of the images, the reconstructed luminance components, partially enlarged details of the original luminance components, and the corresponding enlarged details of the reconstructed luminance components. (af) are seven different underwater images.
Figure 6. Example of luminance reconstruction results for part of the images from the image set. From left to right are six underwater images, from top to bottom: the original underwater images, the RGB components of the original images, the original luminance components of the images, the reconstructed luminance components, partially enlarged details of the original luminance components, and the corresponding enlarged details of the reconstructed luminance components. (af) are seven different underwater images.
Sensors 24 05776 g006
Figure 7. Examples of the results of different methods on underwater images. (a) Illustrative results of algorithmic analysis on the UIEB standard dataset; (b) illustrative results of algorithmic analysis on the UIEB challenging dataset; (c) illustrative results of algorithmic analysis on UCCS and UIQS dataset. From left to right: the original underwater image, results using DCP [33], GDCP [37], two-step [45], fusion-based [49], UTV [38], UNTV [39], PCDE [51], and our method.
Figure 7. Examples of the results of different methods on underwater images. (a) Illustrative results of algorithmic analysis on the UIEB standard dataset; (b) illustrative results of algorithmic analysis on the UIEB challenging dataset; (c) illustrative results of algorithmic analysis on UCCS and UIQS dataset. From left to right: the original underwater image, results using DCP [33], GDCP [37], two-step [45], fusion-based [49], UTV [38], UNTV [39], PCDE [51], and our method.
Sensors 24 05776 g007aSensors 24 05776 g007b
Figure 8. Results of different methods on underwater images from different angles in the same scene. From left to right: the original underwater image, results using DCP [33], GDCP [37], two-step [45], fusion-based [49], UTV [38], UNTV [39], PCDE [51], and our method.
Figure 8. Results of different methods on underwater images from different angles in the same scene. From left to right: the original underwater image, results using DCP [33], GDCP [37], two-step [45], fusion-based [49], UTV [38], UNTV [39], PCDE [51], and our method.
Sensors 24 05776 g008
Table 1. Underwater image enhancement evaluation based on UIQM, UCIQE, FDUM, CSN, and entropy metrics on UIEB standard set, UIEB challenging set, UCCS, and UIQS. The larger the metric the better.
Table 1. Underwater image enhancement evaluation based on UIQM, UCIQE, FDUM, CSN, and entropy metrics on UIEB standard set, UIEB challenging set, UCCS, and UIQS. The larger the metric the better.
Methods UIEB Standard Set UIEB Challenging Set
UIQMUCIQEFDUMCSNEntropyUIQMUCIQEFDUMCSNEntropy
DCP [33]2.9550.5790.6640.5817.2673.9430.5750.5060.5576.780
GDCP [37]2.6520.6040.8660.5997.1951.4790.5690.6610.5867.383
Two-step [45]3.7790.4880.6190.5517.4572.1340.4760.3970.4957.215
Fusion-based [49]4.2190.4500.5200.5487.4192.7240.4270.3270.4847.148
UTV [38]2.3070.5740.6080.5746.1121.1050.5260.3570.4945.313
UNTV [39]3.4120.5360.8820.5880.7442.4210.5110.5700.5707.048
PCDE [51]4.9360.5060.4470.5107.7192.5200.4860.3390.6056.893
Our method4.9620.4970.7720.5917.6553.3300.4770.4400.5507.393
Methods UCCS UIQS
UIQMUCIQEFDUMCSNEntropyUIQMUCIQEFDUMCSNEntropy
DCP [33]1.6560.5790.4530.4327.3201.9800.5770.4890.5227.290
GDCP [37]4.7430.5630.4570.4747.5394.8240.5650.5010.4697.565
Two-step [45]3.6870.4370.4570.4507.2303.8080.4390.4470.4447.245
Fusion-based [49]3.7230.4050.3130.4227.1843.8600.4150.3330.5157.230
UTV [38]−0.3760.5140.2480.3715.294−0.0520.5040.2460.3645.580
UNTV [39]3.8880.5050.7480.4947.6393.9700.5130.7520.4847.580
PCDE [51]3.9250.5050.7010.5027.5723.9410.5040.7070.5037.527
Our method4.9750.4910.6740.5357.7644.8410.4930.7030.5377.719
The most optimal result is red while the second best is blue. Values in the table are dimensionless.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Chen, Z.; Yan, G.; Zhang, J.; Hu, B. Underwater Image Enhancement Based on Luminance Reconstruction by Multi-Resolution Fusion of RGB Channels. Sensors 2024, 24, 5776. https://doi.org/10.3390/s24175776

AMA Style

Wang Y, Chen Z, Yan G, Zhang J, Hu B. Underwater Image Enhancement Based on Luminance Reconstruction by Multi-Resolution Fusion of RGB Channels. Sensors. 2024; 24(17):5776. https://doi.org/10.3390/s24175776

Chicago/Turabian Style

Wang, Yi, Zhihua Chen, Guoxu Yan, Jiarui Zhang, and Bo Hu. 2024. "Underwater Image Enhancement Based on Luminance Reconstruction by Multi-Resolution Fusion of RGB Channels" Sensors 24, no. 17: 5776. https://doi.org/10.3390/s24175776

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop