Next Article in Journal
Numerical Study and Optimization of Speed Control Unit for Submarine Natural Gas Pipeline Pig
Previous Article in Journal
A Novel Open Set Adaptation Network for Marine Machinery Fault Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Enhancement Fusion Method Guided by Salient Region Detection

1
School of Smart Marine Science and Engineering, Fujian University of Technology, Fuzhou 350118, China
2
Fujian Provincial Key Laboratory of Marine Smart Equipment, Fuzhou 350118, China
3
State Key Laboratory of Advanced Design and Manufacturing Technology for Vehicle, Hunan University, Changsha 410008, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(8), 1383; https://doi.org/10.3390/jmse12081383
Submission received: 17 July 2024 / Revised: 4 August 2024 / Accepted: 12 August 2024 / Published: 13 August 2024
(This article belongs to the Section Ocean Engineering)

Abstract

:
Exploring and monitoring underwater environments pose unique challenges due to water’s complex optical properties, which significantly impact image quality. Challenges like light absorption and scattering result in color distortion and decreased visibility. Traditional underwater image acquisition methods face these obstacles, highlighting the need for advanced techniques to solve the image color shift and image detail loss caused by the underwater environment in the image enhancement process. This study proposes a salient region-guided underwater image enhancement fusion method to alleviate these problems. First, this study proposes an advanced dark channel prior method to reduce haze effects in underwater images, significantly improving visibility and detail. Subsequently, a comprehensive RGB color correction restores the underwater scene’s natural appearance. The innovation of our method is that it fuses through a combination of Laplacian and Gaussian pyramids, guided by salient region coefficients, thus preserving and accentuating the visually significant elements of the underwater environment. Comprehensive subjective and objective evaluations demonstrate our method’s superior performance in enhancing contrast, color depth, and overall visual quality compared to existing methods.

1. Introduction

The harsh and complex underwater environment, coupled with low visibility, hampers human activities and the effective extraction of resources underwater. Therefore, underwater activities often require the use of robots, such as Autonomous Underwater Vehicles (AUVs) and Remotely Operated Vehicles (ROVs) [1]. These robots, designed specifically for ocean use, rely heavily on advanced vision capabilities for thorough environmental inspections. These include monitoring oceanic environments [2,3], exploring and developing seabed resources [4,5], tracking and monitoring marine populations [6,7], and even excavating marine archaeological sites [8].
Capturing high-quality visual data in complex underwater environments presents significant challenges. The inherent properties of water significantly affect light propagation, leading to the degradation of underwater image quality. Light of different wavelengths is absorbed at different rates; for example, red light is quickly absorbed within a few meters of water, giving images a dominant bluish or greenish hue [9,10]. This phenomenon significantly impacts both the visual quality of images and the accuracy of color-based analysis in marine research and underwater robotics. Furthermore, suspended particles in water, like plankton and sediment, exacerbate the issue by scattering light [11]. This scattering effect significantly reduces image clarity, causing blurring, detail loss, and decreased contrast. Considering these challenges, optimally extracting information from directly captured underwater images is a formidable task [12]. The degraded quality of these images necessitates the use of image enhancement techniques for tasks such as object recognition, habitat mapping, and species identification. These techniques aim to correct color imbalances and enhance contrast, details, and overall image quality by compensating for water’s adverse effects on light propagation. Many underwater image enhancement methods have been developed to address these challenging issues. Non-physical-model-based methods have proven effective in enhancing the sharpness and luminance of underwater visuals [13,14]; however, such approaches frequently lead to over-enhancement and excessive color saturation. Physical-model-based methods are enhanced according to the physical model given by [15], but it is usually difficult to accurately estimate the parameters of the underwater imaging model [16,17]. Deep learning-based methods still face a scarcity of high-quality training images [18,19]. These methods restore underwater images to approximate their true appearance by compensating for the adverse effects of water on light propagation. This facilitates a more accurate interpretation of underwater scenes and improves the performance of automated analysis systems, paving the way for effective water environment exploration, monitoring, and protection [20,21].
However, there are still some details of the image that will be improperly processed by different methods of underwater image enhancement, and this kind of information is often ignored in the enhancement process, resulting in over-enhancement or insufficient enhancement in different parts of the image. To address this issue, we propose an image enhancement method based on salient region detection. By identifying and focusing on the most visually important regions in an image, enhancement techniques can be applied more effectively, ensuring that key features are properly enhanced.
In this paper, we propose a salient region detection-guided fusion method for underwater image enhancement dedicated to post-exploration image processing, which aims to enhance underwater images with more realistic and rich details. Compared to other methods, our approach emphasizes underwater image color correction to ensure that post-exploration images achieve more accurate and natural colors.
By utilizing dehazing and color correction processes, our method alleviates the problem of the underwater image haze present when using previous fusion methods. The main contributions of our work are shown below:
  • Considering the two issues of color deviation and haze in underwater images, we propose a fusion framework that integrates two distinct enhancement branches. A single underwater image dehazing branch based on prior knowledge is applied to reevaluate the atmospheric light value of the channel and handle the haze problem. A color calibration branch is proposed to alleviate the color deviation caused by special underwater conditions.
  • Guided by the calculated salient region coefficients, the results from the two branches can work together to preserve and enhance the most visually salient elements in the final output, highlighting the key features of the underwater scene.
  • Our proposed method demonstrates significant performance improvements over other methods based on five evaluation metrics, as evidenced by experimental results on three publicly available underwater datasets.
Overall, we combine four different modules to handle complex cases like color cast and haze in underwater images. These modules can be divided into two enhancement branches that deal with different cases while under the guidance of the calculated salient region coefficients; the results of these two branches can work together to retain and enhance the most visually salient elements in the final output, highlight key features of the underwater scene, and further improve the visual quality of underwater images.
The details of the remaining parts are arranged as follows: Section 2 introduces the underwater image degradation model and related methods for underwater image enhancement. Section 3 shows the specifics of our proposed method and describes the implementation process. Section 4 provides a comparison and analysis of the experimental outcomes. Section 5 provides a summary of the findings and explores possible directions for future research.

2. Theoretical Background and Methodology

2.1. Underwater Image Degradation Model

Underwater image acquisition is important for various applications but is challenging due to the optical properties of the underwater environment. Light is selectively absorbed at different wavelengths, and suspended particles cause light scattering, resulting in color deviation and high turbidity in the captured images [22].
The model proposed by Jaffe [15] and McGlamery [23] can be seen in Figure 1, which shows the three main components of light illumination in an underwater image scene: direct illumination  E d , forward scattering  E f s , and backward scattering  E b s , as shown in Equation (1):
E T = E d + E f s + E b s
And, the forward scattering component, which is a minor factor in light deflection, has a negligible effect on image degradation, so it can be disregarded. Then, the formation of underwater images is expressed by Equation (2):
I ( x ) = J ( x ) · t ( x ) + A ( x ) · ( 1 t ( x ) )
where  I ( x )  is the observed intensity of the degraded image centered at x, with c encompassing each RGB channel. A clean image has a luminance of  J ( x ) . The transmission or scattering of light within the underwater environment that reaches the camera is denoted by  t ( x ) J ( x ) · t ( x )  is known as direct attenuation, equivalent to  E d , and contains scene information, representing light illuminating the object and scattering to the camera.  A ( x )  denotes backscattered light, where  A ( x ) · ( 1 t ( x ) )  is backscattered and analogous to  E b s . As a result of backscattered light, light is reflected from particles transmitted to the camera, causing the scene to change color.

2.2. Current Underwater Image Enhancement Methods

Hardware-based methods for underwater information acquisition, e.g., polarization descattering imaging [24,25] and range-gated imaging [26,27], are commonly used. Although these technologies are effective for restoring and enhancing underwater images, they have limitations. Significant expenditure is required to acquire high-end equipment, and obtaining consistent image sequences of an identical scene demands considerable effort.
Physical-model-based methods are employed for underwater image enhancement, as the characteristics of underwater images are similar to those of hazy images. Therefore, physical dehazing models are commonly used in the restoration of underwater images. For example, the dark channel prior [28] (DCP) dehazing method is used to mitigate visual effects in hazy images. The essential concept of this approach is derived from the observation that, in natural scenes, even in haze conditions, certain regions of pixels retain very low brightness. These regions, characterized by significantly low luminance, are commonly referred to as the DCP. Some researchers have used the DCP method in underwater environments and demonstrated its feasibility and shortcomings. They then proposed an improved method based on the DCP. Addressing the issue that underwater red channels appear almost dark with the DCP method, Drews Jr [20] proposed an improved underwater DCP (UDCP) based on the original DCP. In the UDCP method, the DCP is applied to the blue and green channels, thereby enhancing image transmission estimation. The GDCP [29] is based on different assumptions and generalizes the DCP for use with underwater images. Galdran et al. [30] improved the DCP method by inverting the R channel based on previous studies and introduced saturation calculations to avoid interference from artificial light sources in underwater images. They called this method the automatic dark red channel prior method. Bryson.M [6] introduced a technique for estimating depth in underwater environments by analyzing image distortion and light attenuation, which is applicable to restoring and enhancing underwater images using the Image Formation Model (IFM). The above-introduced underwater image optimization methods from previous research are based on physical models. This involves having prior knowledge and making assumptions about environmental conditions, establishing a degradation model, calculating model parameters, and solving the inverse problem.
Non-physical-model-based approaches leverage summation rules within the spatial or frequency domains to modify pixel values or frequencies accordingly [31]. Iqbal [32] compensated for underwater red light attenuation by pulling up the R channel of images in RGB space and then converting them to HIS space to stretch brightness and improve image quality. The method balances uneven illumination and image contrast to some extent and performs color balancing and contrast correction. E. Lang [33] proposed a Retinex theory based on color constancy perception, and many image enhancement methods derived from this theory have been utilized to improve the quality of underwater images [34]. In Garg’s work [35], Contrast-Limited Adaptive Histogram Equalization (CLAHE) was combined with the percentile method to improve contrast and correct color in underwater images. Wang et al. [36] proposed an enhancement method based on minimal color loss and locally adaptive contrast enhancement (MLLE), which uses integral maps to calculate the mean and variance of local image blocks for adaptive contrast adjustment in color correction.
The deep learning method, enabled by hardware enhancements, allows for underwater image enhancement by learning hidden image features. It has seen increasing application to underwater image optimization in recent years. Perez [37] was the first to utilize convolutional neural networks (CNNs) for underwater image enhancement. The method involves training the model with features extracted from both clear and degraded underwater images to achieve enhancement. However, this approach has limitations. To address this, Wang [38] proposed two CNN-based training models for underwater image color restoration and dehazing. Subsequently, Li et al. [39] proposed WaterGAN, an underwater image synthesis method based on generative adversarial networks, and a dual-phase network for image restoration, addressing color correction and depth estimation. Zong [40] proposed a U-shaped Transformer network, which marked the first use of the Transformer model in the domain of underwater image enhancement. Islam [41] proposed FUnIE, a model based on conditional generative adversarial networks. Huang et al. [42] proposed a semi-supervised underwater image restoration (Semi-UIR) method, which utilizes the mean teacher approach to incorporate unlabeled data into the training process. According to Li [43], the Water-Net model, which is a convolutional neural network for enhancing underwater images, is based upon the concept of underwater scene priors. Methods based on deep learning for underwater image enhancement encounter major difficulties because of the limited availability of high-quality training datasets.
In addition, there are researchers who enhance underwater images using fusion methods. Ancuti [44] initially introduced this fusion method, which was later refined to address image artifacts during the fusion process and to enhance the fusion effect [45]. Guo et al. [46] proposed an underwater image enhancement method using a multi-scale fusion approach based on human visual system properties. It fuses results from underwater image enhancement methods addressing color-casting, sharpness, and contrast degradation.
Although existing methods can enhance the visual quality of underwater images, they still have the limitation that the enhanced image is easily over-enhanced and oversaturated. Therefore, we propose a new underwater image enhancement method that divides the original underwater image into two stages, namely, haze removal and color correction, and the enhanced images of the two stages are input into the salient region fusion method. The subsequent section provides a comprehensive explanation of our proposed framework.

3. Models and Methods

In this section, we propose a robust and effective underwater image enhancement method. Our framework builds on the basis of salient region detection-guided fusion to obtain two inputs by enhancing the original underwater image. The proposed method is mainly divided into four phases: underwater dehazing, color correction, salient region detection, and image fusion. Firstly, the color deviation and haze present in the original image are removed by two enhancement methods. Finally, after the saliency weight calculation, a complementary dominance relationship between the two inputs is established, and the resulting image is generated via fusion. This alleviates the problem of underwater color degradation and the loss of details in the salient region during the image enhancement process. Figure 2 presents the system flowchart. Each block will be detailed in Section 3.1, Section 3.2, Section 3.3, Section 3.4.

3.1. Single-Image Dehazing Method

Existing dehazing methods for underwater images often adapt techniques from terrestrial image dehazing. One commonly used approach is the dark channel prior [28], which is effective in removing haze by estimating the transmission map. However, underwater environments present unique challenges, particularly with the red channel, which is significantly absorbed and often close to zero.
We propose a dehazing method for underwater images. In contrast to the main formation model, we aim to alleviate haze from underwater images. Inspired by [20], we apply the DCP only to the green and blue channels due to the difficulty of modeling the behavior of the red channel, and we do not estimate or model the red channel in the secondary part. This severe red degradation phenomenon is mainly related to the high absorption effect of the red channel underwater, which, in many cases, makes it close to zero. In many water regions, at least one color channel exhibits low intensity at certain pixels, as shown in Equation (3):
U d a r k ( x ) = min y Ω ( x ) ( min c G , B I c ( y ) )
where  U d a r k  is the underwater dark channel image, and c is the color channel of the original image I Ω ( x )  is the local patch centered at x.
On this basis, we improve the estimation of the global ambient light value. Different from the GDCP [29] method, the underwater dark channel image used in our ambient light estimation only contains blue and green channels, which is the ambient light value in the water; we estimate the ambient light according to the input pixels corresponding to the top 0.2% brightest pixels in the underwater image, where  U A  represents the underwater ambient light value, and  U d a r k 0.2 %  is the top 0.2% maximum pixel set in the underwater dark channel prior map. The degraded image’s intensity at the coordinate x is indicated by  I c ( x ) . This process can be formulated as follows:
U A = 1 U d a r k 0.2 % x U d a r k 0.2 % I c ( x )
In the underwater image without haze, the underwater dark channel  U d a r k  tends to 0, and the transmittance equation can be obtained as Equation (5):
t ˜ ( x ) = 1 φ min y Ω ( x ) ( min c G , B I c ( y ) U A )
where  t ˜ x  is the transmittance. In fact, even in clear underwater imaging, atmospheric light in water is not absolutely free of particles. If we remove the haze from the water completely, the image may look unnatural, and the sense of depth may be lost. Therefore, we can selectively maintain a very small amount of haze for distant objects by introducing a constant parameter  φ ( 0 < φ 1 )  into the equation, which is fixed to 0.75 in our practices.
We can calculate the rough transmission map from the estimated transmittance, so we perform guided filtering on the rough transmission map and set the filtering radius to 18 to obtain the filtered transmission map content as follows:
I 1 ( x ) = I c ( x ) U A max ( t ( x ) , t 0 ) + U A
where  max ( t ( x ) , t 0 )  is used to refine the content map  t ˜ ( x )  after guided filtering with transmittance  t ( x ) , and  t 0  is taken as 0.125.  I c ( x )  is the original image, and  U A  is the underwater ambient light value, for which we obtain the underwater haze restoration image  I 1 ( x ) , which can be expressed as Equation (6) and is the first enhanced image.

3.2. Removal of Color Deviation

Since the problems of poor contrast and color quality degradation in underwater images exist, we carry out a series of color and contrast adjustments to improve the overall color richness and high-quality visual effects so as to facilitate the subsequent salient region detection step.
Firstly, we apply the color balance method based on the gray world assumption, and the average  I ¯ c  of each image channel is calculated, as shown in Equation (7) below, where N is all pixels in the image:
I ¯ c = 1 N 1 N I c ( x )
Due to the degradation of underwater images, in order to make the brightness of all color channels more comparable, the brightness adjustment ratio of each channel relative to the brightest channel is calculated. Based on the brightness adjustment ratio, a saturation adjustment level  r c  is set for each channel, as shown in Equation (8), in order to adjust the saturation of the image in the subsequent processing step so that the color is more vivid.
r c = 0.3 % max ( I ¯ R , I ¯ G , I ¯ B ) I ¯ c
For each color channel  I c , two quantiles are first calculated according to the saturation adjustment level as the upper and lower thresholds for adjustment.  Q c  is the saturation threshold to determine the color channel:
Q c = [ r c , 1 r c ]
We use the quantile function to determine the upper and lower limits of saturation adjustment to improve the color performance of the image.  T c  is the intensity value, and  I c  is the pixel value of the original color channel.
T c = q ( I c , Q c )
where  T c  distinguishes the saturation threshold of the  I c  color channel c, for which  T c h i g h  and  T c l o w  are defined.  T c l o w  is the minimum threshold, below which  I c  is less than the channel distinguished by  T c , and  T c h i g h  is the maximum threshold, above which  I c  is greater than the channel distinguished by  T c .
The pixel normalization process is one of the key steps in adjusting the color channel values of the image, which ensures that each color channel of the image is appropriately scaled and translated so that all pixel values fall within a standard range.
I = I c T c l o w T c h i g h T c l o w × 255
where  I c  is the saturation-adjusted image channel. Leveraging the human visual system’s ability to adjust to variations in lighting, we transform the color-balanced underwater image  I  into the CIELAB color space.
To increase the brightness and detail of the darker regions of the background, we use contrast stretching for the L channel in the normalized image  I , as shown below:
I 2 L = ( I L L min ) ( L max L min ) + L min L max L min
where  I 2 L  represents the L channel of our final enhanced image,  I L  is the brightness value of the L channel in image  I , and  L min  and  L max  represent the minimum and maximum brightness values of the L channel in image  I , respectively.  L max  and  L min  represent the brightness values of the L channel after stretching, which we set to 100 and 0, respectively. Finally, we convert the image from CIELAB to the RGB form and obtain the final enhanced image  I 2  of this part.

3.3. Salient Region Detection

A saliency map can help to dynamically adjust the weight allocation of different enhanced images to different regions in the fusion method. Higher weights can be given to regions so that they receive more attention and processing resources in the fusion process, thus enhancing the enhancement effect of these regions. Meanwhile, the saliency map can guide the adaptive fusion strategy to differentiate the processing according to different regions of the image content. This can avoid the drawbacks of global uniform processing, making the fused image more natural and delicate. In the non-significant region, the information content is usually low and easily affected by noise. Through saliency map guidance, enhancement processing can be appropriately reduced in these regions to avoid noise amplification, thus improving the overall quality of the image.
Our method incorporates the saliency estimation method introduced by Achantay et al. [47]. The RGB color space of the image is converted to the CIELAB color space. The CIELAB color space is closer to human visual perception, which makes differences between colors easier to quantify. For each pixel, the squared differences between its values in the L, A, and B channels and the respective channel averages are computed and summed. Saliency weights emphasize objects that lose salient features in underwater scenes by preserving this part of the region information. The salient region detection map calculation for the two enhanced images we obtain can be expressed as follows:
S n ( x ) = I n m ( x ) I n g ( x )
where  I n m  represents the average pixel intensity of the image, and  I n g  represents the Gaussian blurred image.  S n  can be considered a measure of the color difference between a pixel and the rest of the image, which is the resulting salient region detection image. The larger the difference, the higher the saliency value of that pixel.
The clarity of underwater images is diminished during imaging because of water turbidity and the scattering effects of visible light. This method uses salient features for fusion, and according to the inconsistency in the information retained in the image pixels enhanced by the two different methods, the retained salient feature points are fused, and  W ^ n ( x )  is regarded as a separate weight of salient features for calculation.  S 1 ( x )  and  S 2 ( x )  are the salient region detection maps of enhanced images  I 1  (using image dehazing) and  I 2  (using color deviation removal), respectively.
W ^ n ( x ) = S n ( x ) S 1 ( x ) + S 2 ( x )

3.4. Enhancement Image Fusion

The main reason that we fuse the two resulting color-corrected and dehazed images is to combine the advantages of two different enhancement techniques to generate a better image. While a single enhancement method may perform poorly in some cases, the fusion method can improve the robustness and stability of the processing results and avoid some possible defects of a single method in terms of visual effects. We compute the salient region weights for the salient region maps of the two enhancement maps. The salient region information usually contains more key information, and fusion based on the salient region weights can ensure that these important regions are better processed and displayed, thus improving the fusion effect and ensuring visual consistency in the image so that the image looks more harmonious and natural on the whole.
The same image contains useful information at different scales. The denser the levels, the more abundant the useful information. A pyramid is a common hierarchical structure, which is a collection of images arranged in a pyramid shape with gradually decreasing resolution. The top-down upsampling operation of the Gaussian pyramid cannot recover the source image, so the Laplacian pyramid is used to realize the reconstruction of the Gaussian pyramid. The bottom layer is the source image to be processed, which has the highest resolution, and the processes of decomposition and fusion are shown in Figure 2.
In this step, the two enhanced images are split by Gaussian pyramid decomposition and Laplacian pyramid decomposition, and the image pyramid  D l ( x )  is reconstructed by combining the saliency weight and the enhancement map  I n ( x ) . And,  I n ( x )  represents the enhanced image  I 1  (using image dehazing) or  I 2  (using color deviation removal).
D l ( x ) = ( L l I 1 ( x ) · G l W ^ 1 ( x ) ) + ( L l I 2 ( x ) · G l W ^ 2 ( x ) )
where  G l W ^ n ( x )  represents Gaussian pyramid decomposition,  L l I n ( x )  represents Laplacian pyramid decomposition, l represents the number of pyramid layers, and  D l ( x )  is the output fusion result of Gaussian and Laplacian decomposition incorporating the information from each subimage.
Finally, the output image after pyramid fusion is upsampled, guided by the salient regions:
F o u t ( x ) = l D l ( x ) d
where  F o u t ( x )  is the result of the Gaussian and Laplacian decomposition of each subimage’s information fused by upsampling, which is the image fusion result and the output of the enhanced image.  d  is the upsampling operation, and d is the number of samples.

4. Experiment and Analysis

In this part, we demonstrate the efficacy of our method through a comparative analysis with various methods for improving underwater images. Figure 3 illustrates the results of each key stage in the process. The final enhanced underwater image clearly shows significant visual improvement.

4.1. Benchmark Datasets

Our method was evaluated using the UIEB [43] and RUIE [48] datasets. The UIEB (Underwater Image Enhancement Benchmark) dataset comprises 890 raw underwater images, each paired with a high-quality reference image, as well as 60 challenging underwater images used for validation in our study. On the other hand, the RUIE [48] dataset includes two data subsets, UIQS [48] and UCCS [48], which provide label files for underwater object detection and images of water degradation. UIQS [48] contains 726 images and assesses how well different methods enhance underwater image visibility. UCCS [48] contains a subset of 100 images in blue, blue-green, and green tones.

4.2. Evaluation Metrics

In the quantitative analysis, five commonly used image quality evaluation indexes are used to demonstrate the advantages of our method in alleviating the haze and color deviation in underwater images and enhancing image details, namely, UCIQE (underwater color image quality evaluation) [49], CCF (colorfulness contrast fog density index) [50], PCQI (patch-based contrast quality index) [51], UIQM (underwater image quality metric) [52], and EI (edge intensity) [53]. UCIQE [49] is a linear combination of color density, saturation, and contrast for quantitatively evaluating uneven color, blurriness, and low contrast in underwater images. The UIQM [52] measures colorfulness, sharpness, and contrast. The CCF [50] combines colorfulness, contrast, and fog density. The PCQI [51] indicates whether the enhanced image has improved visibility. EI [53] indicates the edge intensity.

4.3. Comparison Methods

We compared our method with eight various techniques for underwater image enhancement, including deep learning-based methods (FUnIEGAN [40], Semi-UIR [41], U-shape [39]), a fusion method (CBAF [45]), physical-model-based methods (UDCP [20], GDCP [28], IBLA [12]), and non-physical-model-based methods (MLLE [36]).

4.4. Qualitative Comparisons

In the qualitative comparison, the deep learning-based method demonstrates notable stability in performance. Conversely, the restoration method grounded in physical models faces significant challenges in underwater model estimation, leading to inconsistent image quality. The non-physical-model-based method, while straightforward, often results in suboptimal color restoration. In comparison, the fusion-based method tends to offer more consistent performance and improved color recovery.
Our method was first assessed using the dataset UIEB [43]. The comparison results are shown in Figure 4 and Table 1. It can be seen that FUnIEGAN [40], UDCP [20], GDCP [28], and IBLA [12] are subpar. U-shape [39] and Semi-UIR [41] perform well in restoring images, but the picture details are lost. MLLE [36] preserves the details of the picture, but the color brightness is insufficient. Although CBAF [45] demonstrates good performance, there is a loss of detail and a decrease in image quality. As shown in Table 1, our method achieves the best performance in all evaluation metrics, demonstrating the effectiveness in improving the clarity of underwater images and significantly outperforming the other methods.
In Figure 5, we evaluate the method on the UIQS [48] dataset. We can see that the UDCP [20], FUnIEGAN [40], and the GDCP [28] exhibit subpar performance and have insufficient filtering ability for green. U-shape [39] and Semi-UIR [41] demonstrate superior performance in contrast recovery for scenes, but they can also result in color biases. MLLE [36] performs well for image details, but the image loses color. Although IBLA [12] and CBAF [45] perform well in underwater scenes, the picture still suffers from fog and blur. As shown in Table 2, our method significantly outperforms the other methods.
We evaluated the method on the UCCS [48] dataset, as shown in Figure 6. It can be observed that the UDCP [20], FUnIEGAN [40], the GDCP [28], and IBLA [12] have insufficient filtering abilities for green and blue, and the image color is also dark. MLLE [36] performs well for image details, and its dehazing effect demonstrates significant efficacy, but it diminishes the richness of the image color. U-shape [39], Semi-UIR [41], and CBAF [45] are more effective in restoring the scene’s contrast, but they are still insufficient in preserving the color details. As shown in Table 3, our method is the best in three of the five evaluation indicators compared with other methods, and the dehazing index is second best, because we take into account the color balance and visual light perception and do not blindly choose to completely eliminate color to achieve the ultimate dehazing.

4.5. Comparisons of Detail Enhancement

The detail quality of underwater images is very important for underwater research tasks. Figure 7 compares the detail enhancement achieved by various methods. Globally, our method markedly improves the visual quality of the original image. Locally, our method aims at salient region detection for fusion. Figure 7 demonstrates that our method significantly refines structural details, especially in the area magnified within the red box.

4.6. Ablation Study

To determine the efficacy of our proposed method, we conducted an ablation study to reveal the impact of key components of our method, including (a) the original image, (b) our method excluding single-image dehazing (-w/o SID), (c) our method excluding color cast removal (-w/o CCR), (d) our method excluding saliency-guided fusion (-w/o SGF), and (e) our method with all components.
Figure 8 presents the comparison of the UIEB [43], UIQS [48], and UCCS [48] datasets. The visual results are as follows: (b) -w/o SID cannot correct for haze removal, but the color deviation can be effectively removed; (c) -w/o CCR shows that the haze condition of the underwater image is improved, but the color of the underwater image is not well enhanced; (d) -w/o SGF enhanced the visual effect, but due to the lack of fusion weights, the enhanced image color is not natural; (e) the complete model with all key components achieves a satisfactory visual effect.
Table 4 lists the quantitative scores of ablation models on the UIEB [43], UIQS [48], and UCCS [48] datasets. Table 4 shows that the complete model achieved optimal performance on the three test datasets, demonstrating that each essential element plays a role in the effective performance of our method.

4.7. Running Time Comparison

We evaluated and contrasted the running times of various methods. The deep learning-based enhancement method was run on an Ubuntu 20.04 PC with an NVIDIA Geforce GTX 1660Ti GPU. The traditional enhancement method was run on a Windows 11 PC with Intel(R) Core(TM) i7-12700 CPU, 32 GB memory, and MATLAB 2020b. As shown in Table 5, each method was run fifty times at the corresponding resolution, and the average running time was obtained. Our running time is measured in seconds per image. The deep learning-based enhancement technique leverages GPU parallel acceleration, and our method achieves good performance in running time compared with traditional methods while maintaining good enhancement effects.
In addition, based on all of the above experiments, we separately list MLLE and our method for comparison. In benchmark tests of image quality, our performance is mostly better then MLLE. From the visual effect, compared with the MLLE method, our method has better color balance and image details in the enhanced image. From the perspective of the type of method, MLLE is a non-physical model method, which can quickly perform color correction by summarizing the rules of image pixels in the spatial domain or frequency domain, and the processed image will have a certain deviation in visual quality. Our method incorporates the physical model and solves it according to the underwater image degradation model to achieve more accurate image color correction. In general, our method focuses more on post-exploration image processing, while MLLE is more suitable for real-time processing.
In the task of processing underwater images, our method focuses on the goal of finding the true color of the underwater scenes to make the post-exploration image processing results usable. At the same time, we also begin to expand to the functional perspective, so our method has certain application capabilities in underwater tasks.

4.8. Additional Data Validation

We used a waterproof camera to capture underwater images near Boracay Island in the Philippines and near the coast of Lianjiang County, Fujian Province, with an external lighting system installed on the waterproof camera. Due to the periodic ebb and flow of tides, changes in light and water depth in different areas of the ocean produce different colors. It is rich in marine life, including fish, coral, and so on.
We manually selected about 60 valid images and used them to test our method, as shown in Figure 9. It can be seen that our method also performs well in actual shooting conditions and can accurately recover the underwater color deviation scene, which proves that our method successfully enhances the clarity of underwater images in public datasets and in practical use.

4.9. Application Tests

To validate the effectiveness of our method in different underwater applications, we further conducted experiments on feature matching, image stitching, and target recognition.

4.9.1. Feature-Point-Matching Test

The quality of underwater images is very important for various underwater analysis and application tasks, and high-quality images can significantly improve the accuracy of the analysis. We conducted feature point matching in different underwater visual scenarios to demonstrate our method. To validate the performance of our proposed method in enhancing underwater images, we used the Scale-Invariant Feature Transform (SIFT) operator to perform key feature-point-matching experiments. The effect of image enhancement on feature point detection and matching is evaluated by comparing the matching results of key feature points in the original underwater image and the enhanced image, and the test results are shown in Figure 10. The higher the number of matching feature points, the clearer the texture features of the image. Figure 10 shows the SIFT feature matching of FUnIEGAN [40], Semi-UIR [41], U-shape [39], CBAF [45], MLLE [36], the UDCP [20], the GDCP [28], IBLA [12], and our method. Compared with the other methods, our proposed method has the largest number of matching feature points in enhanced underwater images. The image enhancement effect and clarity are significantly improved, which provides a solid foundation for subsequent underwater image analysis and application.

4.9.2. Image Stitching Test

Image stitching is a technique that combines multiple images into a seamless panoramic image. Image enhancement plays a role in image stitching to improve the effect of feature detection and matching and improve the stitching effect. In order to further verify the matching performance of the proposed algorithm, we tested it with image stitching. We selected several sets of images with sequences from the RUIE [48] dataset and tested our method. In the process of image stitching, we used the image stitching system based on the SIFT and RANSAC (Random Sample Consensus) to stitch our images. Stitching the original image, the results are chaotic due to the color deviation and poor visibility of the image, and the recognition and matching performance are poor. In contrast, our method can complete the orderly stitching of two or three images under the condition of achieving visual improvement and has a certain ability to perform image stitching, as shown in Figure 11.

4.9.3. Object Recognition Test

Finally, the application of our method to object recognition was tested. We used the YOLOv8n network to identify the target of underwater aquaculture, and the application results of target recognition are shown in Figure 12. We can see that the visual effect of our method is significantly improved after enhancement, and at the same time, it has a certain ability to improve the number of recognized targets and the accuracy of target classification.

4.10. Generalization Performance of Our Method

Our proposed method can also perform well in other vision tasks, and we enhance low-light images and haze images. Figure 13 demonstrates that our method can also have good enhancement performance in the case of low light and haze; meanwhile, it is clear from the enhanced figure that the image details and visibility are greatly enhanced, and the contrast and color saturation are also improved. This is because our proposed method provides the basis for image enhancement in terms of haze removal and color correction, and the guided fusion based on salient region detection adds more details to the image enhancement, so the image has a good enhancement effect in terms of details and visibility. This proves that our method has good generalization performance in other vision tasks.

5. Conclusions

We have introduced a method for underwater image enhancement with high-quality and rich color presentation. Our method fully considers the problems of color shift and the presence of haze in underwater images. At the same time, the salient region guides the image fusion process. It can successfully correct the underwater color cast and remove the background haze phenomenon, which is suitable for underwater images under different circumstances. Experimental results demonstrate the effectiveness of our method and prove that our method significantly outperforms previous methods on five evaluation metrics and shows higher color richness for underwater images. The proposed method can serve as a processing step for post-exploration underwater detection. However, our method still has some limitations in handling underwater images that lack primary objects, particularly those acquired under low-illumination conditions and with diverse background environments, such as the appearance of red overtones in special cases. This is due to our focus on the color shift correction in images containing primary objects. This challenging case will be explored in future work.

Author Contributions

Conceptualization, methodology, and writing—original draft preparation, J.Y.; resources, conceptualization, and funding acquisition, H.H.; conceptualization, methodology, writing—review and editing, F.L.; resources, funding acquisition, and writing—review and editing, X.G.; software, validation, and supervision, J.J. and B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Fujian Provincial Department of Science and Technology Announces Major Special Projects (2023HZ025003); Key Scientific and Technological Innovation Projects of Fujian Province (2022G02008); and the Education and Scientific Research Project of the Fujian Provincial Department of Finance (GY-Z220232).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data underlying the results presented in this paper are not publicly available at this time but may be obtained from the authors upon reasonable request.

Acknowledgments

The authors would like to thank the editor and the anonymous reviewers for their valuable comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shkurti, F.; Xu, A.; Meghjani, M.; Higuera, J.C.G.; Girdhar, Y.; Giguere, P.; Dey, B.B.; Li, J.; Kalmbach, A.; Prahacs, C.; et al. Multi-domain monitoring of marine environments using a heterogeneous robot team. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; pp. 1747–1753. [Google Scholar]
  2. Pizarro, O.; Eustice, R.M.; Singh, H. Large Area 3-D Reconstructions From Underwater Optical Surveys. IEEE J. Ocean. Eng. 2009, 34, 150–169. [Google Scholar] [CrossRef]
  3. Guo, D.; Li, K.; Hu, B.; Zhang, Y.; Wang, M. Benchmarking Micro-action Recognition: Dataset, Methods, and Applications. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 6238–6252. [Google Scholar] [CrossRef]
  4. Liu, P.; Wang, G.; Qi, H.; Zhang, C.; Zheng, H.; Yu, Z. Underwater Image Enhancement with a Deep Residual Framework. IEEE Access 2019, 7, 94614–94629. [Google Scholar] [CrossRef]
  5. O’Byrne, M.; Ghosh, B.; Schoefs, F.; Pakrashi, V. Applications of Virtual Data in Subsea Inspections. J. Mar. Sci. Eng. 2020, 8, 328. [Google Scholar] [CrossRef]
  6. Bryson, M.; Johnson-Roberson, M.; Pizarro, O.; Williams, S. Automated registration for multi-year robotic surveys of marine benthic habitats. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 3344–3349. [Google Scholar]
  7. Liu, F.; Fang, M. Semantic Segmentation of Underwater Images Based on Improved Deeplab. J. Mar. Sci. Eng. 2020, 8, 188. [Google Scholar] [CrossRef]
  8. Drap, P.; Merad, D.; Hijazi, B.; Gaoua, L.; Nawaf, M.M.; Saccone, M.; Chemisky, B.; Seinturier, J.; Sourisseau, J.-C.; Gambin, T.; et al. Underwater Photogrammetry and Object Modeling: A Case Study of Xlendi Wreck in Malta. Sensors 2015, 15, 30351–30384. [Google Scholar] [CrossRef] [PubMed]
  9. Jonasz, M.; Fournier, G. Light Scattering by Particles in Water: Theoretical and Experimental Foundations; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  10. Mangeruga, M.; Cozza, M.; Bruno, F. Evaluation of Underwater Image Enhancement Algorithms under Different Environmental Conditions. J. Mar. Sci. Eng. 2018, 6, 10. [Google Scholar] [CrossRef]
  11. Hou, M.; Liu, R.; Fan, X.; Luo, Z. Joint residual learning for underwater image enhancement. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 4043–4047. [Google Scholar]
  12. Peng, Y.T.; Cosman, P.C. Underwater image restoration based on image blurriness and light absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired Unrolling with Cooperative Prior Architecture Search for Low-light Image Enhancement. arXiv 2020, arXiv:2012.00000. [Google Scholar]
  14. Zhang, W.; Pan, X.; Xie, X.; Li, L.; Wang, Z.; Han, C. Color correction and adaptive contrast enhancement for underwater image enhancement. Comput. Electr. Eng. 2021, 91, 106981. [Google Scholar] [CrossRef]
  15. Jaffe, J.S. Computer modeling and the design of optimal underwater imaging systems. IEEE J. Ocean. Eng. 1990, 15, 101–111. [Google Scholar] [CrossRef]
  16. Akkaynak, D.; Treibitz, T. Sea-Thru: A Method for Removing Water From Underwater Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  17. Li, C.-Y.; Guo, J.-C.; Cong, R.-M.; Pang, Y.-W.; Wang, B. Underwater Image Enhancement by Dehazing with Minimum Information Loss and Histogram Distribution Prior. IEEE Trans. Image Process. 2016, 25, 5664–5677. [Google Scholar] [CrossRef] [PubMed]
  18. Li, C.; Anwar, S.; Porikli, F. Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
  19. Guo, C.; Li, C.; Guo, J.; Cong, R.; Fu, H.; Han, P. Hierarchical Features Driven Residual Learning for Depth Map Super-Resolution. IEEE Trans. Image Process. 2019, 28, 2545–2557. [Google Scholar] [CrossRef] [PubMed]
  20. Drews, P., Jr.; do Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission Estimation in Underwater Single Images. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops, Sydney, Australia, 1–8 December 2013. [Google Scholar]
  21. Hu, K.; Weng, C.; Zhang, Y.; Jin, J.; Xia, Q. An Overview of Underwater Vision Enhancement: From Traditional Methods to Recent Deep Learning. J. Mar. Sci. Eng. 2022, 10, 241. [Google Scholar] [CrossRef]
  22. Bailey, G.; Flemming, N. Archaeology of the continental shelf: Marine resources, submerged landscapes and underwater archaeology. Quat. Sci. Rev. 2008, 27, 2153–2165. [Google Scholar] [CrossRef]
  23. McGlamery, B.L. A Computer Model For Underwater Camera Systems. In SPIE Proceedings, Ocean Optics VI; SPIE: Bellingham, WA, USA, 1980. [Google Scholar]
  24. Zhao, Y.; He, W.; Ren, H.; Li, Y.; Fu, Y. Polarization descattering imaging through turbid water without prior knowledge. Opt. Lasers Eng. 2022, 148, 106777. [Google Scholar] [CrossRef]
  25. Han, P.; Liu, F.; Wei, Y.; Shao, X. Optical correlation assists to enhance underwater polarization imaging performance. Opt. Lasers Eng. 2020, 134, 106256. [Google Scholar] [CrossRef]
  26. Han, H.; Zhang, X.; Ge, W. Performance evaluation of underwater range-gated viewing based on image quality metric. In Proceedings of the 2009 9th International Conference on Electronic Measurement and Instruments, Beijing, China, 16–19 August 2009; pp. 4-441–4-444. [Google Scholar]
  27. Tan, C.; Seet, G.; Sluzek, A.; He, D. A novel application of range-gated underwater laser imaging system (ULIS) in near-target turbid medium. Opt. Lasers Eng. 2005, 43, 995–1009. [Google Scholar] [CrossRef]
  28. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  29. Peng, Y.-T.; Cao, K.; Cosman, P.C. Generalization of the Dark Channel Prior for Single Image Restoration. IEEE Trans. Image Process. 2018, 27, 2856–2868. [Google Scholar] [CrossRef]
  30. Galdran, A.; Pardo, D.; Picón, A.; Alvarez-Gila, A. Automatic red-channel underwater image restoration. J. Vis. Commun. Image Represent. 2015, 26, 132–145. [Google Scholar] [CrossRef]
  31. Raveendran, S.; Patil, M.D.; Birajdar, G.K. Underwater image enhancement: A comprehensive review, recent trends, challenges and applications. Artif. Intell. Rev. 2021, 54, 5413–5467. [Google Scholar] [CrossRef]
  32. Iqbal, K.; Salam, R.A.; Osman, A.; Talib, A.Z. Underwater Image Enhancement Using an Integrated Colour Model. IAENG Int. J. Comput. Sci. 2007, 34, 2. [Google Scholar]
  33. Zhuang, P.; Li, C.; Wu, J. Bayesian retinex underwater image enhancement. Eng. Appl. Artif. Intell. 2021, 101, 104171. [Google Scholar] [CrossRef]
  34. Tang, C.; von Lukas, U.F.; Vahl, M.; Wang, S.; Wang, Y.; Tan, M. Efficient underwater image and video enhancement based on Retinex. Signal Image Video Process. 2019, 13, 1011–1018. [Google Scholar] [CrossRef]
  35. Garg, D.; Garg, N.K.; Kumar, M. Underwater image enhancement using blending of CLAHE and percentile methodologies. Multimed. Tools Appl. 2018, 77, 26545–26561. [Google Scholar] [CrossRef]
  36. Zhang, W.; Zhuang, P.; Sun, H.-H.; Li, G.; Kwong, S.; Li, C. Underwater Image Enhancement via Minimal Color Loss and Locally Adaptive Contrast Enhancement. IEEE Trans. Image Process. 2022, 31, 3997–4010. [Google Scholar] [CrossRef] [PubMed]
  37. Perez, J.; Attanasio, A.C.; Nechyporenko, N.; Sanz, P.J. A Deep Learning Approach for Underwater Image Enhancement. In Biomedical Applications Based on Natural and Artificial Computing, Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 183–192. [Google Scholar]
  38. Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017. [Google Scholar]
  39. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised Generative Network to Enable Real-time Color Correction of Monocular Underwater Images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef]
  40. Peng, L.; Zhu, C.; Bian, L. U-shape Transformer for Underwater Image Enhancement. IEEE Trans. Image Process. 2023, 32, 3066–3079. [Google Scholar] [CrossRef]
  41. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  42. Huang, S.; Wang, K.; Liu, H.; Chen, J.; Li, Y. Contrastive Semi-supervised Learning for Underwater Image Restoration via Reliable Bank. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  43. Li, C.; Guo, J.; Guo, R.; Cong, R.; Pang, Y.; Wang, B.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2019, 28, 5590–5601. [Google Scholar] [CrossRef] [PubMed]
  44. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RL, USA, 16–21 June 2012. [Google Scholar]
  45. Ancuti, C.O.; Ancuti, C.; De Vleeschouwer, C.; Bekaert, P. Color Balance and Fusion for Underwater Image Enhancement. IEEE Trans. Image Process. 2018, 27, 379–393. [Google Scholar] [CrossRef] [PubMed]
  46. Guo, P.; Zeng, D.; Tian, Y.; Liu, S.; Liu, H.; Li, D. Multi-scale enhancement fusion for underwater sea cucumber images based on human visual system modelling. Comput. Electron. Agric. 2020, 175, 105608. [Google Scholar] [CrossRef]
  47. Achanta, R.; Hemami, S.; Estrada, F.; Susstrunk, S. Frequency-tuned salient region detection. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009. [Google Scholar]
  48. Liu, R.; Fan, X.; Zhu, M.; Hou, M.; Luo, Z. Real-World Underwater Enhancement: Challenges, Benchmarks, and Solutions Under Natural Light. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 4861–4875. [Google Scholar] [CrossRef]
  49. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef] [PubMed]
  50. Wang, Y.; Ma, K.; Yeganeh, H.; Wang, Z.; Lin, W. An imaging-inspired no-reference underwater color image quality assessment metric. Comput. Electr. Eng. 2018, 67, 904–913. [Google Scholar] [CrossRef]
  51. Wang, S.; Ma, K.; Yeganeh, H.; Wang, Z.; Lin, W. A Patch-Structure Representation Method for Quality Assessment of Contrast Changed Images. IEEE Signal Process. Lett. 2015, 22, 2387–2390. [Google Scholar] [CrossRef]
  52. Panetta, K.; Gao, C.; Agaian, S. Human-Visual-System-Inspired Underwater Image Quality Measures. IEEE J. Ocean. Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  53. Mohd Azmi, K.Z.; Abdul Ghani, A.S.; Md Yusof, Z.; Ibrahim, Z. Natural-based underwater image color enhancement through fusion of swarm-intelligence algorithm. Appl. Soft Comput. 2019, 85, 105810. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of underwater imaging.
Figure 1. Schematic diagram of underwater imaging.
Jmse 12 01383 g001
Figure 2. A flowchart of our proposed method for enhancing underwater images. Firstly, we perform single-image dehazing on the input image. Then, we remove the color deviation from the input image. After that, we detect the salient regions in the previously enhanced maps and calculate the weights. Finally, we fuse the images according to the obtained parameters to achieve the final enhancement result.
Figure 2. A flowchart of our proposed method for enhancing underwater images. Firstly, we perform single-image dehazing on the input image. Then, we remove the color deviation from the input image. After that, we detect the salient regions in the previously enhanced maps and calculate the weights. Finally, we fuse the images according to the obtained parameters to achieve the final enhancement result.
Jmse 12 01383 g002
Figure 3. The results of each key stage in the process: (a) original underwater image, (b) single-image dehazing component, (c) salient region detection component, (d) removed color deviation component, (e) CIELAB component, (f) salient region detection component, (g) enhanced underwater image.
Figure 3. The results of each key stage in the process: (a) original underwater image, (b) single-image dehazing component, (c) salient region detection component, (d) removed color deviation component, (e) CIELAB component, (f) salient region detection component, (g) enhanced underwater image.
Jmse 12 01383 g003
Figure 4. Qualitative comparison results of various methods on the UIEB.
Figure 4. Qualitative comparison results of various methods on the UIEB.
Jmse 12 01383 g004
Figure 5. Qualitative comparison results of various methods on the UIQS.
Figure 5. Qualitative comparison results of various methods on the UIQS.
Jmse 12 01383 g005
Figure 6. Qualitative comparison results of various methods on the UCCS.
Figure 6. Qualitative comparison results of various methods on the UCCS.
Jmse 12 01383 g006
Figure 7. Detail enhancement comparisons.
Figure 7. Detail enhancement comparisons.
Jmse 12 01383 g007
Figure 8. Qualitative ablation results for each key component of our method on the UIEB, UCCS, and UIQS datasets. (a) Original image. (b) -w/o SID. (c) -w/o CCR. (d) -w/o SGF. (e) Our proposed method.
Figure 8. Qualitative ablation results for each key component of our method on the UIEB, UCCS, and UIQS datasets. (a) Original image. (b) -w/o SID. (c) -w/o CCR. (d) -w/o SGF. (e) Our proposed method.
Jmse 12 01383 g008
Figure 9. Additional data validation. (a,b) The top row displays the original image, and the bottom row shows the results of our enhanced underwater image.
Figure 9. Additional data validation. (a,b) The top row displays the original image, and the bottom row shows the results of our enhanced underwater image.
Jmse 12 01383 g009
Figure 10. The results of feature matching.
Figure 10. The results of feature matching.
Jmse 12 01383 g010
Figure 11. The results of image stitching. (a,b,e,f) The original sequence image; (c,d,g,h) the stitching result of the enhanced sequence image.
Figure 11. The results of image stitching. (a,b,e,f) The original sequence image; (c,d,g,h) the stitching result of the enhanced sequence image.
Jmse 12 01383 g011
Figure 12. The results of target recognition. (a) Original target recognition results; (b) target recognition results of our method.
Figure 12. The results of target recognition. (a) Original target recognition results; (b) target recognition results of our method.
Jmse 12 01383 g012
Figure 13. Results of our method in enhancing hazy and low-light images. (a) Comparison of hazy image enhancement; (b) comparison of low-light image enhancement.
Figure 13. Results of our method in enhancing hazy and low-light images. (a) Comparison of hazy image enhancement; (b) comparison of low-light image enhancement.
Jmse 12 01383 g013
Table 1. Average performance metrics of various methods on the UIEB.
Table 1. Average performance metrics of various methods on the UIEB.
FUnIEGANU-shapeSemi-UIRCBAFMLLEUDCPGDCPIBLAOurs
UCIQE0.5650.5640.6170.6020.6100.5980.6120.5420.641
UIQM5.0154.9494.5983.7894.1805.0532.5543.4215.173
CCF21.45321.83428.11028.03148.21541.26033.24339.21146.890
PCQI0.9040.8241.1441.1251.1990.8521.0341.0821.226
EI69.99370.90473.29949.391120.04954.73051.4250.17497.054
Table 2. Average performance metrics of various methods on the UIQS.
Table 2. Average performance metrics of various methods on the UIQS.
FUnIEGANU-shapeSemi-UIRCBAFMLLEUDCPGDCPIBLAOurs
UCIQE0.5160.5460.5660.6020.5810.5060.5910.5830.619
UIQM4.7874.6304.1253.7734.3443.8623.6762.2314.838
CCF18.51320.28523.43929.11245.71029.41021.9826.78336.662
PCQI0.8490.9541.1801.2011.2720.9081.0451.1781.310
EI54.97754.52365.60153.76110.21342.69061.22357.35489.144
Table 3. Average performance metrics of various methods on the UCCS.
Table 3. Average performance metrics of various methods on the UCCS.
FUnIEGANU-shapeSemi-UIRCBAFMLLEUDCPGDCPIBLAOurs
UCIQE0.5030.5390.5530.6010.5770.5290.5730.5890.618
UIQM4.6864.6024.1073.6774.4993.7342.6233.2114.816
CCF17.69420.32121.78927.20443.25131.41525.90125.43437.896
PCQI0.8470.9571.2031.191.2590.9141.0981.1431.297
EI51.03751.15661.17149.876105.04445.40154.33245.76574.974
Table 4. The ablation study on the UCCS, UIQS, and UIEB.
Table 4. The ablation study on the UCCS, UIQS, and UIEB.
AblatedUIEBUIQSUCCS
Models(-w/o SID)(-w/o CCR)(-w/o SGF)(Ours)(-w/o SID)(-w/o CCR)(-w/o SGF)(Ours)(-w/o SID)(-w/o CCR)(-w/o SGF)(Ours)
UCIQE0.5110.5320.4350.6410.5310.4880.5120.6190.5220.4760.5010.618
UIQM4.8313.7314.6325.1734.4354.2314.6304.8384.6304.7314.2714.816
CCF15.6840.28530.28546.89018.22734.68124.28536.66220.28531.23628.28537.896
PCQI1.0520.9840.8111.2261.1511.0040.9541.3101.1040.9930.8541.297
EI98.63156.54162.55397.05484.52348.35266.37289.14475.33561.42748.63274.974
Table 5. Running time comparisons of different methods.
Table 5. Running time comparisons of different methods.
Image SizeFUnIEGANU-shapeSemi-UIRCBAFMLLEUDCPGDCPIBLAOurs
  256 × 256 0.0040.1160.4070.9510.3110.8470.8235.2250.549
  512 × 512 ---2.3340.5092.2131.93224.1241.131
  1024 × 1024 ---6.3751.4066.4825.97189.1533.091
  1980 × 1080 ---12.5222.86111.43511.260242.1135.814
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, J.; Huang, H.; Lin, F.; Gao, X.; Jin, J.; Zhang, B. Underwater Image Enhancement Fusion Method Guided by Salient Region Detection. J. Mar. Sci. Eng. 2024, 12, 1383. https://doi.org/10.3390/jmse12081383

AMA Style

Yang J, Huang H, Lin F, Gao X, Jin J, Zhang B. Underwater Image Enhancement Fusion Method Guided by Salient Region Detection. Journal of Marine Science and Engineering. 2024; 12(8):1383. https://doi.org/10.3390/jmse12081383

Chicago/Turabian Style

Yang, Jiawei, Hongwu Huang, Fanchao Lin, Xiujing Gao, Junjie Jin, and Biwen Zhang. 2024. "Underwater Image Enhancement Fusion Method Guided by Salient Region Detection" Journal of Marine Science and Engineering 12, no. 8: 1383. https://doi.org/10.3390/jmse12081383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop