Next Article in Journal
Multi-Mode Surface Wave Tomography of a Water-Rich Layer of the Jizhong Depression Using Beamforming at a Dense Array
Next Article in Special Issue
Underwater Hyperspectral Imaging System with Liquid Lenses
Previous Article in Journal
Multi-Mission Satellite Detection and Tracking of October 2019 Sabiti Oil Spill in the Red Sea
Previous Article in Special Issue
Sonar Image Target Detection Based on Style Transfer Learning and Random Shape of Noise under Zero Shot Target
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UIR-Net: A Simple and Effective Baseline for Underwater Image Restoration and Enhancement

1
College of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China
2
School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
3
Department of Engineering Mathematics, University of Bristol, Bristol BS8 1TW, UK
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(1), 39; https://doi.org/10.3390/rs15010039
Submission received: 16 November 2022 / Revised: 6 December 2022 / Accepted: 19 December 2022 / Published: 22 December 2022
(This article belongs to the Special Issue Advancement in Undersea Remote Sensing)

Abstract

:
Because of the unique physical and chemical properties of water, obtaining high-quality underwater images directly is not an easy thing. Hence, recovery and enhancement are indispensable steps in underwater image processing and have therefore become research hotspots. Nevertheless, existing image-processing methods generally have high complexity and are difficult to deploy on underwater platforms with limited computing resources. To tackle this issue, this paper proposes a simple and effective baseline named UIR-Net that can recover and enhance underwater images simultaneously. This network uses a channel residual prior to extract the channel of the image to be recovered as a prior, combined with a gradient strategy to reduce parameters and training time to make the operation more lightweight. This method can improve the color performance while maintaining the style and spatial texture of the contents. Through experiments on three datasets (MSRB, MSIRB and UIEBD-Snow), we confirm that UIR-Net can recover clear underwater images from original images with large particle impurities and ocean light spots. Compared to other state-of-the-art methods, UIR-Net can recover underwater images at a similar or higher quality with a significantly lower number of parameters, which is valuable in real-world applications.

1. Introduction

Underwater optical imaging is an important tool enabling humans to understand, exploit and protect the ocean. It has the advantages of intuitive detection of targets, high imaging resolution and high information content. Underwater images are applied in multiple research areas, such as marine biology, archaeology [1,2], fisheries [3], marine resource surveys, underwater target detection [4], etc. However, in high-turbidity waters such as coastal waters or lakes, scatterers in the water (including bubbles, suspended sediment, plankton, etc.) will cause reductions in and scattering of visible light. Severely scattered light enters the imaging system and causes a veiling light effect, producing a lot of noise in the image. As a result, the quality of the captured images is degraded, and the edges and contours of the scene and the targets are blurred. Thus, the features and information in degraded images cannot be effectively extracted and utilized by downstream tasks, such as underwater target detection, segmentation and recognition. This makes the enhancement and restoration of degraded underwater images necessary for marine applications.
There are a great number of solutions, including formation-model-based methods [5] and deep-learning-based methods [6], for enhancing or recovering images on land, but translating these methods for use on underwater images is not an easy task. On one hand, these two kinds of imaging models face different problems; on the other hand, there are no training samples of underwater images for these methods. For these reasons, there exists an urgent need to build relevant high-quality datasets and design enhancement and reconstruction algorithms specifically for underwater images. A number of methods have been proposed for underwater image recovery, which can be broadly classified into traditional methods and depth-based methods. Traditional methods are based on physical models and non-physical models. Non-physical image enhancement methods include white-balancing methods, histogram-equalization methods, Retinex-based methods, etc. Physical image restoration methods construct a mathematical model for the underwater image degradation process and then reverse it to obtain an undegraded image in an ideal state.
With the development and applications of deep neural networks in various scenarios, generative adversarial networks have become increasingly used for image restoration and have achieved impressive visual results [7,8,9,10,11]. Recent work on underwater image recovery using UGAN-P [12], for example, trains a CycleGAN [13] with unpaired clear and degraded underwater images and then generates a training set of degraded underwater images corresponding to the clear underwater images using CycleGAN [13]. To obtain the underwater image training dataset, Water-GAN [14] uses a generative adversarial network to generate images. Jiang [15] has designed a two-stage underwater image recovery network. The first stage is a two-line parallel encoder and decoder structure, and then a dual-channelwise attention operation is carried out. Jiao [16] used Dark channel prior to enhance the image, and refined the rough transmission mapping with the proposed adaptive non-convex non-smooth variational algorithm. The current deep-learning-based approach using generative adversarial networks has three main problems: The first is that it ignores the optical properties of underwater images, thereby easily introducing color biases and artifacts and thus not being ideal for color correction of underwater images. The second is that a CNN often constitutes a neural network with a very large number of parameters and a long training time, making it difficult to achieve a lightweight deployment. The third and most essential problem is that there is no algorithm that combines both tasks, forcing researchers to choose between image enhancement and image restoration.
In order to solve the problems mentioned above, this paper proposes a self-supervised network for underwater image restoration and image enhancement named UIR-Net. In this paper, we consider both tasks as domain-to-domain conversion problems, for which we can achieve high-quality restoration and enhancement of underwater images by mapping high-dimensional features to low-dimensional features through an encoder–decoder structure while preserving the images’ spatial and color features. Firstly, regarding the problem of lacking datasets of clear and degraded underwater images, we created the Marine Spot Impurity Removal Benchmarking (MSIRB) dataset based on the MSRB dataset [17] with scene extension. We also employed UIEBD-Snow, a UIEBD-based dataset [18] that integrates both underwater image enhancement and underwater image restoration to fuse the spots with underwater images. Secondly, to address the GAN-based underwater image restoration methods’ inability to solve the underwater image degradation well, we present a gradient-based policy update scheme and a channel residual prior to extract the channel of the image to be recovered as a prior. Thirdly, to reduce parameters and training time, we introduce a gradient strategy to make the gradient descent algorithm trainable, thus obtaining a more lightweight method. Specifically, we combine a gradient estimation strategy with proximal gradient descent (PGD) algorithm with a learnable step size. Unknown degradation issues in real-world scenarios are solved by combining flexible gradient descent and informative proximal mapping using a deep neural network. Overall, the contributions of this paper are:
1.
In this paper, we present a novel, end-to-end, lightweight underwater image restoration network named UIR-Net, which can generate context-rich and spatially accurate outputs.
2.
UIR-Net is the first network that combines underwater image enhancement and underwater image restoration, which means this can be considered a pioneering study with great significance for practical applications and deployments.
3.
This article proposes the Marine Spot Impurity Removal Benchmarking (MSIRB) dataset and UIEBD-Snow.
4.
Extensive experiments have been carried out on the underwater MSRB [17], the MSIRB dataset, and UIEBD-Snow that demonstrate the effectiveness of our UIR-Net. Our method is able to achieve better results than other state-of-the-art methods. The results of UIR-Net are shown in Figure 1.

2. Related Works

2.1. Underwater Image Restoration

Marine pollutants are one of the main causes of the degradation of underwater images. These pollutants include sinking organic particles, zooplankton remnants, floating excrement, suspended silt and other inorganic materials [19]. They vary in size, shape and transparency. When taking photos (or filming) underwater, the light is scattered by these tiny particles before being reflected back to the camera, resulting in low-contrast and blurry images. Light reflections from these underwater scatterers cause random near-white stains in various shapes, sizes, and opacities in underwater images/videos, decreasing scene perception [20]. It is worth noting that these particles fall to the seafloor in the same way that snow does in atmospheric conditions. There are many existing methods in this area. The authors of [21,22,23,24] proposed a modified median filter to remove the effects of underwater pollutants in these images. Wang et al. [25] advocated for machine vision in marine fisheries and presented an integrated two-channel model to improve underwater picture quality, along with a marine network model for impurity elimination.
DB-ResNet [26] is a “deep detail network” structure designed to remove natural raindrop patterns from captured images. A deep residual network (ResNet) is used as a parameter layer to obtain deeper picture features and to simplify the network’s structure by effectively shortening the mapping distance between input and output features. Other authors [27] proposed a progressive optimization residual network called progressive ResNet (PRN) and a progressive recurrent network (PReNet) to perform image de-raining, which can achieve good de-raining results without increasing network parameters. Maxim [28] is the latest MLP-based U-Net backbone network that implements both global and local perceptual fields and can be applied directly to high-resolution images. Restormer [29] is an efficient transformer that achieves several key improvements in the design of its improved multi-head attention and feed-forward networks MPRNet [30] is a novel collaborative design with a multi-stage structure designed to learn the recovery features of degraded inputs concurrently to decomposing the entire recovery process. Specifically, MPRNet first learns context-dependent features using an encoder–decoder architecture, and then combines it with high-resolution branches that preserve local information better. Compared to the original proximal gradient descent (PGD) algorithm, DGUNet [31] introduces a Deep-Network-based gradient estimation method into its gradient descent progress without loss of interpretability, allowing it to handle complex and real-world image degradation.
All of these methods are very competitive in natural image restoration, but land-based and underwater imaging models are not interchangeable, which means that directly translating image enhancement methods from land-based to underwater image restoration cannot achieve satisfying results; underwater imaging models are simply more complicated.

2.2. Underwater Image Enhancement

Many image enhancement and recovery methods based on marine scenes, such as Neural Networks (NNs) and Dark Channel Priors (DCP) based on Image Formation Model (IFM) class algorithms, have evolved in recent years [32]. In the machine learning literature [33], high-quality restored image data can be generated from a single picture, and the neural network’s generalization was tested using different images. A Convolutional Neural Network (CNN)-based algorithm [34] was proposed in this publication. A Convolutional Neural Network (CNN)-based framework for underwater image improvement implements a pixel interruption method for color correction and defogging. The authors of [14] utilized a Generative Adversarial Network (GAN) to produce a real IFM imaging model. The authors of [35] proposed an underwater image enhancement CNN model based on underwater scene prior, called UWCNN.The authors of [5] generated a huge training dataset encompassing real underwater photos in depth using a GAN for underwater image color correction. An IFM-based method to enhance underwater images by removing background light (BL) based on light transmission parameters is developed in the literature. The authors of [36] simplified the full underwater optical transmission model to extract the transmission map (TM) from the degraded underwater images. Optimal enhancement performance is also guaranteed based on regional background estimation. In [37], the background light is estimated by using a color variable that changes with depth. In [6], Drewa obtains the transmission map with an underwater Dark Channel Prior (UDCP) model by reducing the R channel’s effect on the dark channel. In [38] Carlevaris et al. used a Maximum Intensity Prior (MIP) model to estimate the transmission map to recover underwater images. In [39], Song offered an Underwater Light Attenuation Prior (ULAP) model in order to obtain the scene depth of underwater images. The authors of [40] proposed an Image Blurriness and Light Absorption (IBLA) algorithm, which achieves the estimation of the system parameters by selecting the background light and depth map from different scene regions. The S-model is used to fuse the parameters and estimate the system parameters to complete the image’s recovery.Recently, attention mechanisms have been successfully used in image processing tasks, such as image denoising [41], image super-resolution [42] and image segmentation [43]. The performance of the image enhancement model can be effectively improved by leveraging the focus of the attention mechanism.
However, the existing underwater image enhancement methods only achieve enhancement (color correction). Restoring underwater images degraded by visual impurities to a satisfactory level still remains a challenge.

3. Method

UIR-Net is inspired by DGUNet [31]. Achieving high-quality image restoration and color correction is an essential part of utilizing underwater optical images. The biggest advantage of self-supervised learning is that the results generated by our trained model can match the labels to a greater extent. This paper takes advantage of the self-supervised method, divides the images to be enhanced and the original ones into two domains and forms a domain-to-domain conversion to achieve image restoration and image enhancement after a lot of learning. UIR-Net conforms to the U-Net encoder and decoder principle, as shown in Figure 2. UIR-Net mainly consists of two parts: the first part is the residual prior module, which guides the subsequent underwater image recovery by subtracting the maximum and minimum features as a prior, and the second part is the strategy gradient network, which integrates PGD strategies for iterative training.
In addition, perceptual loss is added to the generator to further constrain the detail and semantic information of generated images. In the discriminator, we use a fully convolutional block-based discriminator to learn the loss of the structure. A great number of approaches combining traditional optimization algorithms with DNNs have been proposed. These methods usually require predefined degradation processes or artificial assumptions and are therefore difficult to handle complex underwater real-world applications.
The structure of UIR-Net is shown in Figure 2.

3.1. Channel Residual Prior

This part refers to Kaiming He’s dark channel defogging algorithm, which can restore the color and visibility of an image and which also uses the fog concentration to estimate the distance of an object. He’s work focuses on the statistical features of fog-free images. It was found that, in a fog-free image, every local region is likely to have shaded areas, something in black or something in pure color. Therefore, it is highly likely that there always exists at least one color channel with very low light values. He named this statistical law Dark Channel Prior or Dark Primary Color Prior. Intuitively, Dark Channel Prior assumes that there is always something very dark in every local area, and it is the essential fundamental law in their study of defogging problems. Since fog is always grayish, once the image is affected by fog, then these features that are supposed to be dark will become grayish. Moreover, based on the physical fog formation formula, they can also determine the fog’s concentration based on how gray these features are. Therefore, their proposed Dark Channel Prior can remove the effect of fog very effectively while using the concentration of features to estimate the distance (depth) of objects.
For dark tones and ocean impurity occlusion, if we think about it, we can identify a statistical rule for underwater images: for most underwater images without impurity particles, at least one of the R, G and B values of any pixel is very low (this is quite understandable, as if all the three values are all high, then that pixel obviously has a tendency to be overly white). The “darker” values (channels) are arranged in a certain way to form a dark channel map corresponding to the original image. Based on this idea and the laws of statistics, the basic model of the image can be reconstructed more clearly without any streaks in the image, but we found that the structures of some constantly changing objects in the reconstructed image are deformed. To address this problem, we propose introducing a channel residual prior (CRP) model. Furthermore, the CRP lacks spots and bubbles, which is beneficial to derive altered background information from the residuals of the maximum and minimum channel values of the picture.
Referring to Li et al. [44], we can express the color intensity of the rainband image as follows:
G ˜ ( x ) = t γ r s ( x ) A α + ( T t ) R π
where G ˜ ( x ) is the intensity of the color vector indicating the presence of color, and γ r s consists of the refraction coefficient, the specular reflection coefficient and the internal reflection coefficient of the spot [45]. A = ( A r , A g , A b ) T expresses the luminance, R = R r + R g + R b · α = A / A and π = R / R denotes the the reflection coming from the background, R = R r + R g + R b . α = A / A and π = R / R are the A and R chromaticity, respectively. The exposure duration is T, and the time it takes for raindrops to flow over pixel x is t. The first term in Equation (1) is the rain streak term, and the second term is the backdrop term. We can achieve the following normalization steps when predicting α using any known color constancy algorithm:
G ( x ) = G ˜ ( x ) α = G r s ( x ) l + G b g ( x )
where l = ( 1 , 1 , 1 ) in Equation (2), T , G r s = t β r s A , and G b g = ( T t ) R / α .
When we normalize the image, the chromaticity of the light deteriorates, erasing the color, and the spectrum’s sensitization impact is likewise eliminated. Therefore, according to Equation (2), given an underwater bubble image G, one can define a residual channel prior P for G:
P p r i o r ( x ) = max c r , g , b G c ( x ) min d r , g , b G d ( x ) .

3.2. Gradient Strategy

From a Bayesian maximum a posteriori perspective, the traditional model-based approach defines image reconstruction as follows:
x ^ = argmax x log P p r i o r ( x y ) = argmax x log P p r i o r ( y x ) + log P p r i o r ( x )
where log P p r i o r ( y x ) represents the data fidelity term and log P p r i o r ( x ) represents the prior term. Usually, the L 2 loss function is used as the data fidelity term. Therefore, the above equation can be further deformed as:
x ^ = argmin x 1 2 y B x 2 2 + λ J ( x )
Based on the proximal gradient descent (PGD) algorithm [31], the above equation can be developed as an iterative solution process:
x ^ k = argmin x 1 2 x x ^ k 1 θ g x ^ k 1 2 2 + λ J ( x )
The first part represents the gradient descent, and the second part represents the proximal mapping. In turn, the above equation can be decomposed into:
v k = x ^ k 1 θ B B x ^ k 1 y x ^ k = prox λ , J v k
where A represents the degradation matrix, and p r o x is the known degradation matrix B. It is difficult to deal with the real scenes where the degradation is not known. In this paper, based on the PGD algorithm, we expand it with a deep network. In detail, the degradation matrix A and the proximal mapping equation are replaced by a deep network, and the step size θ of the gradient descent is set as a learnable parameter. Thus, the gradient descent process is redefined as:
v k = x ^ k 1 ρ k F B k F B k x ^ k 1 y
The proximal mapping module also incorporates inter-stage information to overcome the inter-stage information loss in the deep unfolding network:
x ^ k , F k = prox ω k v k , F k 1
and the structure of PGD is shown in Figure 3, from the structure, the gradient is first passed through a residual layer, then by doing the difference with itself, and then summed with the prior feature, then we assigned a dynamic based on 0–0.5 parameters θ and the output feature matrix just multiplied, after the residual layer output.

3.3. Loss Function

We apply the perceptual loss [46] L p e r c t u a l and Edge loss [30] L e d g e in the network. As the name implies, L p e r c t u a l penalizes outcomes that are not close to the references in the sense of activation in the pretrained network by defining distance measurement. L p e r c t u a l is defined as:
L p e r c t u a l = E i 1 N i ϕ i I g t ϕ i I o u t 1
where ϕ i is the activation map of the i-th layer of the pre-trained network. For our work, ϕ i corresponds to activating the VGG-19 network pre-trained on the ImageNet dataset from the layers relu1-1, relu2-1, relu3-1, relu4-1 and relu5-1. The edge loss is defined as follows:
L edge = Δ X S Δ ( Y ) 2 + ε 2
where ε is empirically set as 10 3 , and Δ means Laplacian (for extracting edges). Thus, the final loss is:
L f i n a l = L edge + λ p e r c t u a l × L p e r c t u a l
There the λ p e r c t u a l equal 0.01, the ablation study of λ p e r c t u a l is shown in Section 4.7.2.

4. Experimental

4.1. Datasets of Underwater Restoration

1.
Marine Snow Removal Benchmarking Dataset (MSRB)
In this section, we present the specifications of the ocean snow artifacts synthesized by the MSRB dataset [17], which has been mainly built for two general tasks of marine snow artifacts removal: MSR task 1 is dedicated removing small-size artifacts, while MSR task 2 is used to cope with various artifacts of different sizes.
Obviously, it would be much more difficult to handle multiple sizes of underwater snow than just focusing on small-sized ones. Corresponding to each MSR task, each sub-dataset is composed of a training set with 2300 pairs of images and a test set with 400 pairs, all with a pixel resolution of 384 × 384. Each image pair includes an original underwater collected picture and one containing synthetic marine snow artifacts, original underwater collected picture is the ground truth of the dataset. Each composite image is added with N marine snow particles, while N, which is the number of added particles, is generated by a discrete uniform distribution of U{100,600}. Based on our preliminary observations, in each synthetic image, H-type and V-type ocean snow spots are randomly generated with a probability of 0.7 for H-type and 0.3 for V-type. Most of our parameters are chosen based on our observations of real-world collected images and the corresponding artificial influence of ocean snow.
2.
Marine Spot Impurity Removal Benchmarking Dataset (MSIRB)
At present, there is no accessible underwater dataset of ocean snow images that can be used to train and test deep neural networks to eliminate ocean snow particles, which represents a major inconvenience to relevant ongoing research on marine light spot removal algorithms. In order to present the diverse morphological features of the real-world ocean snow particles and the complex composition of ocean snow scenes, we propose a new dataset called the Marine Spot Impurity Removal Benchmarking Dataset (MSIRB). The MSIRB dataset contains (1) synthetic ocean light spot images, (2) the corresponding real images and (3) an ocean light spot mask. We reference [47] and use its mask to produce underwater light spot images, each basic mask corresponds to small, medium, and large grain sizes. The dataset is shown in Figure 4.
It is important to focus on the fact that inside the MSIRB dataset’s ground truth and input binary images, there is no difference in color, only the presence or absence of light spot impurities.
3.
An underwater image enhancement benchmark dataset and beyond with Snow (UIEBD-Snow)
To explore the generalization of our work in image enhancement tasks, the same operation that was applied on the MSIRB dataset is performed in this paper on the UIEBD dataset, a dataset with its potential ground truth generated by adopting a dozen of cutting-edge image improvement methods. Various morphological features of real snow particles of the ocean are fused inside the UIEBD to make our new dataset called UIEBD-Snow, which contains 890 underwater images with corresponding reference maps.
It is worth noting that the difference between UIEBD-Snow and MSIRB is mainly in the color correction, and the dataset example can be referred to Figure 4

4.2. Training Parameter Settings

Our UIR-Net is trained end-to-end, and our training strategy is to use the same setting as MPRNet and DGUNet [30,31] in our experiments. Specifically, we use the Adam optimizer with an initialized learning rate of 2 × 10 4 . A typical warm-up strategy [43] is adopted to gradually increase the learning efficiency considering the depth of our network. Our model is trained with images which are randomly cropped into 256 × 256 image patches from our training set. Our batch size is set to 16 and 4 × 10 5 iterations. In image compression perception, the network is trained on 32 × 32 image patches with a learning rate of 1 × 10 4 . For 200 epochs, the batch size is set to 128. Our model training is performed on 3090 GPU and can be completed in 4 hours. Our experimental results are shown in Figure 5 and Figure 6. The source code for UIR-Net can be obtained from https://github.com/meixinkui/UIR-Net, accessed on 1 November 2022.

4.3. Evaluation Metrics

1.
Full-Reference Evaluation
We adopt the standard metrics (PSNR, SSIM and RMSE [48]) as full-reference evaluation.
PSNR is short for Peak Signal to Noise Ratio. The better the PSNR, the less the image distortion. We can obtain the PSNR as follows:
P S N R = 10 log 10 2 n 1 2 M S E
Assume that the current image I and the reference images K, H and W are the height and width of the image respectively. MSE can be calculated as follows:
M S E = 1 H W i = 1 H j = 1 W [ I ( i , j ) K ( i , j ) ] 2
SSIM is short for Structural Similarity, we can obtain the SSIM of the two images (image x and image y) as follows:
SSIM ( x , y ) = 2 μ x μ y + c 1 2 σ x y + c 2 μ x 2 + μ y 2 + c 1 σ x 2 + σ y 2 + c 2
where μ x , μ y are the means of x and y, and σ x , σ y , σ x y represent the variance and covariance of x and y, respectively. C 1 and C 2 are used to maintain stable constants.
RMSE is short for Root Mean Square Error. We can obtain RMSE as follows:
R M S E = 1 N i = 1 n Y i f x i 2
2.
Non-Reference Evaluation
UIR-Net is supervised network, there is an image pair for input and the corresponding target, but we also adopt the UCIQE [49] and UIQM [50] as non-reference evaluation, which are commonly used for underwater image quality assessment.
UCIQE is a linear combination of color intensity, saturation and contrast, used to quantitatively evaluate the non-homogeneous color shift, blurring and low contrast of underwater images. The better the UCIQE, the better the underwater image quality, we can obtain UCIQE as follows:
U C I Q E = c 1 σ c + c 2 con l + c 3 μ s
where c 1 , c 2 and c 3 are the weight factors of each measurement component in the linear combination, σ c is image chromaticity measurement component, con l is image contrast measurement component and μ s is image saturation measurement component.
UIQM is short for underwater image quality measurement. We can obtain it as follows:
U I Q M = c 1 U I C M + c 2 U I S M + c 3 U I c o n M
where c 1 , c 2 and c 3 are the weight factors of each measurement component in the linear combination, UICM is underwater image colorfulness measure, UISM is underwater image sharpness measure and UIconM is underwater image contrast measure. The better the UIQM, the more consistent the results are with human visual perception.

4.4. Comparison with State-of-the-Art Methods on Underwater Image Restoration

We compared UIR-Net with existing methods, including natural image restoration methods: AirNet [51], DB-ResNet [26], PReNet [27], DGUNet [31], Maxim [28], Restormer [29], and MPRNet [30]. In AirNet, the network structure consists of a base layer and a detail layer. The detail layer is fed into this two-layer convolutional neural network, and then, the detail image after de-raining is obtained by deconvolution. The enhanced base layer and the enhanced detail layer are superimposed linearly to obtain the de-rained image. We think this is similar to the underwater image recovery process, and most underwater images are enhanced, so we chose this method for comparison with our proposed method.

4.5. Comparison with State-of-the-Art Methods on Underwater Image Enhancement

We compared UIR-Net with existing methods, including PUIE [52], CWR [53], IFM [54] and FUnIE-GAN [55]. PUIE [52] is a multi-color space encoder network in which the adaptively selected features of different color spaces are fused based on the attention mechanism. A decoder network based on medium transmission guidance is used to guide the model to give more attention to degraded regions. The complementary benefits between underwater imaging domain knowledge and DNNs are discussed in this work. CWR [53] introduces contrast learning into underwater image enhancement, takes the original image as a reference and restores the underwater image through TV loss. FUnIE-GAN [55] uses a structure called FUNI-Gan. The generator uses a five-layer U-net structure. The discriminator part uses PatchGAN, which can check the quality of the generated pictures in blocks. IMF [54] proposes a network consisting of two modules. One aims to estimate backscatter, and the other is used for estimating the direct transmission of light. A carefully designed reconstruction module follows the above two and uses their previous outputs to produce the enhanced underwater image.

4.6. Qualitative Evaluations

4.6.1. Quantitative Comparisons of Underwater Image Restoration on the MSRB Dataset

As we can see in Figure 7, AirNet [51] and PReNet [27] barely remove most of the small near-white spots. PReNet [27] removes the ocean snow grains while placing some dark particles near the original stained location, resulting in an overall darker picture as a result. DB-ResNet [26], maxim [28] and Restormer [29] can remove a small portion of the ocean snow grains while preserving the boundaries of other objects. However, these three methods still perform poorly when removing large ocean snow particles. Our method and MPRNet [30], DGUNet [31] can remove large particles in the ocean very well. In terms of the existing multi-stage techniques, we believe that MPRNet [30] structurally uses an encoder–decoder architecture (the encoding stage effectively propagates contextual information, but does not do much to preserve spatial details) using single scales at the same time. The spatial accuracy and semantic information are also well-preserved, and using both structures at the same time is effective for underwater image restoration.
As we can see in Table 1, UIR-Net achieves the best results at the values of SSIM RMSE and UICQE. The values of PSNR and UIQM are second.

4.6.2. Quantitative Comparisons of Underwater Restoration on MSIRB Dataset

The results of the ocean light spot removal example are shown in Figure 8. AirNet [51], PReNet [27], maxim [28] and Restormer [29] still have defects in their ocean light spot removal tasks. We believe that the rain streak removal task is a little easier than the underwater image recovery task because there is little need to take into account lighting effects and background factors, and as clearly observed, PReNet [27] removes some of the ocean light spots by replacing them with dark patches. Although the Restormer [29] model alleviates the drawbacks of CNNs (i.e., limited perceptual field and maladaptation to the input content), it cannot be directly applied to design recovery tasks for underwater images considering its quadratically increased computational complexity with spatial resolution.
MPRNet [30] shows better performance than Restormer [29] and maxim [28] in underwater image recovery, and we believe that MRNet’s multi-stage progressive restoration is helpful in restoring underwater images. MPRNet first divides the feature map into four copies and then into two copies after restoration by U-Net. It then performs a residual network to complete the refinement. DGUNet [31] is an N-stage progressive generation network that also shows good results for underwater images with dark backgrounds. We attribute this success to its gradient strategy and the encoder and decoder components in each stage, as we can see in Table 2. UIR-Net achieves the best results at the values of PSNR SSIM and RMSE, in terms of UIQM and UICQE, UIR-Net approximate the best results.

4.6.3. Quantitative Comparisons of Underwater Enhancement on UIEBD-Snow Dataset

The results of the ocean light spot removal and image enhancement example are shown in Figure 9. We can see that UIR-Net achieves better results in both underwater image enhancement and underwater image restoration tasks.
As we can see from Table 3, UIR-Net achieves better performance and outperforms the other methods at PSNR SSIM and RMSE, in terms of UIQM and UICQE, UIR-Net approximate the best results. Specifically, the CWR [53] method based on contrast learning relies on the real image used and only takes the color correction problem into account. It performs poorly when confronted with fused light spot impurities in image enhancement, showing unnatural color and saturation and blurred image details. A multi-color space coding proposed by PUIE [52] proves to be effective in image enhancement alone, yet has little effect in the face of light spots, which leads use to believe that IFM [54] is relatively ineffective in image enhancement and restoration and does not improve the contrast, which is severely degraded by these colors. The objective function is also formulated to evaluate the perceptual quality based on global content, color, local texture and style. We believe that FUnIE-GAN [55] is superior to our method in image correction. The results of FUnIE-GAN are obviously different from those of other methods in terms of color, being more colorful and more consistent with the judgment of UIQM in terms of non-reference evaluation indicators, but FUnIE-GAN can only achieve image enhancement and cannot be used for image recovery. It can be seen that while most of the methods can perform contrast enhancement, some of the deficiencies cause severe visual defects such as unsatisfactory colors; our method, on the other hand, works well in all these cases, and our results look cleaner and more natural with fine-grained textures. MPRNet [30] and DGUNet [31] can achieve image enhancement to some extent, but they are not effective in image restoration. In addition, there is still a large portion of blur and missing textures in their outcomes. We believe that MPRNet and DGUNet focus on space in the task of domain-to-domain conversion and are unable to remove light spot occlusion, which is similar to image inpainting.

4.6.4. FLOPs and Params Comparisons with MPRNet and DGUNet

For a better quantitative comparison, we performed a fairer comparison with MPRNet [30] and DGUNet [31] in terms of FLOPs and Params, i.e., time complexity and space complexity, respectively. According to Table 4, we can see that UIR-Net has one-tenth the number parameters of MPRNet in terms of FLOPs, and there is not much difference in the amount of data in Params. We thus measured the training time again. We set the epoch to 200, the batch-size to 8 and the patch size to 128, and all methods adopted the warm up strategy. It can be seen that our training time is half that of MPRNet, and the inference speed is also better, so the operation method and convergence time of this paper are much better than MPRNet, which is very helpful for practical applications.

4.7. Ablation Study

4.7.1. Ablation Study of Channel Prior and Gradient Descent

As shown in Figure 10, while using only the gradient strategy, proximal gradient descent or the residual prior can obtain an improvement over the baseline, combining all of the above strategies can achieve the best performance, and the integration of these components can generate realistic results and restore details in the case of oceanic light spots and large grain impurities. The channel residual prior, based on the observation of the absorption rate of the red channel in a great many of underwater images, is suggested to be able to remain sensitive to low-illumination images.
As we can see in Table 5, after adding the residual prior, the PSNR, SSIM and RMSE were effectively improved to 33.656, 0.983 and 2.954, respectively, enhancing the significant texture features. When only adding the gradient strategy, PSNR climbed to 34.356 and SSIM rose to 0.981, proving the superiority of our model in underwater image restoration. The process of gradient descent is essentially the same as the operation of a neural network, except that when training a neural network, the samples and the parameters to be trained are known. Gradient descent has explicit variables so that it can rely on such variables to complete the iterations when training a neural network, instead of relying purely on samples and parameters. After adding the residual prior and the gradient strategy, PSNR, SSIM and RMSE reached 34.464 and 0.984 and 2.649, respectively.

4.7.2. Ablation Study of λ p e r c t u a l

We need to analyze the paramater of the Loss Function. As we can see from Figure 11, we can obtain that when λ p e r c t u a l equal 0.01, the information of the image is preserved more completely. Moreover we can see in Table 6 we can clearly deduce that when λ p e r c t u a l equal 0.01, all of the paramaters(PSNR SSIM and RMSE) are the best, so there the λ p e r c t u a l equal 0.01.

5. Conclusions

In this paper, we propose a simple and effective baseline named UIR-Net that integrates underwater image restoration and enhancement. For dark tones and ocean impurity occlusion, UIR-Net uses a channel residual prior to extract the channel of the image to be recovered as a prior, combined with a gradient strategy to reduce the number of parameters and training time to achieve a lightweight deployment. Our experimental results prove that UIR-Net can effectively improve images with both large ocean particle impurities and ocean light spots on these datasets (MSRB, MSIRB and UIEBD-Snow). In addition, our approach is able to achieve color correction on image enhancement, maintain the style and spatial texture of the content. In addition, the network in this paper outperforms many existing methods with less computational complexity and faster training speed, which is very useful for practical applications.

Author Contributions

Conceptualization, software and methodology, X.M. and X.Z.; data curation, X.M., J.W. and J.H.; writing—original draft preparation, X.M.; writing—review and editing, X.Y., Y.L. and X.W.; visualization, Y.L. and J.W.; supervision, project administration and funding acquisition, X.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 42276187 and 41876100) and the Fundamental Research Funds for the Central Universities (Grant No. 3072022FSC0401).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hambarde, P.; Murala, S.; Dhall, A. UW-GAN: Single-Image Depth Estimation and Image Enhancement for Underwater Images. IEEE Trans. Instrum. Meas. 2021, 70, 1–12. [Google Scholar] [CrossRef]
  2. Tao, Y.; Dong, L.; Xu, L.; Xu, W. Effective solution for underwater image enhancement. Opt. Express 2021, 29, 32412–32438. [Google Scholar] [CrossRef] [PubMed]
  3. Yuan, J.; Cao, W.; Cai, Z.; Su, B. An Underwater Image Vision Enhancement Algorithm Based on Contour Bougie Morphology. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8117–8128. [Google Scholar] [CrossRef]
  4. Jian, M.J.; Li, D.; Zhang, J.Q. Underwater Image Restoration Based on Non-Uniform Incident Light Imaging Model. Acta Opt. Sin. 2021, 41, 1501003. [Google Scholar]
  5. Zhao, X.; Jin, T.; Qu, S. Deriving inherent optical properties from background color and underwater image enhancement. Ocean. Eng. 2015, 94, 163–172. [Google Scholar] [CrossRef]
  6. Drews, P.; Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission estimation in underwater single images. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops (ICCVW), Sydney, NSW, Australia, 1–8 December 2013; IEEE: New York, NY, USA, 2013; pp. 825–830. [Google Scholar]
  7. Zhang, X.; Chen, F.; Wang, C.; Tao, M.; Jiang, G.P. Sienet: Siamese expansion network for image extrapolation. IEEE Signal Process. Lett. 2020, 27, 1590–1594. [Google Scholar] [CrossRef]
  8. Yan, S.; Zhang, X. PCNet: Partial convolution attention mechanism for image inpainting. Int. J. Comput. Appl. 2021, 44, 738–745. [Google Scholar] [CrossRef]
  9. Zhang, X.F.; Gu, C.C.; Zhu, S.Y. SpA-Former: Transformer image shadow detection and removal via spatial attention. arXiv 2022, arXiv:2206.10910. [Google Scholar]
  10. Zhang, X.; Wu, S.; Ding, H.; Li, Z. Image extrapolation based on multi-column convolutional attention network. In Proceedings of the 2020 IEEE 4th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chongqing, China, 12–14 June 2020; Volume 1, pp. 1938–1942. [Google Scholar]
  11. Shen, R.; Zhang, X.; Xiang, Y. AFFNet: Attention mechanism network based on fusion feature for image cloud removal. Int. J. Pattern Recognit. Artif. Intell. 2022, 36, 2254014. [Google Scholar] [CrossRef]
  12. Fabbri, C.; Islam, M.J.; Sattar, J. Enhancing underwater imagery using generative adversarial networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; pp. 7159–7165. [Google Scholar]
  13. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  14. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef] [Green Version]
  15. Jiang, Z.; Li, Z.; Yang, S.; Fan, X.; Liu, R. Target Oriented Perceptual Adversarial Fusion Network for Underwater Image Enhancement. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 6584–6598. [Google Scholar] [CrossRef]
  16. Jiao, Q.; Liu, M.; Li, P.; Dong, L.; Hui, M.; Kong, L.; Zhao, Y. Underwater image restoration via non-convex non-smooth variation and thermal exchange optimization. J. Mar. Sci. Eng. 2021, 9, 570. [Google Scholar] [CrossRef]
  17. Sato, Y.; Ueda, T.; Tanaka, Y. Marine Snow Removal Benchmarking Dataset. arXiv 2021, arXiv:2103.14249. [Google Scholar]
  18. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Qian, L.J.; Jin, H.H.; Fan, Z.G.; Zhuang, Z.J.; Gong, K.Q. Underwater Image Restoration Method Suppressing Interference of Light Source in Field of View. Acta Opt. Sin. 2021, 41, 1801001. [Google Scholar]
  20. Knauer, G.; Hebel, D.; Cipriano, F. Marine snow: Major site of primary production in coastal waters. Nature 1982, 300, 630–631. [Google Scholar] [CrossRef]
  21. Farhadifard, F.; Radolko, M.; von Lukas, U.F. Single image marine snow removal based on a supervised median filtering scheme. In Proceedings of the 12th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications, Porto, Portugal, 27 February–1 March 2017; pp. 280–287. [Google Scholar]
  22. Banerjee, S.; Sanyal, G.; Ghosh, S.; Ray, R.; Shome, S.N. Elimination of marine snow effect from underwater image-an adaptive probabilistic approach. In Proceedings of the IEEE Students’ Conference on Electrical, Electronics and Computer Science, Bhopal, India, 1–2 March 2014; pp. 1–4. [Google Scholar]
  23. Farhadifard, F.; Radolko, M.; von Lukas, U.F. Marine snow detection and removal: Underwater image restoration using background modeling. In Proceedings of the 25th International Conference in Central Europe on Computer Graphics, Visualization and Computer Visionin Co-Operation with Eurographics Association, Prague, Czech Republic, 29 May–2 June 2017; pp. 81–89. [Google Scholar]
  24. Cyganek, B.; Gongola, K. Real-time marine snow noise removal from underwater video sequences. J. Electron. Imaging 2018, 27, 043002. [Google Scholar] [CrossRef]
  25. Wang, Y.; Yu, X.; An, D.; Wei, Y. Underwater image enhancement and marine snow removal for fishery based on integrated dual-channel neural network. Comput. Electron. Agric. 2021, 186, 106182. [Google Scholar] [CrossRef]
  26. Fu, X.; Huang, J.; Zeng, D.; Huang, Y.; Ding, X.; Paisley, J. Removing rain from single images via a deep detail network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3855–3863. [Google Scholar]
  27. Ren, D.; Zuo, W.; Hu, Q.; Zhu, P.; Meng, D. Progressive image deraining networks: A better and simpler baseline. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3937–3946. [Google Scholar]
  28. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. Maxim: Multi-axis mlp for image processing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 5769–5780. [Google Scholar]
  29. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 5728–5739. [Google Scholar]
  30. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Multi-stage progressive image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 14821–14831. [Google Scholar]
  31. Mou, C.; Wang, Q.; Zhang, J. Deep Generalized Unfolding Networks for Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 17399–17410. [Google Scholar]
  32. Wei, G.Y.Z.; Chen, S.Y.; Liu, Y.T. Survey of underwater image enhancement and restoration algorithms. Appl. Res. Comput. 2021, 38, 2561–2569, 2589. [Google Scholar]
  33. Perez, J.; Attanasio, A.C.; Nechyporenko, N.; Sanz, P.J. A deep learning approach for underwater image enhancement. In Proceedings of the 2017 International Work-Conference on the Interplay Between Natural and Artificial Computation (IWINAC), Corunna, Spain, 19–23 June 2017; pp. 183–192. [Google Scholar]
  34. Wang, Y.; Zhang, J.; Cao, Y.; Wang, Z. A deep CNN method for underwater image enhancement. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 1382–1386. [Google Scholar]
  35. Li, C.; Anwar, S.; Porikli, F. Underwater scenes prior inspired deep underwater image and video enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
  36. Wang, N.; Zheng, H.; Zheng, B. Underwater Image Restoration via Maximum Attenuation Identification. IEEE Access 2017, 5, 18941–18952. [Google Scholar] [CrossRef]
  37. Peng, Y.-T.; Cao, K.; Cosman, P.C. Generalization of the Dark Channel Prior for Single Image Restoration. IEEE Trans. Image Process. 2018, 27, 2856–2868. [Google Scholar] [CrossRef] [PubMed]
  38. Carlevaris-Bianco, N.; Mohan, A.; Eustice, R.M. Initial results in underwater single image dehazing. In Proceedings of the MTS/IEEE Seattle, OCEANS 2010, Seattle, WA, USA, 20–23 September 2010; IEEE: New York, NY, USA, 2010; pp. 1–8. [Google Scholar]
  39. Song, W.; Wang, Y.; Huang, D.; Tjondronegoro, D. A rapid scene depth estimation model based on underwater light attenuation prior for underwater image restoration. In Proceedings of the 2018 Pacific Rim Conference on Multimedia (PCM), Hefei, China, 21–22 September 2018; Springer: Berlin, Germany, 2018; pp. 678–688. [Google Scholar]
  40. Peng, Y.-T.; Cosman, P.C. Underwater Image Restoration Based on Image Blurriness and Light Absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef]
  41. Zhang, B.; Jin, S.; Xia, Y.; Huang, Y.; Xiong, Z. Attention Mechanism Enhanced Kernel Prediction Networks for Denoising of Burst Images. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 2083–2087. [Google Scholar]
  42. Chen, D.; He, Z.; Cao, Y.; Yang, J.; Cao, Y.; Yang, M.Y.; Zhuang, Y. Deep neural network for fast and accurate single image super-resolution via channel-attention-based fusion of orientation-aware features. arXiv 2019, arXiv:1912.04016. [Google Scholar]
  43. Li, H.; Qiu, K.; Chen, L.; Mei, X.; Hong, L.; Tao, C. SCAttNet: Semantic segmentation network with spatial and channel attention mechanism for high-resolution remote sensing images. IEEE Geosci. Remote Sens. Lett. 2020, 18, 905–909. [Google Scholar] [CrossRef]
  44. Li, R.; Tan, R.T.; Cheong, L.F. Robust optical flow in rainy scenes. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Volume 2, pp. 288–304. [Google Scholar]
  45. Garg, K.; Nayar, S.K. Photometric model of a rain drop. In CMU Technical Report; Citeseer: Princeton, NJ, USA, 2003; Volume 4. [Google Scholar]
  46. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 694–711. [Google Scholar]
  47. Chen, W.T.; Fang, H.Y.; Ding, J.J.; Tsai, C.C.; Kuo, S.Y. JSTASR: Joint size and transparency-aware snow removal algorithm based on modified partial convolution and veiling effect removal. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; Springer: Cham, Switzerland, 2020; pp. 754–770. [Google Scholar]
  48. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  49. Yang, M.; Sowmya, A. An underwater color image quality evaluation metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef]
  50. Panetta, K.; Gao, C.; Agaian, S. Human-visual-system-inspired underwater image quality measures. IEEE J. Oceanic Eng. 2016, 41, 541–551. [Google Scholar] [CrossRef]
  51. Fu, X.; Huang, J.; Ding, X.; Liao, Y.; Paisley, J. Clearing the skies: A deep network architecture for single-image rain removal. IEEE Trans. Image Process. 2017, 26, 2944–2956. [Google Scholar] [CrossRef] [Green Version]
  52. Fu, Z.; Wang, W.; Huang, Y.; Ding, X.; Ma, K.-K. Uncertainty Inspired Underwater Image Enhancement. In Proceedings of the European Conference on Computer Vision(ECCV), Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 465–482. [Google Scholar]
  53. Han, J.; Shoeiby, M.; Malthus, T.; Botha, E.; Anstee, J.; Anwar, S.; Wei, R.; Petersson, L.; Armin, M.A. Single underwater image restoration by contrastive learning. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 2385–2388. [Google Scholar]
  54. Chen, X.; Zhang, P.; Quan, L.; Yi, C.; Lu, C. Underwater image enhancement based on deep learning and image formation model. arXiv 2021, arXiv:2101.00991. [Google Scholar]
  55. Islam, M.J.; Xia, Y.; Sattar, J. Fast underwater image enhancement for improved visual perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
Figure 1. Results of UIR-Net on MSRB and MSIRB datasets.
Figure 1. Results of UIR-Net on MSRB and MSIRB datasets.
Remotesensing 15 00039 g001
Figure 2. The structure of UIR-Net. It follows the encoder–decoder rule.
Figure 2. The structure of UIR-Net. It follows the encoder–decoder rule.
Remotesensing 15 00039 g002
Figure 3. The structure of gradient strategy.
Figure 3. The structure of gradient strategy.
Remotesensing 15 00039 g003
Figure 4. Examples from MSRB, MSIRB and UIEBD-Snow.
Figure 4. Examples from MSRB, MSIRB and UIEBD-Snow.
Remotesensing 15 00039 g004
Figure 5. Image restoration results of UIR-Net on MSRB and MSIRB datasets.
Figure 5. Image restoration results of UIR-Net on MSRB and MSIRB datasets.
Remotesensing 15 00039 g005
Figure 6. Image enhancement results of UIR-Net on UIEDB-Snow.
Figure 6. Image enhancement results of UIR-Net on UIEDB-Snow.
Remotesensing 15 00039 g006
Figure 7. Comparison results of UIR-Net with the seven other methods on the MSRB dataset.
Figure 7. Comparison results of UIR-Net with the seven other methods on the MSRB dataset.
Remotesensing 15 00039 g007
Figure 8. Comparing the results of UIR-Net with those of seven other methods on the MSIRB dataset.
Figure 8. Comparing the results of UIR-Net with those of seven other methods on the MSIRB dataset.
Remotesensing 15 00039 g008
Figure 9. Comparison results of UIR-Net with four other methods on UIBED-Snow. The underwater color images range from 1 to 4, which represent the difficulty of enhanced color correction: 1 represents the smallest color difference between the image to be corrected and the original image, and 5 represents the largest color difference between the image to be corrected and the original image.
Figure 9. Comparison results of UIR-Net with four other methods on UIBED-Snow. The underwater color images range from 1 to 4, which represent the difficulty of enhanced color correction: 1 represents the smallest color difference between the image to be corrected and the original image, and 5 represents the largest color difference between the image to be corrected and the original image.
Remotesensing 15 00039 g009
Figure 10. Ablation study of UIR-Net on the MSIRB dataset.
Figure 10. Ablation study of UIR-Net on the MSIRB dataset.
Remotesensing 15 00039 g010
Figure 11. Ablation study of λ p e r c t u a l on UIEBD-Snow dataset.
Figure 11. Ablation study of λ p e r c t u a l on UIEBD-Snow dataset.
Remotesensing 15 00039 g011
Table 1. Comparison results of UIR-Net with other SOTA methods on the MSRB dataset.
Table 1. Comparison results of UIR-Net with other SOTA methods on the MSRB dataset.
MethodPSNRSSIMRMSEUIQMUICQE
Nature image restoration
AirNet [51]24.3590.9546.7453.4180.549
DB-ResNet [26]23.2580.9209.0603.5310.553
PReNet [27]21.8370.89913.9323.7220.524
Maxim [28]24.5780.9317.8363.6060.553
Restormer [29]24.4110.9546.7833.9770.550
DGUNet [31]33.2920.9773.4083.7550.567
MPRNet [30]35.3220.9842.6593.7150.566
Our34.4640.9842.6493.8770.576
Table 2. Comparison results of UIR-Net with other SOTA methods on the MSIRB dataset.
Table 2. Comparison results of UIR-Net with other SOTA methods on the MSIRB dataset.
MethodPSNRSSIMRMSEUIQMUICQE
Nature image restoration
AirNet [51]20.5360.9229.8783.7610.575
DB-ResNet [26]20.5060.89710.7433.7580.574
PReNet [27]20.3100.88814.2374.0750.550
Maxim [28]23.5990.9248.3163.8320.565
Restormer [29]20.3960.9279.8173.8100.574
DGUNet [31]33.7850.9842.5843.9110.562
MPRNet [30]33.5610.9852.5923.7250.554
Our34.5130.9862.3913.8370.565
Table 3. Comparison results of UIR-Net with other SOTA methods on the UIEBD-Snow dataset.
Table 3. Comparison results of UIR-Net with other SOTA methods on the UIEBD-Snow dataset.
MethodPSNRSSIMRMSEUIQMUICQE
Underwater image enhancement
PUIE [52]16.9260.72318.0303.5940.606
CWR [53]15.3740.55623.2764.6200.507
IMF [54]17.0060.70718.9043.3540.633
FUnIE-GAN [55]15.6190.48624.7625.1060.596
Nature image restoration
MPRNet [30]20.0270.77515.0483.4000.598
DGUNet [31]20.7110.79514.0683.3630.597
Our21.2000.80713.1423.6100.596
Table 4. FLOPs and Params comparisons.
Table 4. FLOPs and Params comparisons.
MethodFLOPsTotal ParamsTraining Time
DGUNet [31] 8.4278 × 10 11 12,176,11994,600 s
MPRNet [30] 2.1946 × 10 12 3,637,24941,400 s
UIR-Net 2.4120 × 10 11 3,632,74022,000 s
Table 5. Ablation study of UIR-Net on the MSRB and MSIRB datasets.
Table 5. Ablation study of UIR-Net on the MSRB and MSIRB datasets.
residual priorproximal gradient descentPSNRSSIMRMSE
33.3190.9812.953
33.3190.9812.953
34.3560.9812.931
34.4640.9842.649
residual priorproximal gradient descentPSNRSSIMRMSE
32.4770.9822.912
33.2950.9832.649
34.1210.9842.765
34.5130.9862.391
Table 6. Ablation results of paramater of λ p e r c t u a l on the UIEBD-Snow dataset.
Table 6. Ablation results of paramater of λ p e r c t u a l on the UIEBD-Snow dataset.
ParamaterPSNRSSIMRMSE
λ p e r c t u a l = 0.120.2030.78114.030
λ p e r c t u a l = 0.0519.3710.79113.301
λ p e r c t u a l = 0.0220.0390.79713.904
λ p e r c t u a l = 0.0121.2000.80713.142
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mei, X.; Ye, X.; Zhang, X.; Liu, Y.; Wang, J.; Hou, J.; Wang, X. UIR-Net: A Simple and Effective Baseline for Underwater Image Restoration and Enhancement. Remote Sens. 2023, 15, 39. https://doi.org/10.3390/rs15010039

AMA Style

Mei X, Ye X, Zhang X, Liu Y, Wang J, Hou J, Wang X. UIR-Net: A Simple and Effective Baseline for Underwater Image Restoration and Enhancement. Remote Sensing. 2023; 15(1):39. https://doi.org/10.3390/rs15010039

Chicago/Turabian Style

Mei, Xinkui, Xiufen Ye, Xiaofeng Zhang, Yusong Liu, Junting Wang, Jun Hou, and Xuli Wang. 2023. "UIR-Net: A Simple and Effective Baseline for Underwater Image Restoration and Enhancement" Remote Sensing 15, no. 1: 39. https://doi.org/10.3390/rs15010039

APA Style

Mei, X., Ye, X., Zhang, X., Liu, Y., Wang, J., Hou, J., & Wang, X. (2023). UIR-Net: A Simple and Effective Baseline for Underwater Image Restoration and Enhancement. Remote Sensing, 15(1), 39. https://doi.org/10.3390/rs15010039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop