Next Article in Journal
Numerical Study on the Performance of an OWC under Breaking and Non-Breaking Waves
Next Article in Special Issue
Breakthrough Underwater Physical Environment Limitations on Optical Information Representations: An Overview and Suggestions
Previous Article in Journal
Structural Analysis of Deck Reinforcement on Composite Yacht for Crane Installation
Previous Article in Special Issue
Deblurring Network for UV Images of Colorless Chemicals Floating on Sea Surface
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Underwater Image Restoration and Depth Estimation with Light Field Images

1
State Key Laboratory of Advanced Design and Manufacturing for the Vehicle Body, Hunan University, Changsha 410082, China
2
School of Electrical and Data Engineering, University of Technology Sydney, Ultimo, NSW 2007, Australia
3
School of Smart Marine Science and Engineering, Fujian University of Technology, Fuzhou 350118, China
4
Fujian Provincial Key Laboratory of Marine Smart Equipment, Fuzhou 350118, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(6), 935; https://doi.org/10.3390/jmse12060935
Submission received: 27 April 2024 / Revised: 28 May 2024 / Accepted: 31 May 2024 / Published: 2 June 2024
(This article belongs to the Special Issue Underwater Engineering and Image Processing)

Abstract

:
Methods based on light field information have shown promising results in depth estimation and underwater image restoration. However, improvements are still needed in terms of depth estimation accuracy and image restoration quality. Previous work on underwater image restoration employed an image formation model (IFM) that overlooked the effects of light attenuation and scattering coefficients in underwater environments, leading to unavoidable color deviation and distortion in the restored images. Additionally, the high blurriness and associated distortions in underwater images make depth information extraction and estimation very challenging. In this paper, we refine the light propagation model and propose a method to estimate the attenuation and backscattering coefficients of the underwater IFM. We simplify these coefficients into distance-related functions and design a relationship between distance and the darkest channel to estimate the water coefficients, effectively suppressing color deviation and distortion in the restoration results. Furthermore, to increase the accuracy of depth estimation, we propose using blur cues to construct a cost for refocusing in the depth direction, reducing the impact of high signal-to-noise ratio environments on depth information extraction, and effectively enhancing the accuracy and robustness of depth estimation. Finally, experimental comparisons show that our method achieves more accurate depth estimation and image restoration closer to real scenes compared to state-of-the-art methods.

1. Introduction

Underwater image restoration and depth estimation are crucial research areas that have broad applications in underwater scene reconstruction, heritage conservation, and seabed surveying, among others. However, underwater optical imaging is often compromised by water quality and lighting conditions, which can lead to color distortion and image blurring. To address these challenges, numerous solutions have been explored, including the use of polarization [1], light fields (LF), and stereoscopic imaging techniques [2]. Among these, light field applications in underwater imaging have garnered attention due to the micro-lens array of light field cameras, which can capture sufficient spatial and angular information in a single shot. This capability makes it possible to extract accurate depth information from a single viewpoint, significantly enhancing the quality of restoration in underwater environments.
A considerable amount of research has utilized light field data to enhance underwater image restoration [3,4,5,6]. These algorithms typically rely on the Image Formation Model (IFM) proposed by Narasimhan and Nayar [7], which assumes constant and identical attenuation and scattering coefficients. However, in real underwater environments, the attenuation coefficient varies with distance and differs from the scattering coefficient, leading to inaccuracies in the IFM that result in noticeable color biases in the restored images.
In the field of underwater depth estimation, the clarity of target scenes is often compromised by backscatter from dense water bodies, leading to blurred images. This effect renders most traditional refocusing algorithms that rely on correlation cues and methods using the dark channel prior for long-range depth estimation ineffective, especially in areas with significant backscatter noise. Recently, researchers like Lu [5] and Huang [8] have proposed using deep neural networks to estimate underwater image depth. However, like all methods that employ deep learning for underwater depth estimation, they face challenges due to the scarcity of adequate paired training data. Consequently, these methods perform poorly in underwater environments where no training data pairs are available.
In this paper, we introduce a method for estimating intrinsic parameters of an underwater image formation model and propose a depth estimation method for underwater light field images based on blur cues. Inspired by [9], we use distance as a cue to estimate the inherent parameters of the water body. However, compared to [9], our method further simplifies the parameters required by the model while maintaining accuracy. Furthermore, leveraging the positive correlation between underwater image depth and blurriness, we propose using blur cues as a cost component for light field depth refocusing, effectively enhancing the accuracy of depth estimation in underwater scenes. As illustrated in Figure 1, our approach achieves higher accuracy in depth estimation and improves image restoration outcomes in underwater settings. The contributions of this paper are outlined as follows:
  • We have introduced an underwater light propagation model for image restoration and developed a method to estimate water body attenuation parameters and backscatter;
  • We proposed a method for underwater image depth estimation that utilizes the relationship between blurriness and scene depth as one of the clues for estimating the depth of underwater light field images, thereby improving the accuracy of underwater scene depth estimation;
  • In an experimental water tank environment, we demonstrated through extensive experimental data that our method achieves higher depth estimation accuracy and better restoration effects compared to previous methods.

2. Related Work

The problem of underwater image degradation has been approached from various research angles, such as hardware-based and software-based restoration techniques. Image distortions caused by forward attenuation are generally addressed using software approaches, including the establishment of attenuation functions [10,11,12], image enhancement [13], or deep learning methods [8]. However, software solutions face limitations in handling information loss caused by scattering noise, and deep learning approaches struggle to acquire high-quality data pairs for training.
To mitigate backscatter noise, Schechner and Karpel [1] utilized the polarization properties of light in the medium, capturing images at different angles with polarizers to retrieve effective image information. Hitam et al. [14] explored range-gated imaging, using pulsed lasers and fast-gated cameras to precisely control the imaging process and reduce backscatter. Additionally, Roser and colleagues [2] employed stereoscopic imaging techniques to determine object distances by comparing the disparity between images captured from different angles. These methods have achieved notable results, but harsh underwater conditions can significantly reduce the imaging quality, thus impacting the performance of traditional sensors and environmental sensing accuracy.
In recent years, as light field cameras have entered the consumer market, the field of light field imaging has garnered increased attention. Compared to traditional imaging, light field cameras can capture not only the intensity of light but also its angular information, allowing for far superior estimation accuracy in image depth. Tao and others [15] combined digital refocusing blur cues with corresponding cues to estimate scene depth. Georgiev [16] further applied this to super-resolution imaging.
However, there has been limited research on using light field cameras for underwater image restoration. Tian [3] and Lu [4] were the first to attempt applying light field methods to underwater image restoration, combining the dark channel prior (DCP) [17] method and digital refocusing cues to estimate depth maps and restore images, which showed better robustness and restoration effects compared to monocular image strategies. Skinner [18] applied 2D defogging methods to each sub-aperture image and processed them using guided image filtering, resulting in 4D defogged underwater light field images. Feng [19] established an underwater imaging system based on Mie scattering theory for underwater 3D reconstruction and image restoration. Additionally, Lu [20] introduced depth convolutional neural fields in image restoration tasks to address the desaturation problem in light field images. To tackle the challenge of obtaining underwater image data, Ye [21] proposed a deep learning-based method to transform images taken in air into underwater images. However, due to limited training data, the effectiveness of deep learning methods was not satisfactory. While these methods have achieved good results, they have neglected the impact of light imaging model parameters in the restoration process. Moreover, the accuracy of light field depth estimation methods needs improvement due to the noise caused by backscattering.
Despite the superior performance of light field cameras compared to traditional cameras, there remain significant limitations. In this paper, we employ blur cues from light field refocusing combined with traditional priors to achieve more accurate depth maps. We then introduce a refined underwater optical imaging model and estimate more precise parameters for light attenuation and scattering to improve the quality of underwater image restoration.
The remainder of this paper is organized as follows. In Section 3, we describe the proposed method. In Section 4, we set up the experiments and conduct extensive performance comparisons of the results. In Section 5, we provide some discussions and conclusions. Specifically, in Section 3.1, we describe the image restoration model and its related parameters, and the subsequent subsection details the solving process for these parameters. In Section 3.2, we explain the methods and processes for solving the scattering and attenuation coefficients in the model. In Section 3.3, we present our proposed underwater light field depth estimation method. In Section 3.4, we describe the refocusing restoration process for light field images.

3. Method

3.1. Underwater Image Formation Model

Figure 2 depicts the underwater image formation model, where the camera captures underwater image information comprising attenuated direct signals and scattered signals from the water body. The imaging model is represented as follows:
I c ( x ) = D c ( x ) + B c ( x ) ,
where c { R , G , B } represents the color channels, I c is the distorted image captured by the camera, D c is the target information attenuated during propagation through the water, and B c is the scattering signal from light reflected off particles suspended in the water, which degrades and blurs the image.
The extended form of Equation (1) is given as follows [9]:
I c ( x ) = J c ( x ) e β c D z + B c 1 e β c B z ,
where z denotes the distance from the camera to the target, B c represents the external input light source, J c represents the undegraded scene image, β c D is the beam attenuation coefficient, and β c B is the backscattering coefficient of the water body.
The attenuation coefficient β c D in the attenuation term D c ( x ) depends on wavelength λ , depth d, and distance z. Similarly, the scattering coefficient β c B is related to the wavelength and the background light B c . These coefficients can be expressed as follows [22]:
β c D = ln λ 1 λ 2 S c ( λ ) ρ ( λ ) E ( d , λ ) e β ( λ ) z d λ λ 1 λ 2 S c ( λ ) ρ ( λ ) E ( d , λ ) e β ( λ ) ( z + Δ z ) d λ / Δ z β c B = ln 1 λ 1 λ 2 S c ( λ ) B ( λ ) ( 1 e β ( λ ) z ) d λ λ 1 λ 2 B ( λ ) S c ( λ ) d λ / z ,
where ρ represents the reflectance of the object’s surface, E ( d , λ ) is the intensity of the incident light at depth d, S c is the spectral response of the camera, β is the physical attenuation coefficient, and λ 1 and λ 2 represent the visible wavelength range (400 to 700 nm).
The background light B c is the sum of the light received at the position where the distance z , represented by the following:
B c = λ 1 λ 2 B ( λ ) S c ( λ ) d λ .

3.2. Coefficients Estimation

According to Equation (2), to effectively restore underwater images, it is essential to determine or estimate the parameters associated with light attenuation and scattering during propagation. As mentioned in Equation (3), ( β c D ) and ( β c B ) depend on the parameters { z , ρ , E ( d ) , S c , β } and { E ( d ) , S c , β } , respectively. The parameters β , E ( d ) (defined by the optical water type), the depth d at which photos are taken, the reflectance ρ of each target in the scene, and the camera response S c are considered constants once determined at the time of capture. These parameters are seldom known precisely at the time an underwater photo is taken. β c D is significantly influenced by the distance z, and β c B is highly correlated with the type of water body and illumination. Here we detail the methods for estimating these coefficients.

3.2.1. β c B and B c Estimation

To estimate the backscattering parameters, it is crucial to isolate other influencing factors. Inspired by He’s [23] dark channel prior method, where the minimum RGB values in a hazy image’s dark channel represent the relative distance between the camera and the target and the maximum values indicate the background light region, our approach differs in that we know the target distance range rather than estimating it.
From Equations (1) and (2):
B c ( x ) = J c ( x ) e β c D z + I c ( x ) .
It is evident that the backscatter part is related to distance z and reflectance ρ , and tends to the captured distorted image as distance increases or target reflectance decreases. Therefore, for reflectance ρ 0 or E d 0 (in shadow regions), the captured target pixel intensity tends toward the backscatter pixel intensity B c ( x ) I c ( x ) . This is similar to the DCP method, but here we look for the lowest 1% of RGB triplets in the image area represented as Ω where ρ c 0 or E d 0 , hence B ^ c ( Ω ) I c ( Ω ) . Concurrently, to obtain a more accurate scattering parameter relative to distance, we take the target distance as a known parameter. Thus, the backscatter in the region Ω can be further expressed as follows:
B ^ c B c ( 1 e β c B z i ) + J c e β c D z i ,
where z i is the known distance variable and J c e β c D z i represents a minor residual amount, retained due to dense scattering conditions in water. Using nonlinear least squares, we then estimate parameters B c , β c B , J c , β c D . Note that since the forward signal is minimal, β c D can be temporarily considered constant.

3.2.2. β c D ( z ) Estimation

According to [9], β c D is strongly related to distance z, and its dependency on any range z is described using a dual-exponential form:
β c D ( z ) = α 1 exp ( α 2 z ) + α 3 exp ( α 4 z ) ,
where α 1 , α 2 , α 3 , and α 4 are constants.
Assuming B c and β c B have been obtained, B c can be removed from the original image I c , and β c D ( z ) can be estimated by processing the direct signal D c . According to Equations (2) and (3), the restored image can be expressed as follows:
J ^ c ( x ) = ( I c ( x ) B c ) e β c D ( z ) · z .
In our experimental setup, where absolute distances are not extensive, we assume the corresponding image captured in clear water undergoes minimal scattering and attenuation, using it as a reference image for restoration effects set as J ^ c ( x ) . Rewriting the formula, we obtain the following:
z ˜ = ln J ^ c ( x ) I c ( x ) B c / e β c D ( z ) .
By minimizing this expression, we obtain discrete β c D data related to range z.
min β c D ( z ) z ˜ z
Combining with Equations (7) and (9), we derive constants α 1 , α 2 , α 3 , α 4 for β c D .

3.3. Depth Estimation

In densely scattering media such as water, the random scattering of light during propagation leads to restored images with relatively low signal-to-noise ratios, causing traditional defocus and correspondence cues to perform poorly. To address this, we propose using blur cues and incorporating a depth threshold method to reduce the influence of noise on depth estimation, thereby improving accuracy. Below, we detail our approach.

3.3.1. Construction of Blur Clue Cost

We utilize a parallel plane parameterization to represent the four-dimensional light field L F ( x , y , u , v ) , where ( x , y ) and ( u , v ) correspond to spatial and angular coordinates, respectively. Following Ng et al. [24], we segment the light field data at different depths as follows:
L α ( x , y , u , v ) = L F x + u 1 1 α , y + v 1 1 α , u , v ,
where α represents the current focus depth ratio and L α is the refocused light field image at the depth ratio α , with the central view located at coordinates ( u , v ) = ( 0 , 0 ) .
Considering that the further the distance of a target pixel due to attenuation, the greater its blurriness, and that the blurriness of a target pixel after light field refocusing is inversely proportional to its true distance, we choose to use blurriness as a cost clue for depth estimation. We employ a local frequency analysis method using the Discrete Cosine Transform (DCT) [25] to analyze the characteristics of the area around a pixel. By evaluating the absence of high-frequency components, we can quantify the image’s blurriness. The DCT representation for the area Ω around a clue point ( x , y ) is given as follows:
T ( p , q ) = c ( p ) c ( q ) x = 0 N 1 y = 0 N 1 Ω ( x , y ) cos π ( 2 x + 1 ) p 2 N cos π ( 2 y + 1 ) q 2 N ,
where p , q are frequency variables ranging from 0 to N 1 , Ω ( x , y ) are the pixel values in an N × N area, and c ( p ) , c ( q ) are normalization factors defined as follows:
c ( p ) = 1 N if p = 0 2 N otherwise .
The blurriness at that location is then expressed as follows:
M α ( x , y ) = p = 1 N 1 q = 1 N 1 | T ( p , q ) | 2 .
For the refocused light field image at depth ratio α , the cost clue is represented as follows:
C α M ( x , y ) = 1 N u , v M α ( x , y , u , v ) .
This method systematically utilizes the inherent blurriness of the underwater images caused by scattering to establish a robust depth estimation framework. By integrating blur as a quantitative measure into the cost function for depth estimation, we can enhance the accuracy of light field-based depth estimation in challenging underwater conditions.

3.3.2. Single Image Preprocessing

Given the initial high blurriness of underwater images, it is crucial to obtain more texture detail information through preprocessing before depth estimation. As discussed in Section 3.2, having acquired the parameters of the light propagation model, we now only need an initial depth estimate to retrieve more detailed textures from underwater images.
Inspired by the methods of Peng [11] and Carlevaris-Bianco [26], we combine the characteristics of these two approaches for our image preprocessing depth estimation.
First, considering the relationship between imaging distance and blurriness in the captured images, Peng [11] introduced an image formation model based on blurriness prior, where the pixel blurriness related to depth distance is represented as follows:
F ( x ) = max y Ω ( x ) 1 n i = 1 n I grey ( y ) G ( I grey ( y ) ) ,
where I grey is the grayscale version of the input image I c , G is a spatial filter of size h × h with variance σ 2 , set as h = 2 i + n + 1 , with n set to 4. Ω ( x ) is a local patch centered at x.
Next, we estimate depth distances based on the absorption rate differences of light at various wavelengths in the medium, where attenuation of red wavelengths is greater than that of blue and green wavelengths. Carlevaris-Bianco [26] suggested using the difference between the maximum intensities of the red channel and the green and blue channels as depth cues, represented as follows:
Q ( x ) = max y Ω ( x ) I r ( y ) max y Ω ( x ) { I g ( y ) , I b ( y ) } .
We combine the distance cues from Equations (16) and (17) to estimate an initial depth map, expressed as follows:
z ˜ F = 1 N norm ( F ( x ) ) z ˜ Q = 1 N norm ( Q ( x ) ) ,
where N norm is a stretching normalization function. Then, depending on the input’s average red channel value, we choose between light absorption-based or image blurriness-based methods to estimate the underwater scene depth:
z ˜ init = d ˜ F + ( 1 θ ) d ˜ Q ,
where θ = S ( avg ( I r ) , 0.1 ) , a v g is the mean function, and S is the sigmoid function defined as follows:
S ( a , b ) = 1 1 + e s ( a b ) ,
where s controls the slope of the activation function, thereby controlling the amplitude of the output change when the input values have slight differences. We experimented with different values and found that the model performs best when s = 32 .
Finally, based on Equation (19), we can obtain the initial depth estimation map. By substituting the depth results into the process described in Section 3.1, we can obtain the preliminary restored underwater image.

3.3.3. Depth Cue Fusion and Depth Estimation

After obtaining the blur-based depth cues to enhance the robustness of depth estimation, we also introduce Tao’s [15] defocus cost as a supplementary metric. The spatial point defocus cost response is represented as follows:
C α F ( x , y ) = 1 | W F | x , y W F Δ L ¯ α ( x , y ) ,
where W F is the window size around the current pixel, Δ is the Laplacian operator in both horizontal and vertical directions, and L ¯ α ( x , y ) is the mean of all angular positions for that pixel, expressed as follows:
L ¯ α ( x , y ) = 1 N u , v L α ( x , y , u , v ) .
Combining the depth cost clues from Equations (15) and (21), the depth relationship is finally established as follows:
z α ( x , y ) = arg min α C α F ( x , y ) + C α M ( x , y ) .
By inputting the restored light field image into this formula, we obtain the depth estimation map for the underwater light field image. This comprehensive approach effectively merges various depth cues to produce a more accurate and robust estimation of underwater depths.

3.4. Underwater LF Image Restoration

Although attenuation and scattering effects have been removed in light field imaging, the random scattering of light by particles in water often still results in noise in the restored images, especially in waters with high turbidity. To address this, we take into account that the noise patterns and their distributions formed at the same object point across different sub-apertures vary. By averaging multiple light field views through a light field refocusing algorithm, we can further suppress noise and reduce the impact of backscatter noise.

4. Experimental Results

This section introduces the experimental setup, displays results from our experimental apparatus, compares them against various state-of-the-art algorithms, and discusses the limitations of our approach.

4.1. Experimental Methodology

In this study, we conducted experiments in a dark environment using a water tank with dimensions of 0.7 × 0.4 × 0.4 m. All sides of the tank, except the shooting face, were covered with opaque black film to prevent external light reflections. We used white LED lights as the light source. The Lytro Illum camera was positioned 2 cm away from the tank. Figure 3 illustrates this experimental setup. Note that the selection of the tank size was based on the physical space constraints of the laboratory and the practical feasibility of operations. Ideally, a larger water tank would better simulate the light propagation paths in real environments. However, due to the limitations of our laboratory conditions, we could only use the available size of the water tank for our experiments.
In the experiment, we gradually added milk to the water tank, totaling 4 mL. This incremental addition method allowed us to precisely control the turbidity of the water, thereby simulating different levels of scattering effects.
We chose milk as the scattering medium because it can evenly disperse in water, providing consistent optical scattering effects. Although other studies have used substances like red wine, milk is more suitable for simulating suspended particles and turbidity in water. Additionally, milk is easy to control and measure, and its optical properties better meet the requirements of our experiment.

4.2. Results Comparisons

In this experiment, we evaluated the performance of our proposed method against the latest techniques in image restoration and depth estimation. We selected prior-based single image methods for depth estimation and restoration [11,27], light field (LF)-based methods [3], and deep learning-based methods [8,28] for comparison. It should be noted that the LF-based methods were not optimized specifically for this task during replication, leading to discrepancies between their replication results and practical outcomes. Note that the methods selected above all have publicly available source code or authors willing to provide the code. This allows us to accurately reproduce the results of these methods and make fair comparisons.
The original image data used in Figure 4 and Figure 5, labeled LF1 to LF5, were collected in a medium turbidity environment created by adding 4ml of milk. To verify robustness to distance, the targets in these five scenes were randomly placed at different positions in the water tank, with designs including both single target and varying positions of front and back targets.
Figure 4 presents a comparison of depth estimation results between our method and other methods, corresponding to the original images in column (a). In columns (b–c) of Figure 4, methods based on blur and light attenuation priors show relatively stable performance but are significantly affected by high blur in the background regions. In column (d), the light field-based depth estimation method performs well in blurred background regions, second only to our method. However, it exhibits poor robustness in the halo regions caused by light scattering around the target. Column (e) shows good detail in the target area but performs poorly in noisy regions. Column (f) has the worst performance due to the limited generalization capability caused by the underwater image data constraints. Column (g) of our method exhibits more target depth details and smoother, more accurate results in noisy background regions. From a subjective evaluation perspective, our method exhibits certain advantages over other methods in terms of depth estimation accuracy in areas with background light noise interference and target details.
Figure 5 provides a comparison of restoration effects between our method and other methods based on the original images shown in column (a). Columns (b–c) in Figure 5 show that prior-based methods result in noticeable red color casts in the restored images. Additionally, due to incorrect depth map estimation, the target blur remains significant. Columns (d) and (f) display a certain degree of blue-green cast, and the image in column (f) is relatively blurred. Column (e) performs well in the background region but still exhibits some red color cast in the target area. Column (g) of our method effectively removes backscatter from both the target and background, reducing noise. Additionally, light field refocusing brings about better clarity, and the colors are more aligned with human visual perception. Therefore, from a subjective evaluation perspective, our method demonstrates better restoration effects compared to the other methods.
Underwater images have various objective evaluation methods. The underwater color image quality evaluation (UCIQE) [29] is the most widely used set of criteria, providing a no-reference metric specifically for assessing underwater image quality by analyzing color, saturation, and contrast. This standard addresses image quality degradation caused by light attenuation, scattering, and color shifts. A higher score indicates better image quality. Table 1 presents the evaluation scores for Figure 5. It can be observed that the image restoration results of our method achieve better overall performance across the five different image scenes compared to other methods.
Furthermore, We also tested the robustness of our method in environments with varying degrees of turbidity. We evaluated the restoration performance of our approach in waters of different turbidity levels using the underwater color image quality evaluation (UCIQE) and the structural similarity index (SSIM) [30] as quality metrics. SSIM, which mimics the human visual system to measure the structural similarity between two images, is a widely used reference-based index. We used images captured in clear water as the real image reference. The restoration results, as shown in Figure 6 and Figure 7, demonstrate that our method outperforms others across various turbidity levels. However, our method has limitations; when the water reaches a certain level of turbidity, the quality of our restored images rapidly deteriorates.
In summary, we subjectively compared the latest methods with our proposed approach in terms of depth maps and image restoration. Our method demonstrates better accuracy and image texture details in the depth maps compared to other methods. Additionally, due to the full-focus characteristics of the light field, our method significantly outperformed single image approaches in controlling noise in the final restored images. Objectively, we used the reference-free UCIQE to compare the restoration effects of different models and the reference-based SSIM to assess the restoration performance under various turbidity levels. The results conclusively showed that our method performed better than existing methods in both respects.

5. Discussion and Conclusions

In this paper, we proposed a method for underwater depth estimation and image restoration in near-field illumination environments. Our contributions include introducing a novel approach that utilizes blurriness and backscatter as depth estimation cues, proposing an image restoration scheme based on state-of-the-art underwater imaging models in near-field illumination and dense scattering environments, and a method to improve the accuracy of light field depth estimation in dense noisy environments. Through comparisons with existing methods, our approach demonstrates improved accuracy in depth estimation, image restoration, and stability compared to color prior-based methods and several state-of-the-art light field-based restoration methods. However, our results also have some errors and limitations, such as inaccuracies in depth estimation in shadow regions and a significant decline in restoration performance in highly turbid water. Future research will further optimize our method and validate its effectiveness under more experimental conditions. Our next step is to establish a database of underwater light field images to analyze the impact of different lighting conditions and water parameters on imaging.

Author Contributions

Conceptualization, B.X. and X.G.; methodology, B.X. and H.H.; software, B.X. and X.G.; validation, B.X. and H.H.; formal analysis, B.X. and X.G.; investigation, B.X. and H.H.; resources, X.G.; data curation, H.H.; writing—original draft preparation, B.X. and X.G.; writing—review & editing, H.H. and X.G.; visualization, B.X. and H.H.; supervision, H.H.; project administration, H.H. and X.G. All authors have read and agreed to the published version of the manuscript.

Funding

The support provided by China Scholarship Council (CSC) during a visit of Bo Xiao to University of Technology Sydney is acknowledged. This research by authors Xiujing Gao and Hongwu Huang was supported by Fujian Provincial Department of Science and Technology Major Special Projects (2023HZ025003), Key Scientific and Technological Innovation Projects of Fujian Province (2022G02008), Education and Scientific Research Project of Fujian Provincial Department of Finance (GY-Z220233), and Fujian University of Technology (GY-Z23027).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to the protection of intellectual property.

Acknowledgments

Thanks to my supervisor Stuart Perry for his careful guidance and valuable advice on my research. Additionally, we once again thank Stuart for the invaluable support we received from him during the various phases of the preparation of this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Schechner, Y.; Karpel, N. Clear underwater vision. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; Volume 1, p. I. [Google Scholar] [CrossRef]
  2. Roser, M.; Dunbabin, M.; Geiger, A. Simultaneous underwater visibility assessment, enhancement and improved stereo. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 3840–3847. [Google Scholar] [CrossRef]
  3. Tian, J.; Murez, Z.; Cui, T.; Zhang, Z.; Kriegman, D.; Ramamoorthi, R. Depth and Image Restoration from Light Field in a Scattering Medium. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2420–2429. [Google Scholar] [CrossRef]
  4. Lu, H.; Li, Y.; Uemura, T.; Kim, H.; Serikawa, S. Low illumination underwater light field images reconstruction using deep convolutional neural networks. Future Gener. Comput. Syst. 2018, 82, 142–148. [Google Scholar] [CrossRef]
  5. Lu, H.; Li, Y.; Zhang, Y.; Chen, M.; Serikawa, S.; Kim, H. Underwater Optical Image Processing: A Comprehensive Review. Mob. Netw. Appl. 2017, 22, 1204–1211. [Google Scholar] [CrossRef]
  6. Tian, Y.; Liu, B.; Su, X.; Wang, L.; Li, K. Underwater Imaging Based on LF and Polarization. IEEE Photonics J. 2019, 11, 1–9. [Google Scholar] [CrossRef]
  7. Narasimhan, S.; Nayar, S. Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 713–724. [Google Scholar] [CrossRef]
  8. Huang, S.; Wang, K.; Liu, H.; Chen, J.; Li, Y. Contrastive Semi-Supervised Learning for Underwater Image Restoration via Reliable Bank. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 18145–18155. [Google Scholar]
  9. Akkaynak, D.; Treibitz, T. A Revised Underwater Image Formation Model. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6723–6732. [Google Scholar] [CrossRef]
  10. Song, W.; Wang, Y.; Huang, D.; Tjondronegoro, D. A Rapid Scene Depth Estimation Model Based on Underwater Light Attenuation Prior for Underwater Image Restoration. In Proceedings of the Advances in Multimedia Information Processing—PCM 2018; Lecture Notes in Computer Science; Hong, R., Cheng, W.H., Yamasaki, T., Wang, M., Ngo, C.W., Eds.; Springer: Cham, Switzerland, 2018; pp. 678–688. [Google Scholar] [CrossRef]
  11. Peng, Y.T.; Cosman, P.C. Underwater Image Restoration Based on Image Blurriness and Light Absorption. IEEE Trans. Image Process. 2017, 26, 1579–1594. [Google Scholar] [CrossRef]
  12. Drews, P.L.; Nascimento, E.R.; Botelho, S.S.; Montenegro Campos, M.F. Underwater Depth Estimation and Image Restoration Based on Single Images. IEEE Comput. Graph. Appl. 2016, 36, 24–35. [Google Scholar] [CrossRef]
  13. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2020, 29, 4376–4389. [Google Scholar] [CrossRef]
  14. Hitam, M.S.; Awalludin, E.A.; Jawahir Hj Wan Yussof, W.N.; Bachok, Z. Mixture contrast limited adaptive histogram equalization for underwater image enhancement. In Proceedings of the 2013 International Conference on Computer Applications Technology (ICCAT), Sousse, Tunisia, 20–22 January 2013; pp. 1–5. [Google Scholar] [CrossRef]
  15. Tao, M.W.; Hadap, S.; Malik, J.; Ramamoorthi, R. Depth from Combining Defocus and Correspondence Using Light-Field Cameras. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 673–680. [Google Scholar] [CrossRef]
  16. Georgiev, T.; Chunev, G.; Lumsdaine, A. Superresolution with the focused plenoptic camera. In Proceedings of the Computational Imaging IX; International Society for Optics and Photonics; Bouman, C.A., Pollak, I., Wolfe, P.J., Eds.; SPIE: Bellingham, WA, USA, 2011; Volume 7873, p. 78730X. [Google Scholar] [CrossRef]
  17. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 1956–1963. [Google Scholar] [CrossRef]
  18. Skinner, K.A.; Johnson-Roberson, M. Underwater Image Dehazing With a Light Field Camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  19. Ouyang, F.; Yu, J.; Liu, H.; Ma, Z.; Yu, X. Underwater Imaging System Based on Light Field Technology. IEEE Sens. J. 2021, 21, 13753–13760. [Google Scholar] [CrossRef]
  20. Lu, H.; Li, Y.; Kim, H.; Serikawa, S. Underwater light field depth map restoration using deep convolutional neural fields. Artif. Intell. Robot. 2018, 2018, 305–312. [Google Scholar]
  21. Ye, T.; Chen, S.; Liu, Y.; Ye, Y.; Chen, E.; Li, Y. Underwater Light Field Retention: Neural Rendering for Underwater Imaging. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, New Orleans, LA, USA, 19–20 June 2022; pp. 488–497. [Google Scholar]
  22. Akkaynak, D.; Treibitz, T. Sea-Thru: A Method for Removing Water From Underwater Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1682–1691. [Google Scholar] [CrossRef]
  23. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409. [Google Scholar] [CrossRef]
  24. Ng, R.; Levoy, M.; Brédif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light Field Photography with a Hand-held Plenoptic Camera; Stanford University Computer Science Tech Report; Stanford University: Stanford, CA, USA, 2005. [Google Scholar]
  25. Ahmed, N.; Natarajan, T.; Rao, K. Discrete Cosine Transform. IEEE Trans. Comput. 1974, C-23, 90–93. [Google Scholar] [CrossRef]
  26. Carlevaris-Bianco, N.; Mohan, A.; Eustice, R.M. Initial results in underwater single image dehazing. In Proceedings of the OCEANS 2010 MTS/IEEE SEATTLE, Seattle, WA, USA, 20–23 September 2010; pp. 1–8. [Google Scholar] [CrossRef]
  27. Drews, P., Jr.; do Nascimento, E.; Moraes, F.; Botelho, S.; Campos, M. Transmission Estimation in Underwater Single Images. In Proceedings of the 2013 IEEE International Conference on Computer Vision Workshops, Sydney, NSW, Australia, 2–8 December 2013; pp. 825–830. [Google Scholar] [CrossRef]
  28. Gupta, H.; Mitra, K. Unsupervised Single Image Underwater Depth Estimation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 624–628. [Google Scholar] [CrossRef]
  29. Yang, M.; Sowmya, A. An Underwater Color Image Quality Evaluation Metric. IEEE Trans. Image Process. 2015, 24, 6062–6071. [Google Scholar] [CrossRef]
  30. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
Figure 1. Comparison of Depth Estimation and Image Restoration. (a) Central view of an underwater light field image captured with a light field camera. (b) Restored image using Huang’s deep learning-based method [8]. (c) Restored image using Tian’s light field refocusing method [3]. (d) Restored image using our method. (e) Depth estimation by Tian [3]. (f) Depth estimation image using our method.
Figure 1. Comparison of Depth Estimation and Image Restoration. (a) Central view of an underwater light field image captured with a light field camera. (b) Restored image using Huang’s deep learning-based method [8]. (c) Restored image using Tian’s light field refocusing method [3]. (d) Restored image using our method. (e) Depth estimation by Tian [3]. (f) Depth estimation image using our method.
Jmse 12 00935 g001
Figure 2. Schematic diagram of underwater imaging and light propagation in Scattering media.
Figure 2. Schematic diagram of underwater imaging and light propagation in Scattering media.
Jmse 12 00935 g002
Figure 3. Our experimental setup.
Figure 3. Our experimental setup.
Jmse 12 00935 g003
Figure 4. Comparison of depth estimation results. (a) Original image. (b) UDCP [27]. (c) ULAP [10]. (d) IBLA [11]. (e) Tian’s method [3]. (f) UW-Net [28]. (g) Our method.
Figure 4. Comparison of depth estimation results. (a) Original image. (b) UDCP [27]. (c) ULAP [10]. (d) IBLA [11]. (e) Tian’s method [3]. (f) UW-Net [28]. (g) Our method.
Jmse 12 00935 g004
Figure 5. Comparison of restoration results. (a) Original image. (b) UDCP [27]. (c) ULAP [10]. (d) IBLA [11]. (e) Tian’s method [3]. (f) Semi-UIR [8]. (g) Our method.
Figure 5. Comparison of restoration results. (a) Original image. (b) UDCP [27]. (c) ULAP [10]. (d) IBLA [11]. (e) Tian’s method [3]. (f) Semi-UIR [8]. (g) Our method.
Jmse 12 00935 g005
Figure 6. Comparison of restoration results across various turbidity levels using the SSIM index. For better display, our data begins recording from the third addition of an equal amount of milk. Note that aside from our results and Tian’s results, all other views are from the LF’s central perspective. Note that all images are from the same original target with turbidity being the only variable.
Figure 6. Comparison of restoration results across various turbidity levels using the SSIM index. For better display, our data begins recording from the third addition of an equal amount of milk. Note that aside from our results and Tian’s results, all other views are from the LF’s central perspective. Note that all images are from the same original target with turbidity being the only variable.
Jmse 12 00935 g006
Figure 7. Comparison of restoration effects across various turbidity levels using the UCIQE index. Note that all image data is the same as in Figure 6.
Figure 7. Comparison of restoration effects across various turbidity levels using the UCIQE index. Note that all image data is the same as in Figure 6.
Jmse 12 00935 g007
Table 1. Comparison of restoration results using the UCIQE scores corresponding to Figure 5.
Table 1. Comparison of restoration results using the UCIQE scores corresponding to Figure 5.
MethodLF1LF2LF3LF4LF5Average
UDCP20.3220.7521.1619.0720.4420.34
ULAP25.0932.4323.8925.1122.0325.31
IBLA18.3514.3810.0210.129.44712.86
Tian’s method26.1128.7221.8819.4330.0925.24
Semi-UIR16.7225.1324.6421.1619.4321.01
Our method27.2127.9930.2728.4237.2430.22
The best result is shown in Red.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiao, B.; Gao, X.; Huang, H. Optimizing Underwater Image Restoration and Depth Estimation with Light Field Images. J. Mar. Sci. Eng. 2024, 12, 935. https://doi.org/10.3390/jmse12060935

AMA Style

Xiao B, Gao X, Huang H. Optimizing Underwater Image Restoration and Depth Estimation with Light Field Images. Journal of Marine Science and Engineering. 2024; 12(6):935. https://doi.org/10.3390/jmse12060935

Chicago/Turabian Style

Xiao, Bo, Xiujing Gao, and Hongwu Huang. 2024. "Optimizing Underwater Image Restoration and Depth Estimation with Light Field Images" Journal of Marine Science and Engineering 12, no. 6: 935. https://doi.org/10.3390/jmse12060935

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop