Next Article in Journal
Multi-Parametric Methodology for the Feasibility Assessment of Alternative-Fuelled Ships
Next Article in Special Issue
Optimizing Underwater Image Restoration and Depth Estimation with Light Field Images
Previous Article in Journal
A Study on Fishing Vessel Energy System Optimization Using Bond Graphs
Previous Article in Special Issue
EF-UODA: Underwater Object Detection Based on Enhanced Feature
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deblurring Network for UV Images of Colorless Chemicals Floating on Sea Surface

1
Ocean College, Zhejiang University, Zhoushan 316021, China
2
Hainan Institute, Zhejiang University, Sanya 572025, China
3
Key Laboratory of Ocean Observation–Imaging Testbed of Zhejiang Province, Zhoushan 316021, China
4
Shanghai Shipbuilding Technology Research Institute, Zhoushan Ship Engineering Research Center, Zhoushan 316100, China
5
Zhejiang Lvhai Salt Manufacturing Co., Ltd., Zhoushan 316021, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2024, 12(6), 904; https://doi.org/10.3390/jmse12060904
Submission received: 17 March 2024 / Revised: 10 May 2024 / Accepted: 17 May 2024 / Published: 28 May 2024
(This article belongs to the Special Issue Underwater Engineering and Image Processing)

Abstract

:
Automatic imaging monitoring technology has been developed as a mature solution to address oil spill accidents. However, when airborne equipment is used to monitor the area of leakage, there is a chance of relative motion between the equipment and the target, leading to blurred images. This can hinder subsequent monitoring tasks, such as the identification and segmentation of hazardous chemical targets. Therefore, it is crucial to deblur these images, as sharper images can substantially improve the precision and accuracy of these tasks. In light of this, this paper presents a novel multi-scale recurrent deblurring network, leveraging the Convolutional Block Attention Module (CBAM) attention mechanism, to restore clear images in an end-to-end manner. The CBAM attention mechanism has been incorporated, which learns features in both the spatial and channel domains, thereby enhancing the deblurring ability of the network by combining the attention mechanisms of multiple domains. To evaluate the effectiveness of the proposed method, we applied the deblurring network to our self-built dataset of blurred ultraviolet (UV) images of colorless chemicals floating on water surfaces. Compared to the SRN-Deblur deblurring network, the proposed approach yields improved PSNR and SSIM results. Moreover, when compared with the state-of-the-art DeblurGANv2 method, it performs better in the SSIM. These results demonstrate that the proposed multi-scale recurrent deblurring network, based on the CBAM attention mechanism, exhibits a superior deblurring performance.

1. Introduction

As technology advances, computer vision tasks increasingly demand high-quality and clear images. However, in many computer vision applications, such as target recognition and segmentation, the input images often contain a significant number of low-quality, blurred images. The unclear boundaries of the targets in these blurred images can impede the quality of downstream computer vision tasks.
Motion blur is a common cause of image blurring, which occurs when there is relative motion between the object being photographed and the capturing equipment, such as an optical camera, during the shooting process.
Deblurring techniques are designed to restore and reconstruct clear images from blurred ones. By deblurring an image, the quality of the images improves, which increases the performance of computer vision tasks.
The rapid development of neural networks has garnered significant attention from researchers, with many applying this technology to image-processing tasks, including image deblurring. In this paper, we build upon the work in [1], which proposed a multi-scale recurrent neural network for deblurring by incorporating the CBAM attention mechanism [2], which combines the channel and spatial attention mechanisms. Unlike the network proposed in [1], our approach utilizes a long short-term memory (LSTM) network to establish connections between the intermediate features of the three resolutions, progressing from coarse to fine.
This paper proposes a deblurring network designed for UV images. The following is a brief summary of our work:
(1)
Our proposed network builds upon the multi-scale recurrent neural network, with the introduction of the CBAM attention mechanism. This module enables attention to be focused on the latent features of both the channel and spatial dimensions;
(2)
The multi-scale recurrent neural network utilized in this work employs an LSTM network, which is capable of capturing and utilizing relational temporal features;
(3)
We constructed a UV-image dataset consisting of colorless chemicals floating on water surfaces and utilized it to test the deblurring network’s ability to restore clear images.
In conclusion, we compared our proposed method to other advanced methods on a self-collected UV-image dataset of colorless chemicals floating on water, and the results demonstrate its superior deblurring effectiveness.

2. Related Work

In this section, we review related works, including image-deblurring methods and attention mechanisms.

2.1. Image-Deblurring Methods

Image deblurring belongs to the methodology aimed at improving the image quality and falls under the category of image enhancement methods [3,4,5]. Image-deblurring methods are broadly divided into traditional deblurring methods and deep learning deblurring methods.
Traditional deblurring methods for non-blind deblurring problems essentially involve the exploration of various deconvolution algorithms. The more classical algorithms include the Lucy–Richardson algorithm, which recovers in an iterative convergence manner [6], the Wiener filtering method [7], inverse filtering, etc. Inverse filtering is the earliest deblurring method, and directly applying inverse filtering is the simplest deblurring method. In 1967, Harris and colleagues [8] implemented image deblurring using Wiener filtering (minimum mean square error) through the approach that minimizes the mean square error between the estimate and the original to recover images. The Lucy–Richardson algorithm based on Bayesian theory uses a non-linear iterative process to solve for the maximum-likelihood solution of blurred images, resulting in clear images. In 2017, Zhang et al. [9] used fully convolutional networks to deblur known blur kernels, resulting in clear images. To address the problem of biased blur kernel estimation in non-blind deblurring, Vasu et al. [10] proposed a deblurring method based on convolutional neural networks in 2018.
All of the above are non-blind deblurring algorithms that deblur based on known blur kernels. However, there are many causes for blurred images, such as the shaking of the shooting device, the movement of the target, lens defocusing, etc., and in most real shooting scenes, the blur kernel is unknown. Therefore, non-blind deblurring algorithms are limited in practical applications.
The crux of blind deblurring algorithms is the estimation of the blur kernel function, which is used in accordance with the image degradation model to estimate the clear images. The main research aim of blind deblurring algorithms is how to accurately estimate the blur kernel function using the image’s blur information. Nah et al. [11] proposed DeepDeblur, which refines the output gradually from the coarse scale to the fine scale, and first accomplished image-deblurring tasks in an end-to-end training manner. In 2019, inspired by Spatial Pyramid Matching [12], Zhang et al. [13] proposed the Deep Multi-Patch Hierarchical Network (DMPHN), which processes the blurred images from fine to coarse. In 2021, Zamir et al. [14] built upon the DMPHN model and, in order to strike a better balance between the spatial details and high-level contextual information, proposed a multi-stage progressive image recovery model (MPRNet).
CNNs are typically designed to extract local features from images, which may not generalize well to the problem of dynamic-scene deblurring. Moreover, image blur can have various causes and vary from image to image, making it challenging to directly address this issue using CNNs. To address these challenges, we propose a method that combines CNNs with attention mechanisms to extract image features from multiple domains. By leveraging attention mechanisms across multiple domains, our approach can improve the model’s deblurring ability.

2.2. Attention Mechanism

The attention mechanism in deep learning is inspired by human behavior, and its various implementations differ in the domains they focus on. Some methods, such as those proposed in [15,16], concentrate on the spatial domain. These methods transform the spatial information in the original image to another space while retaining the most important features. Other methods, such as those proposed in [2,17,18], concentrate on the channel domain. These methods focus on establishing relationships between channels.
In our work, we primarily focus on the attention mechanism in both the spatial and channel domains. Our key idea is to leverage the attention mechanism to identify image regions with residual blur. We learn features in both the spatial and channel domains, and we further improve the deblurring performance of the model by combining attention mechanisms from multiple domains.

3. Proposed Deblurring Network

Based on SRN-DeblurNet [1], we propose the network architecture shown in Figure 1. The main structure of SRN-DeblurNet is augmented with the CBAM attention mechanism to learn features in both the spatial and channel domains. By combining the attention mechanisms of multiple domains, the model’s deblurring ability is further improved.

3.1. CBAM Attention Mechanism

The CBAM attention mechanism, as an important part of the network, is structured as shown in Figure 2. The initial restored image output by this module realizes a multi-scale structure, allowing the network to have different receptive fields. The output attention feature map adjusts the spatial importance of the feature map during the process of reconstructing the image on the encoding–decoding side, making it more adaptable. The Convolutional Block Attention Module consists of two sub-modules, namely, the channel attention module and the spatial attention module, making it a lightweight and general-purpose module. The module takes the intermediate feature map generated by the convolutional layer of the neural network as input and infers the attention map through the channel and spatial dimensions. The attention map is then multiplied by the input feature map to achieve the adaptive refinement of the features. Its lightweight and general characteristics make it easy to integrate into CNNs with negligible computational overhead.
The channel attention module is designed to capture the inter-channel relationship of the features and generate channel–channel attention maps, which help to identify the most relevant channels in the feature map. Channel attention is concerned with the “what” of the input image, such as detecting specific objects or patterns. In contrast, the spatial attention module is designed to capture the spatial relationship of the features and generate spatial attention maps, which highlight the most informative spatial locations in the feature map. Spatial attention is concerned with the “where” of the input image, such as identifying the important regions of an image. Together, the channel and spatial attention modules work in a complementary manner to improve the overall feature representation of the network [2].

3.2. Image-Deblurring Network

The UV-image-deblurring network adopts a “coarse-to-fine” strategy. “Coarse-to-fine” refers to extracting the characteristics of the input UV image of floating colorless chemicals at three different resolution scales. The network structure is based on the U-net network structure and consists of two parts: an encoder and a decoder.
First, the encoder side uses residual dense blocks to improve the feature extraction and learning capabilities of the input blurred image and retain high-frequency detailed information in the image. The encoder consists of three encoder modules. As shown in Figure 2, each encoder module is composed of a convolution layer and three residual dense blocks in series; each residual dense block is composed of two convolution layers activated by a ReLU A structure composed of function series and connected through a skip connection. The input of this module is the output of the previous residual dense block.
An LSTM module is added between the encoder and the decoder to establish the connection between the intermediate features of the scales at three resolutions “from coarse to fine”, so that the features of different scales can interact with each other.
The decoder side consists of three decoder modules. As shown in Figure 2, each decoder module consists of a residual dense block, a CBAM, two residual dense blocks, and a deconvolution layer in series. The addition of the CBAM attention mechanism enables the deblurring network to learn features in both the spatial domain and channel domain, thereby improving the network’s deblurring ability by integrating attention mechanisms from multiple fields. The CBAM is an important part of the deblurring network. The preliminary restored image output by this module has a multi-scale structure, so the network has different receptive fields. The output attention feature map adjusts the spatial importance of the feature map during the image reconstruction process on the encoding–decoding sides, resulting in stronger adaptability to different degrees of blur.
The entire network is composed of three identical sub-networks. After the input blurred image is output by the first sub-network, the de-blurred image is upsampled and used as the input of the next sub-network. Similarly, the output of this sub-network is input to the next sub-network through upsampling. In the encoding stage, the image is gradually downsampled to reduce the network parameters. Correspondingly, in the decoding stage, the image is gradually upsampled to restore a clear image. The entire network uses a “coarse-to-fine” strategy to deblur the blurred image at three scales.

3.3. Loss and Training Details

We adopt a Euclidean loss function that downsamples both the network output and the ground-truth (GT) image to the same size using bilinear interpolation at each scale. The loss function is defined as follows:
L = i = 1 n κ i N i I i I * i 2 2
where I i and I * i are the network output and the ground-truth image at the i-th scale, respectively; { κ i } is the weight of each scale, which we set empirically to 1.0; N i is the number of elements in I i to be normalized. Although we also experimented with full mutation and adversarial losses, we found that the L2-norm loss is sufficient to produce clear and unambiguous results.

4. Experimental Results

Our experiments were conducted on a PC equipped with an NVIDIA GTX1080Ti GPU (Santa Clara, CA, USA) and an Intel i7–[email protected] CPU (Santa Clara, CA, USA). We implemented our framework [13] on the TensorFlow platform. Our evaluation was comprehensive and aimed to validate different network structures and various network parameters. To ensure fairness, all experiments were performed on the same dataset with the same training configuration unless otherwise specified.
The input UV images of colorless chemicals floating on water surfaces have a different channel number compared to that of ordinary RGB images. While RGB images have three channels, UV images have only one channel. Therefore, for UV images, the model’s channel number is set to one.

4.1. Dataset

In our study of ultraviolet (UV) images of colorless chemicals floating on water surfaces, we utilized a variety of technical tools for data collection. These included Unmanned Aerial Vehicles (UAVs) and handheld UV cameras, which both played significant roles. The former facilitated broad high-altitude photography, while the latter allowed for the acquisition of detailed, close-up images.
Additionally, we took the diversity of the collection environments into account. The collection points ranged from self-constructed swimming pools to aqueducts, man-made lakes, and seawater. The swimming pool served as a controlled environment where we could conduct preliminary experiments, and the wide surface of the water offered richer samples. Aqueducts, with their diverse environmental conditions and certain dynamism, introduced fresh challenges to our collection efforts. However, sampling from artificial lakes and seawater enabled our data to cover broader and more profound areas, which was significantly beneficial to our research.
Our team faced various challenges proactively, right from the controlled environment of the self-built swimming pool to the aqueducts and natural environments of artificial lakes and seawater. We were successful in acquiring a considerable amount of high-quality data. These data provided an essential foundation for our subsequent in-depth examination of colorless chemicals floating on water surfaces.
We collected clear ultraviolet images of colorless chemicals floating on water surfaces and applied a simulated motion-blur algorithm (motionblur) to blur the clear images, resulting in a total of 1079 pairs of clear–blurred UV images. Figure 3 illustrates the effects of the motion-blur processing on the clear images.

4.2. Evaluation Methods

The peak signal-to-noise ratio (PSNR) and the structural similarity index (SSIM) were used to evaluate the effectiveness of the deblurring algorithm.
The PSNR is an engineering term used to measure the fidelity of a signal by comparing it to a reference signal and calculating the ratio of the maximum possible power of the signal to the noise power that affects its accuracy. A higher value indicates lower distortion. A PSNR value higher than 40 dB indicates excellent image quality (i.e., very close to the original image); a range of 30–40 dB usually indicates good image quality (i.e., noticeable but acceptable distortion); a range of 20–30 dB indicates poor image quality; and finally, images with PSNRs below 20 dB are considered unacceptable.
The mean square error (MSE) is a measure of the difference between a clean image (I) and a noisy image (K) of the size m × n and is defined as follows:
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I i , j K ( i , j ) ] 2
Then, the PSNR(dB) is defined as follows:
P S N R = 10 · l o g 10 ( M A X I 2 M S E )
The term M A X I 2 denotes the maximum possible pixel value of the image. In the case of an eight-bit binary representation of each pixel, this value is 255. In general, if each pixel is represented by B-bit binary, then M A X I = 2 B 1 . For uint8 data, the maximum pixel value is usually 255, whereas for floating-point data, the maximum pixel value is 1. The PSNR(dB) formula represents the ratio of the maximum possible power of a signal to its mean square error, where higher values indicate better image quality.
The above is the calculation method for grayscale images. If it is a color image, there are typically three methods for calculation:
(1)
Calculate the PSNR separately for each of the three RGB channels, and then compute the average of the three PSNR values;
(2)
Calculate the MSE for each of the three RGB channels separately, then average the three MSE values and divide by 3;
(3)
Convert the image to the YCbCr color space, and then compute the PSNR only for the Y component, which represents the image brightness.
The SSIM measures both the degrees of distortion and similarity between two images. It evaluates the image similarity from three aspects: luminance, contrast, and structure. Unlike the MSE and PSNR, which measure the absolute error, the SSIM is a perceptual model that takes into account the visual perception of the human eye. The SSIM values range between 0 and 1, with larger values indicating smaller image distortion and higher similarity between the two images:
S S I M x , y = [ l ( x , y ) α · c ( x , y ) β · s ( x , y ) γ ]
The terms in the above equation are given by the following:
l ( x , y ) α = 2 μ x μ y + c 1 μ x 2 + μ y 2 + c 1 ,   c ( x , y ) β = 2 σ x σ y + c 2 σ x 2 + σ y 2 + c 2 ,   s ( x , y ) γ = σ x y + c 3 σ x σ y + c 3
μ x is the mean value of x;
μ y is the mean value of y;
σ x 2 is the variance of x;
σ y 2 is the variance of y;
σ x σ y is the covariance of x and y;
c 1 = ( k 1 L ) 2 ,     c 2 = ( k 2 L ) 2 are two constants; avoid division by zero;
l is the range of pixel values, 2 B 1 ;
k 1 = 0.01 ,     k 2 = 0.03 is the default value.

4.3. Ablation Experiments

In this section, we evaluate the performances of our key methods. We used a self-built dataset of UV images of colorless chemicals floating on water surfaces for the ablation research. To investigate the influence of the CBAM attention mechanism on the model, we conducted ablation experiments with and without adding the CBAM attention mechanism to the encoder and decoder sides. We kept the rest of the methods fixed for a fair comparison when changing the value of one method. The results are presented in Table 1. Figure 4 and Figure 5 depict the network structure diagrams of the deblurring network with the CBAM added to the encoder side and the CBAM added to the encoder and decoder sides, respectively.
In each of our ablation experiments, we maintain a consistent number of training epochs at 14,000. Ensuring this consistency allows for an accurate and impartial evaluation of the impact of different network components and their respective contributions to overall performance.
As shown in Table 1, with the same training steps, adding the CBAM on the decoder side, such as in ① and ③, can significantly improve the model’s performance in terms of the PSNR. It was also found that model ③ with only the CBAM added on the decoder side performed the best in terms of the PSNR, and the SSIM also improved significantly, with no significant difference from the optimal SSIM. However, adding the CBAM on the encoder side resulted in worse PSNR and SSIM performances of the model.
In summary, adding the CBAM only on the decoder side significantly improved the deblurring effect of the model, demonstrating that the deblurring network based on the CBAM attention mechanism proposed in this work is superior to the original deblurring network.

4.4. Deblurring Results

The clear ultraviolet images of colorless chemicals floating on water surfaces underwent simulated motion-blur processing, and the effect of this processing is shown in Figure 3. It is evident that the outline boundary of the ultraviolet image processed by motion blur becomes unclear, which simulates the blur caused by the relative motion of the shooting equipment and the target object during shooting.
The blurred ultraviolet image of colorless chemicals floating on a water surface was fed into the deblurring network that utilizes the CBAM attention mechanism proposed by us, and the deblurring process was performed. The deblurring effect is shown in Figure 6.
The improvement in the image clarity after passing the blurred ultraviolet image through the proposed deblurring network is evident from Figure 6. The resulting image appears sharper, and the details, such as the outline and area of hazardous chemicals, are easier to observe. These deblurred images will be immensely helpful in subsequent monitoring tasks, such as hazardous chemical target segmentation and target detection, as they significantly enhance the precision and accuracy of these tasks.
In order to highlight the merit of the proposed method compared to other deblurring methods designed for visible-light images, the performance was tested on a dataset of ultraviolet images of colorless chemicals floating on water surfaces that were motion-blurred. Contemporary, advanced visible-light deblurring methods were adopted for this test. The metrics included the PSNR, SSIM, inference time for a single blurry image, and number of model parameters selected as benchmarks for comparison. The comparative results are shown in Table 2.
Table 2 illustrates the significant advantage of the enhanced UV-image-deblurring network that was designed and is explained in this chapter for floating chemical UV images, compared to the current advanced deblurring methods for visible-light images. It shows a remarkable advantage in the PSNR value and SSIM value. The PSNR reached 32.173 dB, an improvement of 1.6% compared to the next best method, MPRNet. The SSIM reached 0.962, an improvement of approximately 0.4% compared to the next best method. Moreover, the UV-image-deblurring network has the smallest number of parameters, meeting the design objective of being lightweight, and its inference time does not significantly differ from those of the other methods. This validates the certain advantage of the improved UV-image-deblurring method designed according to the characteristics of UV images.
To further validate the performance of the UV-image-deblurring network, various deblurring methods were tested on the chemical UV images. Figure 7 shows the comparison results of the different deblurring methods, where Figure 7a is the input of the original blurry UV image, and Figure 7b–e depict the results of applying various deblurring methods to UV images. As shown in Figure 7b, when the DeblurGANv2 method is applied for UV-image deblurring, the image detail recovery is mediocre, with significant blurring issues still existing in the details. The overall image edge has artifacts and ringing phenomena, resulting in poor image quality. Figure 7c,d present the results of using the DMPN and MPRNet methods, which are currently known for their excellent recovery effects on RGB images. It can be seen that when these two methods are applied for the deblurring of blurry UV images, the deblurring results in a significant improvement in the details compared to the DeblurGANv2 method, with clearer detail information. However, just like with the DeblurGAN method, the edges of the entire image still have severe artifacts and ringing phenomena, making the deblurring results unsatisfactory. Figure 7e displays the recovery effect of the deblurring method proposed for UV images in this chapter. It can be observed that, compared to the previous methods, the detail information of the image was better restored and preserved after deblurring. Furthermore, there are no prominent artifacts and ringing phenomena at the image edge. The overall image is smoother, which leads to better deblurring results.

5. Conclusions

After a chemical spill, most of the chemicals will float on the surface of the sea, presenting as colorless, and at this time, it is difficult to distinguish between the chemicals and seawater through images under visible light. Despite the fact that UV sensors can detect thin liquid films more effectively, due to equipment vibrations and the impact of waves, the obtained UV images can be blurred, causing the boundary of the chemicals to be unclear and bringing challenges to subsequent segmentation tasks. Deep learning holds great potential for image deblurring, but most applications concentrate on visible-light images, lacking specific methods for deblurring UV images. Therefore, it is extremely important to study methods specifically for deblurring UV images of colorless chemicals floating on water surfaces. This will help to recover the contour information of the chemicals in the UV images, improve the precision of the segmentation, and more accurately determine the area of the leak, which is of great significance for the emergency response to marine chemical spill incidents. For this reason, we introduced the CBAM attention mechanism and improved the U-net deep learning deblurring network framework based on the methods of deblurring visible-light images. We extracted image features from both the spatial and channel domains to restore the contour information of the chemicals in the blurred UV images, thereby improving the image quality. After comparison and verification, we found that our method outperforms other methods in terms of its deblurring effects and segmentation precision. Compared to the most advanced deblurring methods currently available, such as DeblurGANv2, DeepDeblur, DMPHN, and MPRNet, our method improved by 1.6% and 0.4%, respectively, in the PSNR and SSIM evaluations.
It has been verified that the improved UV-image-deblurring method, designed specifically for UV images of colorless chemicals floating on water surfaces, results in superior visual impressions and the better recovery of detail information in actual UV images after deblurring, compared with the existing deblurring methods designed for visible-light images.

Author Contributions

H.H.: conceptualization, validation, formal analysis, resources, data curation, writing—review and editing, supervision; Y.Z.: conceptualization, methodology, software, validation, investigation, writing—original draft preparation, visualization; C.L.: project administration, funding acquisition; F.L.: project administration, funding acquisition; Y.F.: project administration, funding acquisition; H.S.: writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science and technology projects in the “seed breeding, deep-sea and aerospace” industries, grant number “DSTIC-SSPT-2022006”, and the Science and technology projects of Zhoushan in the “Microscopic hyperspectral imaging analysis and detection of salt”, grant number “2022C01023”.

Data Availability Statement

The datasets generated during the current study are available from the corresponding author upon reasonable request. The ultraviolet-image data were obtained by a drone-borne ultraviolet camera but restrictions apply to the availability of these data, which were used under license for the current study and are thus not publicly available.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Tao, X.; Gao, H.; Shen, X.; Wang, J.; Jia, J. Scale-Recurrent Network for Deep Image Deblurring. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; IEEE: New York, NY, USA, 2018; pp. 8174–8182. [Google Scholar]
  2. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing AG: Cham, The Netherlands, 2018; Volume 11211, pp. 3–19. [Google Scholar]
  3. Zhang, W.; Wang, Y.; Li, C. Underwater Image Enhancement by Attenuated Color Channel Correction and Detail Preserved Contrast Enhancement. IEEE J. Ocean. Eng. 2022, 47, 718–735. [Google Scholar] [CrossRef]
  4. Zhou, J.; Zhang, D.; Zhang, W. Cross-View Enhancement Network for Underwater Images. Eng. Appl. Artif. Intell. 2023, 121, 105952. [Google Scholar] [CrossRef]
  5. Zhang, W.; Zhuang, P.; Sun, H.-H.; Li, G.; Kwong, S.; Li, C. Underwater Image Enhancement via Minimal Color Loss and Locally Adaptive Contrast Enhancement. IEEE Trans. Image Process. 2022, 31, 3997–4010. [Google Scholar] [CrossRef] [PubMed]
  6. Lucy, L.B. An Iterative Technique for the Rectification of Observed Distributions. Astron. J. 1974, 79, 745. [Google Scholar] [CrossRef]
  7. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear Total Variation Based Noise Removal Algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  8. Harris, J.L. Image Evaluation and Restoration. JOSA 1966, 56, 569–574. [Google Scholar] [CrossRef]
  9. Zhang, J.; Pan, J.; Lai, W.-S.; Lau, R.W.H.; Yang, M.-H. Learning Fully Convolutional Networks for Iterative Non-Blind Deconvolution. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3817–3825. [Google Scholar]
  10. Vasu, S.; Maligireddy, V.R.; Rajagopalan, A.N. Non-Blind Deblurring: Handling Kernel Uncertainty with CNNs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3272–3281. [Google Scholar]
  11. Nah, S.; Kim, T.H.; Lee, K.M. Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 257–265. [Google Scholar]
  12. Lazebnik, S.; Schmid, C.; Ponce, J. Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; Volume 2, pp. 2169–2178. [Google Scholar]
  13. Zhang, H.; Dai, Y.; Li, H.; Koniusz, P. Deep Stacked Hierarchical Multi-Patch Network for Image Deblurring. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2019), Long Beach, CA, USA, 15–20 June 2019; IEEE Computer Society: Los Alamitos, CA, USA, 2019; pp. 5971–5979. [Google Scholar]
  14. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Multi-Stage Progressive Image Restoration. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14816–14826. [Google Scholar]
  15. Jaderberg, M.; Simonyan, K.; Zisserman, A. Spatial Transformer Networks. In Proceedings of the Advances in Neural Information Processing Systems 28 (NIPS 2015), Montreal, QC, Canada, 7–12 December 2015; Volume 28. [Google Scholar]
  16. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-Local Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  17. Du, Y.; Yuan, C.; Li, B.; Zhao, L.; Li, Y.; Hu, W. Interaction-Aware Spatio-Temporal Pyramid Attention Networks for Action Classification. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; IEEE: New York, NY, USA, 2018; pp. 373–389. [Google Scholar]
  18. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
Figure 1. Deblurring network with CBAM attention mechanism (CBAM on decoder side).
Figure 1. Deblurring network with CBAM attention mechanism (CBAM on decoder side).
Jmse 12 00904 g001
Figure 2. Encoder–decoder module structure with CBAM attention mechanism added (The description of modules in different colors is consistent with Figure 1).
Figure 2. Encoder–decoder module structure with CBAM attention mechanism added (The description of modules in different colors is consistent with Figure 1).
Jmse 12 00904 g002
Figure 3. Motion-blur effects: (a) original clear UV image, (b) blurred UV image after blurring treatment.
Figure 3. Motion-blur effects: (a) original clear UV image, (b) blurred UV image after blurring treatment.
Jmse 12 00904 g003
Figure 4. Deblurring network with CBAM attention mechanism (CBAM on encoder side).
Figure 4. Deblurring network with CBAM attention mechanism (CBAM on encoder side).
Jmse 12 00904 g004
Figure 5. Deblurring network with CBAM attention mechanism on (CBAM on encoder and decoder sides).
Figure 5. Deblurring network with CBAM attention mechanism on (CBAM on encoder and decoder sides).
Jmse 12 00904 g005
Figure 6. Deblurring results: (a) original blurred UV image, (b) UV image after deblurring using the method proposed in this study.
Figure 6. Deblurring results: (a) original blurred UV image, (b) UV image after deblurring using the method proposed in this study.
Jmse 12 00904 g006
Figure 7. Comparison of different deblurring methods.
Figure 7. Comparison of different deblurring methods.
Jmse 12 00904 g007
Table 1. Results of ablation experiments.
Table 1. Results of ablation experiments.
Experiment NumberEncoder (CBAM)Decoder (CBAM)LSTMPSNRSSIMParams
(MB)
Inference Time (s)
28.1530.88292.2951.501
×27.8460.88192.2481.449
x28.2540.88592.2471.483
xx27.8920.88492.2001.422
Table 2. Deblurring effect of each model.
Table 2. Deblurring effect of each model.
ModelPSNR (dB)SSIMTime (s)Parameter (MB)
DeepDeblur29.0510.9160.42111.713
DeblurGANv231.1120.9350.35060.921
DMPHN31.2040.9400.42221.742
MPRNet31.6630.9580.89320.121
Our Model32.1730.9620.5106.832
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, H.; Zhang, Y.; Lin, C.; Lin, F.; Fang, Y.; Song, H. Deblurring Network for UV Images of Colorless Chemicals Floating on Sea Surface. J. Mar. Sci. Eng. 2024, 12, 904. https://doi.org/10.3390/jmse12060904

AMA Style

Huang H, Zhang Y, Lin C, Lin F, Fang Y, Song H. Deblurring Network for UV Images of Colorless Chemicals Floating on Sea Surface. Journal of Marine Science and Engineering. 2024; 12(6):904. https://doi.org/10.3390/jmse12060904

Chicago/Turabian Style

Huang, Hui, Yu Zhang, Chuansong Lin, Fuliang Lin, Yue Fang, and Hong Song. 2024. "Deblurring Network for UV Images of Colorless Chemicals Floating on Sea Surface" Journal of Marine Science and Engineering 12, no. 6: 904. https://doi.org/10.3390/jmse12060904

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop