Next Article in Journal
Developments in Marine Propulsors
Previous Article in Journal
Assessment of Diversity of Marine Organisms among Natural and Transplanted Seagrass Meadows
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Image Translation via Multi-Scale Generative Adversarial Network

1
College of Computer Science and Technology, Harbin Engineering University, Harbin 150001, China
2
School of Computing and Information Systems, Singapore Management University, Singapore 178902, Singapore
3
National Key Laboratory of Underwater Acoustic Technology, Harbin Engineering University, Harbin 150001, China
*
Authors to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2023, 11(10), 1929; https://doi.org/10.3390/jmse11101929
Submission received: 31 August 2023 / Revised: 27 September 2023 / Accepted: 2 October 2023 / Published: 6 October 2023

Abstract

:
The role that underwater image translation plays assists in generating rare images for marine applications. However, such translation tasks are still challenging due to data lacking, insufficient feature extraction ability, and the loss of content details. To address these issues, we propose a novel multi-scale image translation model based on style-independent discriminators and attention modules (SID-AM-MSITM), which learns the mapping relationship between two unpaired images for translation. We introduce Convolution Block Attention Modules (CBAM) to the generators and discriminators of SID-AM-MSITM to improve its feature extraction ability. Moreover, we construct style-independent discriminators that enable the discriminant results of SID-AM-MSITM to be not affected by the style of images and retain content details. Through ablation experiments and comparative experiments, we demonstrate that attention modules and style-independent discriminators are introduced reasonably and SID-AM-MSITM performs better than multiple baseline methods.

1. Introduction

Underwater images play critical roles in diverse marine-related military and scientific applications, such as seabed sediment classification [1], submarine cable detection [2], and mine recognition [3]. However, the complex underwater environment limits the use of camera devices, including Kinect units [4] and binocular stereo cameras [5], which makes it difficult to obtain real underwater images. Intuitively, image translation [6] provides a viable direction to obtain such scarce data. Specifically, image translation methods translate source-domain non-underwater images into target-domain underwater ones, which re-assigns particular attributes of underwater images. The translated underwater images are valuable for advanced visual tasks such as target detection, 3D reconstruction, and target segmentation [7,8].
Existing image translation methods are generally categorized into two groups, i.e., conventional and Generative Adversarial Network (GAN)-based [9] methods. First, conventional methods [10,11] extract low-level features to transfer input images’ texture or devise various Convolutional Neural Network-based (CNN) [12], such as image style transfer [13], to make use of semantic content for image translation. Second, GAN-based methods, such as supervised models including Pix2Pix [14] and Pix2PixHD [15], and unsupervised models including StyleGAN [16] and StarGAN [17], use generators to translate images from one image domain to another image domain. By comparison, GAN-based methods do not require researchers to design complex loss functions, which saves manpower.
In view of the immense potential of GAN, researchers attempt to apply it to underwater image translation tasks. Li et al. [18] propose an unsupervised WaterGAN that uses in-air RGB images and depth maps to generate corresponding realistic underwater images. Wang et al. [19] use an unsupervised image translation method that also takes in-air RGB-D images to generate realistic underwater images. Li et al. [20] work to generate images with underwater style using in-air RGB images.
Although existing methods have achieved a certain success, they still face multiple challenges. (1) Data lacking. The necessary paired in-air and depth images are too scarce to train translation models. In the absence of data, using the general image translation methods, it is difficult to achieve good results. (2) Insufficient feature extraction ability. Underwater optics images and underwater sonar images present obvious colors, which reduce the visibility of objects in translated images and reduce the quality of the translated images. These translated images limit the performance of subsequent advanced computer vision tasks [21]. Therefore, the colors of the underwater images pose challenges to the feature extraction ability of image translation. (3) Loss of content details. GAN models are sensitive to the style of images (such as color and texture) [22], which makes image translation models ignore the content information of images. Similar to the lack of feature extraction ability, the loss of content details also limits the performance of subsequent advanced computer vision tasks.
With the aim of addressing the above challenges, in this paper, we propose a multi-scale underwater image translation model based on style-independent discriminators and attention modules (SID-AM-MSITM). Specifically, our contributions mainly include the following:
  • In response to the data lacking challenge, we construct the backbone model of SID-AM-MSITM based on a fundamental image translator, TuiGAN [23]. TuiGAN conducts image translation tasks based on only two unpaired images, and we thus make further improvements on its encoders and decoders.
  • In response to the challenge of insufficient feature extraction ability, we propose to apply Convolution Block Attention Modules (CBAM) [24] to the generators and discriminators of SID-AM-MSITM. CBAM assigns the weight distribution of feature maps in the two dimensions of channel and space and increases the weight of important features, so as to make SID-AM-MSITM pay attention to meaningful information.
  • In response to the loss of content details, we further improve SID-AM-MSITM by constructing style-independent discriminators. The discriminators give similar results when discriminating images with the same content and different styles, so as to make SID-AM-MSITM focus on the content information instead of the style information.
  • We conduct systematical experiments based on multiple datasets, including submarine, underwater optics, sunken ship, crashed plane, and underwater sonar images. Compared with multiple baseline models, SID-AM-MSITM improves the ability to access effective information and retain content details.
The rest of this paper is organized as follows: Section 2 details the methodology of SID-AM-MSITM. Section 3 presents our ablation and comparative experiments, and their corresponding analysis. Section 4 concludes this work.

2. Methodology

In this section, we will present our proposed SID-AM-MSITM in detail. Based on TuiGAN as a backbone architecture, SID-AM-MSITM improves its generators and discriminators by introducing CBAM and makes further improvements by devising style-independent discriminators. Figure 1 shows the architecture of SID-AM-MSITM and the process of translating a non-underwater image into an underwater image.
In the underwater image translation task, the source domain image means the non-underwater image I X , and the target domain image means the underwater image I Y . We use SID-AM-MSITM to translate non-underwater images into underwater images, also known as generating underwater images. Using SID-AM-MSITM, we also reconstruct translated underwater images into non-underwater images. The original image represents the initial image that has not been processed by SID-AM-MSITM.
As in TuiGAN, generally, SID-AM-MSITM downsamples two images into different scales { I X 0 , I X 1 , I X N . I Y 0 , I Y 1 , I Y N } , and each scale corresponds to two generators { G X Y n , G Y X n } and two discriminators { D Y n , D X n } . The generators { G X Y 0 , G X Y 1 , G X Y N 1 } utilize the downsampled source domain images { I X 0 , I X 1 , I X N 1 } as well as their previous-scale upsampled generated images { I X Y 1 , I X Y 2 , I X Y N } to generate new images. At the lowest scale N, the previous-scale upsampled generated image is replaced with an image with pixel values of 0.
The discriminators { D Y 0 , D Y 1 , D Y N } learn the distribution of the target domain using a variety of loss functions for model training, including adversarial loss WGAN-GP [25], cycle-consistency loss, identity loss, and total variation loss [26]. Among them, the cycle-consistency loss helps SID-AM-MSITM to avoid the mode collapse, the identity loss helps it to align colors and textures, and TV loss smooths the generated images. Finally, the translated underwater image I X Y 0 is obtained at the highest scale. In Figure 1, the discriminator processes images translated by a generator of the same color as it.
In order to obtain the generators { G Y X 0 , G Y X 1 , G Y X N } , we also train the generators { G Y X 0 , G Y X 1 , G Y X N } and discriminators { D X 0 , D X 1 , D X N } in a similar way.

2.1. Generators and Discriminators with CBAM Modules

We start with presenting generators and discriminators with CBAM modules. CBAM enables SID-AM-MSITM to focus on critical features in a given image, so as to improve the quality of generated images [27].
As illustrated in Figure 2, each CBAM contains two modules, a channel attention module [28] and a spatial attention module. Specifically, the channel attention module enables SID-AM-MSITM to focus on critical features in a given image, so as to obtain accurate weights of channel features. The spatial attention module performs Max Pooling and Average Pooling on the spatial dimension of compressed features. The yellow blocks represent the input features of the source domain, and the blue blocks represent the input features of the target domain.
Figure 3 shows the generator and discriminator of SID-AM-MSITM, where the generator G X Y n implements the translation of a source domain image I X n to a generated image I X Y n . Firstly, SID-AM-MSITM simply processes I X n from source domain X to obtain an intermediate image I X Y , ϕ n through the CBAM module and convolution operations. Then, SID-AM-MSITM utilizes I X Y , ϕ n , I X n , and an upsampled image after generating at the previous scale I X Y n + 1 to concatenate in the direction of the channel, and obtain a mask image X n through the CBAM module and convolution operations. Finally, a generated image I X Y n is obtained using the linear combination of X n , I X Y , ϕ n , and I X Y n + 1 . I X Y n and the target domain image I Y n are input into the discriminator D Y n to obtain discriminant results for model training, where 0 n < N .
G X Y n and G Y X n share the same architecture but with different weights. The working principle of G X Y n is as follows:
I X Y , ϕ n = ϕ I X n ,
X n = A n I X Y , ϕ n , I X n , I X Y n + 1 ,
I X Y n = X n I X Y n + 1 + 1 X n I X Y , ϕ n ,
where 0 n < N , ⊗ represents pixel-level multiplication. At the lowest scale N, I X Y n + 1 is replaced with an image with pixel values of 0. First, SID-AM-MSITM uses the encoder ϕ to preprocess I X n to I X Y , ϕ n . Then, SID-AM-MSITM uses the encoder A n to generate mask X n . Finally, SID-AM-MSITM uses the linear combination to obtain output I X Y n .
Similarly, the implementation of the translation of I Y I Y X at scale n is as follows:
I Y X , ϕ n = ϕ I Y n ,
Y n = A n I Y X , ϕ n , I Y n , I Y X n + 1 ,
I Y X n = Y n I Y X n + 1 + 1 Y n I Y X , ϕ n ,
where 0 n < N . At the lowest scale N, I Y X n + 1 is also replaced with an image with pixel values of 0.
In this way, the generators focus on the regions that synthesize current scale details in the images. Meanwhile, it maintains the previously learned global structures as unaffected.

2.2. Style-Independent Discriminators

Next, we present our proposed style-independent discriminators, which focus on the images’ content information rather than their style information, so as to enable SIM-AM-MSITM to avoid losing the content details in non-underwater images.
When two sets of images share the same content but different styles, it is ideal to make discriminators give similar discriminant results to the images. Thus, SID-AM-MSITM uses instance-level as well as vector-level style difference losses to train style-independent discriminators.
Figure 4 illustrates style and content differences. As illustrated in the figure, the first two generated images share the same content information, while the latter two share the same style information. s x and s y represent the style information of images x and y respectively. c x and c y represent the content information of images x and y, respectively. G s , c indicates the image generated using style information s and content information c. D G s , c represents the discriminant result of G s , c given by a discriminator.
Figure 5 shows a style-independent discriminator D Y n of SIM-AM-MSITM, which requires a source domain image I X n , a generation image at current scale I X Y n , an instance-level mixed style image I ^ X n , and a vector-level mixed style image I X n to make the discriminator D Y n style-independent.
Discriminator D X n has a similar structure and requires a target domain image I Y n , a generation image at current scale I Y X n , an instance-level mixed style images I ^ Y n , and a vector-level mixed style image I Y n , where 0 n N .
Then, we will describe how style-independent discriminators eliminate style differences at the instance level as well as the vector level, respectively. Instance-level style difference refers to the style difference between the image obtained by stylizing its pixels and the original image. Vector-level style difference refers to the style difference between the image obtained by stylizing its encoded vectors and the original image.

2.2.1. Instance-Level Style-Independent Discriminators

Instance-level style-independent discriminators use a special regularization term, so as to reduce style differences between the images obtained by stylizing its pixels and the original images.
We first adjust weight α to gradually increase the proportion of generated images among the mixed ones at multi-scales. Then, we make discriminators to reduce the differences between the original non-underwater images or the underwater images and the final mixed images, which are constrained by a consistency loss. Such progress is formulated as follows:
I ^ X n = α I X n + ( 1 α ) I X Y n , 0 n N ,
L c o n = D Y n I X n D Y n I ^ X n 1 ,
I ^ Y n = α I Y n + ( 1 α ) I Y X n , 0 n N ,
L c o n = D X n I Y n D X n I ^ Y n 1 ,
where 0 n N , 0 α < 1 . I X n and I Y n represent the source-domain and target-domain images at the current scale, respectively. I X Y n and I Y X n denote the generation images at the current scale. α indicates the weight of the linear combination and gradually becomes smaller as the scale rises. L c o n uses an L 1 paradigm to process the instance-level style-independent loss, and D · represents the discriminant results of discriminators.
As the scale rises, the style of instance-level mixed-style images I ^ X n is closer to the target-domain ones and away from the source-domain ones. The style of instance-level mixed-style images I ^ Y n is closer to the source-domain ones and away from the target-domain ones. Discriminators penalize the distances between the source-domain image outputs or the target-domain image outputs and the mixed-style image outputs.
Figure 6 shows the training process of instance-level style-independent discriminators D Y . ↑ and ↓ indicate the value rising and descending as the scale rises, respectively.

2.2.2. Vector-Level Style-Independent Discriminators

Based on the above instance-level discriminators, it is not enough to generate images since this is limited by style-independent pixels. Therefore, we devise vector-level style-independent discriminators that further mix the encoded vectors of the source-domain and target-domain images at each scale. We put the mixed encoded vectors into a decoder, and utilize its generated images as well as the source-domain ones or the target-domain ones for model training.
First, we encode the source domain-images and the target-domain images, respectively using VGG 19 [29] and then process them using AdaIN [30]. The results obtained are linearly combined with the encoded vectors of the source- or target-domain images. Then, we put the results of the linear combinations into a decoder to get the vector-level mixed-style images I X n and I Y n . The decoder is a convolutional network that is symmetric to VGG 19 and upsamples the mixed encoded vectors into images. Finally, we utilize discriminators to penalize the distances between the source-domain image outputs or the target-domain image outputs and the mixed-style image outputs. Such progress is formulated as follows:
I X n = D e c o d e r α E n c o d e r I X n + ( 1 α ) A d a I N E n c o d e r ( I X n ) , E n c o d e r ( I Y n ) ,
L c o n = D Y I X n D Y I X n 1 ,
I Y n = D e c o d e r α E n c o d e r I Y n + ( 1 α ) A d a I N E n c o d e r ( I Y n ) , E n c o d e r ( I X n ) ,
L c o n = D X I Y n D X I Y n 1 ,
where 0 n N , 0 α < 1 . I X n and I Y n represent the source-domain and target-domain images at the current scale, respectively. E n c o d e r · is VGG 19 and D e c o d e r · is symmetric to VGG 19. α is the weight coefficient that becomes smaller as the scale rises. L c o n uses an L 1 paradigm to process the vector-level style-independent loss. D · represents the discriminant results of discriminators.
Figure 7 shows the training process of vector-level style-independent discriminators. ↑ and ↓ indicate the value rising and descending as the scale rises, respectively.

2.3. Implementation

To implement SID-AM-MSITM, we utilize Adam [31] as its optimizer and LeakyReLU [32] as its activation function. At the lowest scale, images of the model are 100 × 100-pixel ones. And at the highest scale, the size of images is 250 × 250 pixels. The model uses 6 scales.

3. Experiment and Result Analysis

In this section, we will present our experiments and the corresponding results analysis. We first introduce the evaluation metrics and then present the ablation and comparative experiments, respectively.

3.1. Evaluation Metric

We utilize four metrics for quantitative evaluation, including peak signal-to-noise ratio (PSNR) [33], structure similarity index measure (SSIM) [34], information entropy (Entropy) [35], and single image Fréchet Inception distance (SIFID) [36].
(1)
PSNR: PSNR measures the distance between the distributions of two images. We use PSNR to calculate the distance between the source-domain images and reconstructed images. A larger PSNR value indicates a smaller difference between the two images.
(2)
SSIM: SSIM measures the similarity of two images. The value of SSIM is between 0 and 1, and a larger SSIM indicates a better reconstruction effect, which suggests the translation effect of an image translation model.
(3)
Entropy: Information entropy measures the complexity of an image. Larger information entropy indicates complex images that contain more information.
(4)
SIFID: Single Image Fréchet Inception Distance (SIFID) is a special type of Fréchet Inception distance (FID) [37]. It measures the deviation between the feature distribution of two single images, and smaller SIFID indicates the better effect of generated images.

3.2. Ablation Experiment

We use five different datasets, including submarine, underwater optics, sunken ship, crashed plane, and underwater sonar datasets. The underwater optics images are from the URPC2020 dataset [38]. These images are difficult to collect and are not large in number. Specifically, the submarine, sunken ship, and crashed plane images are content categories, and the underwater sonar and underwater optics images are style ones. Meanwhile, the submarine, sunken ship, and crashed plane images are non-underwater images, and the underwater sonar and underwater optics images are underwater ones.
Figure 8 presents the utilized datasets, which are grouped into six different combinations, i.e., (1) sunken ships with underwater sonars, (2) sunken ships with underwater optics, (3) crashed planes with underwater sonars, (4) crashed planes with underwater optics, (5) submarines with underwater sonars, and (6) submarines with underwater optics.
Based on the above six combinations of images, we perform ablation experiments on four modules, including (1) TuiGAN only, (2) TuiGAN with CBAM, (3) TuiGAN with style-independent discriminators, and (4) SID-AM-MSITM, so as to comprehensively evaluate the improvements we have made.
Figure 9, Figure 10, Figure 11, Figure 12, Figure 13 and Figure 14, respectively, present the generated and reconstructed images based on different combinations of datasets. In the figures, the first-column and second-column images are source-domain and target-domain ones, respectively. The third to the tenth columns, respectively, present the images generated or reconstructed using TuiGAN, TuiGAN with CBAM modules, TuiGAN with style-independent discriminators, and SID-AM-MSITM.
It is observed that SID-AM-MSITM has the ability to translate non-underwater images into underwater sonar images and underwater optics images, and objects such as submarines, crashed planes, and sunken ships are evident in underwater images. Meanwhile, SID-AM-MSITM is also capable of reconstructing translated underwater images into non-underwater images with little difference from the original non-underwater images.
Table 1 presents the PSNR results of each model. It is observed that the PSNR values of the images reconstructed using the TuiGAN with CBAM modules on the three combinations of the datasets are higher than that of TuiGAN, and the maximum difference reaches 8.63. The PSNR values of the images reconstructed using TuiGAN with style-independent discriminators on the five combinations of datasets are higher than that of TuiGAN, and the maximum difference reaches 4.7. The PSNR values of the images reconstructed using SID-AM-MSITM on all combinations of the datasets are not less than that of TuiGAN, and the maximum difference reaches 4.58. The promising results indicate that the combination of CBAM modules and style-independent discriminators significantly improves the effective information acquisition ability of backbone TuiGAN and is suitable for all combinations of datasets.
Table 2 presents the SSIM results of each model. SSIM is also a metric to measure the effect of model reconstruction. It is observed that the SSIM results of the images reconstructed using TuiGAN with CBAM modules on four combinations of datasets are higher than that of TuiGAN, and the maximum difference reaches 0.17. The SSIM results of the images reconstructed using TuiGAN with style-independent discriminators on all combinations of datasets are not less than that of TuiGAN, and the maximum difference reaches 0.09. The SSIM values of the images reconstructed using SID-AM-MSITM on all combinations of datasets are higher than that of TuiGAN, and the maximum difference reaches 0.15. These promising results also indicate that CBAM modules and style-independent discriminators improve the ability to access effective information.
Table 3 presents the Entropy results of each model. It is observed that the Entropy result of the images generated using TuiGAN with CBAM modules on only one combination of datasets is higher than that of TuiGAN, and the difference reaches 0.21. The Entropy results of the images generated using TuiGAN with style-independent discriminators on four combinations of datasets are higher than that of TuiGAN, and the maximum difference reaches 0.86. The Entropy results of the images reconstructed using SID-AM-MSITM on all combinations of datasets are higher than that of TuiGAN, and the maximum difference reaches 0.78. These promising results indicate that style-independent discriminators improve the diversity of generated images. The style-independent discriminators improve TuiGAN’s ability to retain content details and are suitable for the combinations of all datasets.
In summary, based on the above ablation results, our proposed SID-AM-MSITM achieves promising underwater image translation performance in terms of PSNR, SSIM, and Entropy. The ablation experiments demonstrate that CBAM modules enhance the feature extraction ability of the network, so as to enhance the ability to access effective information. Moreover, we prove that style-independent discriminators improve the diversity of the generated images without weakening the reconstruction performance, which indicates SID-AM-MSITM retains the content details of non-underwater images.

3.3. Comparative Experiment

The above ablation experiments demonstrate the overall effect of SID-AM-MSITM. In the following, we further compare it with multiple advanced image translation models, including CycleGAN [39], FUNIT [40], AdaIN [30], and SinDiffusion [41]. These models are selected as baselines since they present promising performance and cover general image translation models, CycleGAN, FUNIT, and AdaIN, as well as the emerging SinDiffusion.
(1)
CycleGAN: CycleGAN is one of the most typical translation models using cycle consistency. The model assumes the potential correspondence between source-domain and target-domain images.
(2)
FUNIT: FUNIT is an unsupervised few-shot image translation model that achieves satisfactory performance based on limited data.
(3)
AdaIN: AdaIN is an image translation model that achieves real-time and arbitrary style transfer.
(4)
SinDiffusion: SinDiffusion is a diffusion model that works on a single natural image.
Figure 15 presents the comparison results between the images translated using SID-AM-MS- ITM and other baseline models. Through visual effect comparison, it is observed that SID-AM-MSITM has learned the style of target-domain images and retains the content of source-domain images. Moreover, the images translated using SID-AM-MSITM show little difference between adjacent pixels as well as excellent smoothness, which is superior to CycleGAN which shows obvious adjacent pixels difference after amplification. Compared with FUNIT and AdaIN, SID-AM-MSITM retains the content of source-domain images and learns better texture and color information from the target domain.
Next, we use SIFID to quantitatively compare SID-AM-MSITM with these baselines. Table 4 shows the SIFID results. It is observed that SID-AM-MSITM achieves the best (smallest) SIFID values. For example, the SIFID values of the images translated using SID-AM-MSITM are roughly 0.02 × 10 2 to 0.058 × 10 2 smaller than that of CycleGAN, 17.408 × 10 2 to 17.55 × 10 2 smaller than that of FUNIT, 9.31 × 10 2 to 9.418 × 10 2 smaller than that of AdaIN, 0.002 × 10 2 to 0.018 × 10 2 smaller than that of TuiGAN, and 1.118 × 10 2 to 1.25 × 10 2 smaller than that of SinDiffusion. This demonstrates that the images translated using SID-AM-MSITM are closer to the source-domain images and retain more content information than other models.
In summary, SID-AM-MSITM is superior to multiple baseline models in improving the ability to access effective information and avoiding the loss of content details.

4. Conclusions

In this work, we propose a novel multi-scale image translation model with attention modules and style-independent discriminators (SID-AM-MSITM), to complete the underwater image translation task. We use a multi-scale generative adversarial network, TuiGAN, to construct a backbone architecture, which translates images from low scales to high scales. We introduce CBAM modules into the generators and discriminators at multi-scales and devise style-independent discriminators to improve the generative and discriminant effects. Based on systematical ablation and comparative experiments, we demonstrate that SID-AM-MSITM has the ability to acquire effective information and retain the content details of non-underwater images during the underwater image translation process, and it requires only two unpaired images to complete the image translation.
However, there are still some problems in the current research. In the use of style-independent discriminators, SID-AM-MSITM uses the number between 0 and 1 in the linear combination to achieve the translation from the source domain to the target domain. We will continue to study whether there is a more appropriate interval to train style-independent discriminators. We only select several source-domain and target-domain images as the dataset, which has certain limitations. In order to measure the performance of the model comprehensively, we will use other underwater target images to verify the versatility of SID-AM-MSITM, such as the UIEB database [42].

Author Contributions

Conceptualization, D.Y., T.Z. and M.L.; Data curation, M.L. and X.W.; Formal analysis, T.Z. and M.L.; Funding acquisition, D.Y. and X.W.; Investigation, M.L., W.C. and X.L.; Methodology, D.Y., T.Z., M.L. and X.W.; Project administration, B.L. and X.W.; Resources, M.L. and X.W.; Software, M.L. and W.C.; Supervision, B.L. and X.W.; Validation, T.Z. and M.L.; Visualization, M.L.; Writing—original draft, D.Y. and M.L.; Writing—review and editing, T.Z., B.L., W.C., X.L. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by a grant from the Key Laboratory of Avionics System Integrated Technology, the Fundamental Research Funds for the Central Universities in China, Grant No. 3072022JC0601, and the Ministry of Industry and Information Technology High-tech Ship Project [2019] No. 331.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available from the corresponding author upon reasonable request.

Acknowledgments

The authors are grateful to the editors and anonymous reviewers for their constructive comments.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SID-AM-MSITMmulti-scale image translation model based on style-independent discriminators and attention modules
CBAMconvolution block attention module
GANgenerative adversarial network
CNNconvolutional neural network
LeakyReLUleaky rectified linear unit
CycleGANcycle-consistent adversarial network
TV losstotal variation loss
PSNRpeak signal-to-noise ratio
SSIMstructure similarity index measure
Entropyinformation entropy
AdaINadaptive instance normalization
CycleGANcycle-consistent adversarial networks
FIDthe Fréchet Inception distance
SIFIDsingle image Fréchet Inception distance

References

  1. Zhao, Y.; Zhu, K.; Zhao, T.; Zheng, L.; Deng, X. Small-Sample Seabed Sediment Classification Based on Deep Learning. Remote. Sens. 2023, 15, 2178. [Google Scholar] [CrossRef]
  2. Chen, B.; Li, R.; Bai, W.; Zhang, X.; Li, J.; Guo, R. Research on recognition method of optical detection image of underwater robot for submarine cable. In Proceedings of the 2019 IEEE 3rd Advanced Information Management, Communicates, Electronic and Automation Control Conference (IMCEC), Chongqing, China, 11–13 October 2019; pp. 1973–1976. [Google Scholar]
  3. Teng, B.; Zhao, H. Underwater target recognition methods based on the framework of deep learning: A survey. Int. J. Adv. Robot. Syst. 2020, 17, 1729881420976307. [Google Scholar] [CrossRef]
  4. Cruz, L.; Lucio, D.; Velho, L. Kinect and rgbd images: Challenges and applications. In Proceedings of the IEEE 2012 25th SIBGRAPI Conference on Graphics, Patterns and Images Tutorials, Ouro Preto, Brazil, 22–25 August 2012; pp. 36–49. [Google Scholar]
  5. Yang, L.; Wang, B.; Zhang, R.; Zhou, H.; Wang, R. Analysis on location accuracy for the binocular stereo vision system. IEEE Photonics J. 2017, 10, 1–16. [Google Scholar] [CrossRef]
  6. Lin, E. Comparative Analysis of Pix2Pix and CycleGAN for Image-to-Image Translation. Highlights Sci. Eng. Technol. 2023, 39, 915–925. [Google Scholar] [CrossRef]
  7. Multi-view underwater image enhancement method via embedded fusion mechanism. Eng. Appl. Artif. Intell. 2023, 121, 105946. [CrossRef]
  8. Zhou, J.; Liu, Q.; Jiang, Q.; Ren, W.; Lam, K.M.; Zhang, W. Underwater camera: Improving visual perception via adaptive dark pixel prior and color correction. Int. J. Comput. Vis. 2023. [Google Scholar] [CrossRef]
  9. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, USA, 8–13 December 2014; Volume 27. [Google Scholar]
  10. Hertzmann, A.; Jacobs, C.E.; Oliver, N.; Curless, B.; Salesin, D.H. Image analogies. In Proceedings of the 28th Annual Conference on Computer Graphics and Interactive Techniques, Los Angeles, CA, USA, 12–17 August 2001; pp. 327–340. [Google Scholar]
  11. Resales; Achan; Frey. Unsupervised image translation. In Proceedings of the 9th IEEE International Conference on Computer Vision, Nice, France, 14–17 October 2003; pp. 472–478. [Google Scholar]
  12. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  13. Gatys, L.A.; Ecker, A.S.; Bethge, M. Image style transfer using convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2414–2423. [Google Scholar]
  14. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-image translation with conditional adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  15. Wang, T.C.; Liu, M.Y.; Zhu, J.Y.; Tao, A.; Kautz, J.; Catanzaro, B. High-resolution image synthesis and semantic manipulation with conditional gans. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8798–8807. [Google Scholar]
  16. Karras, T.; Laine, S.; Aila, T. A style-based generator architecture for generative adversarial networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4401–4410. [Google Scholar]
  17. Choi, Y.; Choi, M.; Kim, M.; Ha, J.W.; Kim, S.; Choo, J. Stargan: Unified generative adversarial networks for multi-domain image-to-image translation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8789–8797. [Google Scholar]
  18. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised generative network to enable real-time color correction of monocular underwater images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef]
  19. Wang, N.; Zhou, Y.; Han, F.; Zhu, H.; Yao, J. UWGAN: Underwater GAN for real-world underwater color restoration and dehazing. arXiv 2019, arXiv:1912.10269. [Google Scholar]
  20. Li, N.; Zheng, Z.; Zhang, S.; Yu, Z.; Zheng, H.; Zheng, B. The synthesis of unpaired underwater images using a multistyle generative adversarial network. IEEE Access 2018, 6, 54241–54257. [Google Scholar] [CrossRef]
  21. Zhou, J.; Li, B.; Zhang, D.; Yuan, J.; Zhang, W.; Cai, Z.; Shi, J. UGIF-Net: An Efficient Fully Guided Information Flow Network for Underwater Image Enhancement. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 1–17. [Google Scholar] [CrossRef]
  22. Geirhos, R.; Rubisch, P.; Michaelis, C.; Bethge, M.; Wichmann, F.A.; Brendel, W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. arXiv 2018, arXiv:1811.12231. [Google Scholar]
  23. Lin, J.; Pang, Y.; Xia, Y.; Chen, Z.; Luo, J. Tuigan: Learning versatile image-to-image translation with two unpaired images. In Proceedings of the 16th European Conference of Computer Vision (ECCV 2020), Glasgow, UK, 23–28 August 2020; Proceedings, Part IV 16. Springer: Cham, Switzerland, 2020; pp. 18–35. [Google Scholar]
  24. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  25. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved training of wasserstein gans. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  26. Mahendran, A.; Vedaldi, A. Understanding deep image representations by inverting them. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5188–5196. [Google Scholar]
  27. You, Q.; Wan, C.; Sun, J.; Shen, J.; Ye, H.; Yu, Q. Fundus image enhancement method based on CycleGAN. In Proceedings of the 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Berlin, Germany, 23–27 July 2019; pp. 4500–4503. [Google Scholar]
  28. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  29. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  30. Huang, X.; Belongie, S. Arbitrary style transfer in real-time with adaptive instance normalization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 1501–1510. [Google Scholar]
  31. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  32. Xu, J.; Li, Z.; Du, B.; Zhang, M.; Liu, J. Reluplex made more practical: Leaky ReLU. In Proceedings of the 2020 IEEE Symposium on Computers and Communications (ISCC), Rennes, France, 7–10 July 2020; pp. 1–7. [Google Scholar]
  33. Najafipour, A.; Babaee, A.; Shahrtash, S.M. Comparing the trustworthiness of signal-to-noise ratio and peak signal-to-noise ratio in processing noisy partial discharge signals. IET Sci. Meas. Technol. 2013, 7, 112–118. [Google Scholar] [CrossRef]
  34. Khadtare, M.S. GPU based image quality assessment using structural similarity (SSIM) index. In Emerging Research Surrounding Power Consumption and Performance Issues in Utility Computing; IGI Global: Hershey, PA, USA, 2016; pp. 276–282. [Google Scholar]
  35. Xu, N.; Zhuang, J.; Xiao, J.; Peng, C. Regional Differential Information Entropy for Super-Resolution Image Quality Assessment. arXiv 2021, arXiv:2107.03642. [Google Scholar]
  36. Shaham, T.R.; Dekel, T.; Michaeli, T. Singan: Learning a generative model from a single natural image. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 4570–4580. [Google Scholar]
  37. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  38. Zhang, J.; Zhang, J.; Zhou, K.; Zhang, Y.; Chen, H.; Yan, X. An Improved YOLOv5-Based Underwater Object-Detection Framework. Sensors 2023, 23, 3693. [Google Scholar] [CrossRef] [PubMed]
  39. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired image-to-image translation using cycle-consistent adversarial networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  40. Liu, M.Y.; Huang, X.; Mallya, A.; Karras, T.; Aila, T.; Lehtinen, J.; Kautz, J. Few-shot unsupervised image-to-image translation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 10551–10560. [Google Scholar]
  41. Wang, W.; Bao, J.; Zhou, W.; Chen, D.; Chen, D.; Yuan, L.; Li, H. Sindiffusion: Learning a diffusion model from a single natural image. arXiv 2022, arXiv:2211.12445. [Google Scholar]
  42. Zhou, J.; Pang, L.; Zhang, D.; Zhang, W. Underwater Image Enhancement Method via Multi-Interval Subhistogram Perspective Equalization. IEEE J. Ocean. Eng. 2023, 48, 474–488. [Google Scholar] [CrossRef]
Figure 1. The architecture of SID-AM-MSITM (the process of translating a non-underwater image into an underwater image).
Figure 1. The architecture of SID-AM-MSITM (the process of translating a non-underwater image into an underwater image).
Jmse 11 01929 g001
Figure 2. The structure of CBAM modules. CBAM represents the Convolution Block Attention Module. Conv represents convolution operations.
Figure 2. The structure of CBAM modules. CBAM represents the Convolution Block Attention Module. Conv represents convolution operations.
Jmse 11 01929 g002
Figure 3. Generator and discriminator of SID-AM-MSITM (the process of translating a source domain image I X n to a generated image I X Y n ).
Figure 3. Generator and discriminator of SID-AM-MSITM (the process of translating a source domain image I X n to a generated image I X Y n ).
Jmse 11 01929 g003
Figure 4. The illustration of style and content differences.
Figure 4. The illustration of style and content differences.
Jmse 11 01929 g004
Figure 5. The training process of SID-AM-MSITM’s style-independent discriminator D Y n .
Figure 5. The training process of SID-AM-MSITM’s style-independent discriminator D Y n .
Jmse 11 01929 g005
Figure 6. The training process of instance-level style-independent discriminators.
Figure 6. The training process of instance-level style-independent discriminators.
Jmse 11 01929 g006
Figure 7. The training process of vector-level style-independent discriminators.
Figure 7. The training process of vector-level style-independent discriminators.
Jmse 11 01929 g007
Figure 8. Datasets.
Figure 8. Datasets.
Jmse 11 01929 g008
Figure 9. Generated and reconstructed images using the sunken ships and the underwater sonars.
Figure 9. Generated and reconstructed images using the sunken ships and the underwater sonars.
Jmse 11 01929 g009
Figure 10. Generated and reconstructed images using the sunken ships and the underwater optics.
Figure 10. Generated and reconstructed images using the sunken ships and the underwater optics.
Jmse 11 01929 g010
Figure 11. Generated and reconstructed images using the crashed planes and the underwater sonars.
Figure 11. Generated and reconstructed images using the crashed planes and the underwater sonars.
Jmse 11 01929 g011
Figure 12. Generated and reconstructed images using the crashed planes and the underwater optics.
Figure 12. Generated and reconstructed images using the crashed planes and the underwater optics.
Jmse 11 01929 g012
Figure 13. Generated and reconstructed images using the submarines and the underwater sonars.
Figure 13. Generated and reconstructed images using the submarines and the underwater sonars.
Jmse 11 01929 g013
Figure 14. Generated and reconstructed images using the submarines and the underwater optics.
Figure 14. Generated and reconstructed images using the submarines and the underwater optics.
Jmse 11 01929 g014
Figure 15. The comparison results of the images generated using SID-AM-MSITM and other models.
Figure 15. The comparison results of the images generated using SID-AM-MSITM and other models.
Jmse 11 01929 g015
Table 1. Ablation experiments on different datasets (PSNR).
Table 1. Ablation experiments on different datasets (PSNR).
CategoryTuiGANTuiGAN with CBAM ModulesTuiGAN with Style-Independent DiscriminatorsSID-AM-MSITM
Sunken Ship + Sonar17.2020.2819.5322.77
Sunken Ship + Optics21.2418.9119.9822.87
Crashed plane + Sonar20.0223.0020.5024.26
Crashed plane + Optics26.4625.0429.1726.46
Submarine + Sonar27.2725.8831.1131.85
Submarine + Optics23.2031.8327.9025.96
Table 2. Ablation experiments on different datasets (SSIM).
Table 2. Ablation experiments on different datasets (SSIM).
CategoryTuiGANTuiGAN with CBAM ModulesTuiGAN with Style-Independent DiscriminatorsSID-AM-MSITM
Sunken Ship + Sonar0.680.820.750.83
Sunken Ship + Optics0.850.880.870.90
Crashed plane + Sonar0.830.910.830.90
Crashed plane + Optics0.910.890.920.92
Submarine + Sonar0.850.840.880.89
Submarine + Optics0.700.870.790.84
Table 3. Ablation experiments on different datasets (Entropy).
Table 3. Ablation experiments on different datasets (Entropy).
CategoryTuiGANTuiGAN with CBAM ModulesTuiGAN with Style-Independent DiscriminatorsSID-AM-MSITM
Sunken Ship + Sonar6.756.707.127.00
Sunken Ship + Optics6.006.216.866.78
Crashed plane + Sonar7.316.537.237.38
Crashed plane + Optics5.575.535.705.63
Submarine + Sonar6.475.546.966.59
Submarine + Optics6.326.295.446.33
Table 4. SIFID comparison of underwater images generated using SID-AM-MSITM and other baseline models ( × 10 2 ) .
Table 4. SIFID comparison of underwater images generated using SID-AM-MSITM and other baseline models ( × 10 2 ) .
Our ModelCycleGANFUNITAdaINTuiGANSinDiffusion
0.0920.13017.59.510.1011.21
0.0540.08017.69.410.0721.23
0.0500.10817.69.360.0521.30
0.0150.03517.59.350.0171.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, D.; Zhang, T.; Li, B.; Li, M.; Chen, W.; Li, X.; Wang, X. Underwater Image Translation via Multi-Scale Generative Adversarial Network. J. Mar. Sci. Eng. 2023, 11, 1929. https://doi.org/10.3390/jmse11101929

AMA Style

Yang D, Zhang T, Li B, Li M, Chen W, Li X, Wang X. Underwater Image Translation via Multi-Scale Generative Adversarial Network. Journal of Marine Science and Engineering. 2023; 11(10):1929. https://doi.org/10.3390/jmse11101929

Chicago/Turabian Style

Yang, Dongmei, Tianzi Zhang, Boquan Li, Menghao Li, Weijing Chen, Xiaoqing Li, and Xingmei Wang. 2023. "Underwater Image Translation via Multi-Scale Generative Adversarial Network" Journal of Marine Science and Engineering 11, no. 10: 1929. https://doi.org/10.3390/jmse11101929

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop