Next Article in Journal
Long-Acting Real-Time Microscopic Monitoring Inside the Proton Exchange Membrane Water Electrolyzer
Next Article in Special Issue
Automatic Suppression Method for Water Surface Glints Using a Division of Focal Plane Visible Polarimeter
Previous Article in Journal
TFC-GCN: Lightweight Temporal Feature Cross-Extraction Graph Convolutional Network for Skeleton-Based Action Recognition
Previous Article in Special Issue
Freeform Wide Field-of-View Spaceborne Imaging Telescope: From Design to Demonstrator
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Underwater Target Detection Utilizing Polarization Image Fusion Algorithm Based on Unsupervised Learning and Attention Mechanism

1
College of Engineering, Ocean University of China, Qingdao 266100, China
2
Key Laboratory for Micro/Nano Technology and System of Liaoning Province, Dalian University of Technology, Dalian 116024, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(12), 5594; https://doi.org/10.3390/s23125594
Submission received: 17 April 2023 / Revised: 23 May 2023 / Accepted: 13 June 2023 / Published: 15 June 2023
(This article belongs to the Special Issue Optical Imaging and Sensing)

Abstract

:
Since light propagation in water bodies is subject to absorption and scattering effects, underwater images using only conventional intensity cameras will suffer from low brightness, blurred images, and loss of details. In this paper, a deep fusion network is applied to underwater polarization images; that is, the underwater polarization images are fused with intensity images using the deep learning method. To construct a training dataset, we establish an experimental setup to obtain underwater polarization images and perform appropriate transformations to expand the dataset. Next, an end-to-end learning framework based on unsupervised learning and guided by an attention mechanism is constructed for fusing polarization and light intensity images. The loss function and weight parameters are elaborated. The produced dataset is used to train the network under different loss weight parameters, and the fused images are evaluated based on different image evaluation metrics. The results show that the fused underwater images are more detailed. Compared with light intensity images, the information entropy and standard deviation of the proposed method increase by 24.48% and 139%. The image processing results are better than other fusion-based methods. In addition, the improved U-net network structure is used to extract features for image segmentation. The results show that the target segmentation based on the proposed method is feasible under turbid water. The proposed method does not require manual adjustment of weight parameters, has faster operation speed, and has strong robustness and self-adaptability, which is important for research in vision fields, such as ocean detection and underwater target recognition.

1. Introduction

The ocean covers more than 70% of the earth’s total area. The marine ecosystem is one of the most productive and dynamic ecosystems on earth. Many scholars have conducted research in marine resource exploration, biological investigation, underwater vehicle navigation, and other fields [1,2,3,4]. Underwater optical images are one of the important media for exploring the ocean at present. However, due to the influence of a large number of floating particles in the water, the actual underwater images are seriously degraded, with problems such as high background noise, low contrast, and loss of details [1]. Therefore, the research of underwater image enhancement technology is of great significance and value for ocean exploration.
Researchers have shown that underwater polarization imaging technology can reduce the influence of backscattered light on underwater imaging to a certain extent by using the polarization characteristics of scattered light [5,6]. The degree of linear polarization (DoLP) images is used to characterize polarization characteristics and provide detailed features, which provide complementary scene information from different aspects. In order to obtain more complete information about the same scene, the intensity and DoLP images need to be fused. In the past decade, many image fusion methods have emerged in the field of image processing. Image fusion methods can be divided into different levels according to the fusion process: pixel level, feature level, and decision level. Among them, pixel-level image fusion has the most research and application. It can be divided into fusion methods based on the transform domain (image pyramid [7], wavelet transform [8], Ridgelets [9], curvelet transform [10]) and fusion methods based on the spatial domain (HIS transform [11], principal component analysis inverse transform [12], pixel value weighting, statistical model). These methods can realize multisource image fusion from the pixel level. The information on the fused image is more comprehensive, accurate, and reliable.
In recent years, deep learning technology has developed rapidly and made great breakthroughs in many problems in the field of computer vision and image processing. At present, there is more and more research on image fusion technology based on deep learning in digital imaging (multifocus image fusion [13], multiple exposure image fusion [14], etc.), multimodal imaging (visible light and infrared image fusion [15], medicine multimodal image fusion [16], polarization image fusion [17], etc.), remote sensing imaging [18] (multispectral and panchromatic image fusion, multispectral and hyperspectral image fusion, etc.), and other directions. Compared with the traditional methods, which have the limited ability to characterize the complex mapping relationship between the input image and the target image, the deep learning model has a strong ability to characterize the complex relationship between different signals. It can automatically extract the most effective features from the data to solve the difficulties of manual design. However, the polarization image fusion method based on deep learning has not been reported so far, which is one of the main contributions of this paper.
In this paper, the fusion network is applied to underwater polarization image processing, and a network model guided by unsupervised learning and attention mechanism is proposed. It is mainly divided into three modules: feature extraction, feature fusion, and image reconstruction. The feature extraction module incorporates an attention mechanism. The relevant loss function and weight parameters are constructed. The experimental results show that the proposed method can effectively fuse underwater light intensity and polarization image information and improve the quality of visual imaging. Compared with other traditional fusion algorithms, the proposed method has strong robustness and adaptability, which is of great significance for the application research of marine detection, underwater target recognition, and other visual fields.

2. Underwater Polarization Imaging

2.1. Underwater Imaging Model

The Jaffe–McGlamery model [19,20] is one of the commonly used underwater imaging models, as shown in Figure 1. Many underwater image restoration algorithms are proposed based on this model. The Jaffe–McGlamery model states that the final image I(x, y) received by the detector is a linear combination of three components: the target reflected light S(x, y) received by the detector, the backscattered light B(x, y) scattered by the water body before the light source reaches the target, and the forward scattered light F(x, y) scattered by part of the target reflected light reaching the detector through the water body. The image can be expressed as
I ( x , y ) = S ( x , y ) + B ( x , y ) + F ( x , y ) .
The initial irradiance of the target is assumed to be J(x, y), and part of the energy is lost due to scattering and absorption when the light propagates from the target to the detector. Therefore, the reflected light S(x, y) of the target can be expressed as
S ( x , y ) = J ( x , y ) t ( x , y ) ,
t(x, y) is the transmittance of the medium, which is determined by the attenuation coefficient β ( x , y ) and the propagation distance ρ ( x , y ) . In a single uniform medium, the attenuation coefficient β ( x , y ) is invariant in space, so β ( x , y ) = β 0 . The propagation distance refers to the underwater part of the optical path between the object and the camera:
t ( x , y ) = e β ( x , y ) ρ ( x , y ) .
Backscattered light B(x, y) is the background light scattered by water particles to the detector. It can be expressed as
B ( x , y ) = B ( 1 t ( x , y ) ) ,
where B represents the underwater ambient light intensity at the infinite distance. Since the effect of forward scattering on imaging quality is minimal, its effect is usually ignored, so Equation (1) can be simplified as
I ( x , y ) = J ( x , y ) t ( x , y ) + B ( 1 t ( x , y ) ) .
Thus, the initial irradiance J(x, y) of the object can be expressed as
J ( x , y ) = I ( x , y ) B ( 1 t ( x , y ) ) t ( x , y ) .

2.2. Polarization Imaging Model

The Stokes vector method is one of the most commonly used polarization characterization methods in the field of polarization detection. This method can fully characterize the polarization characteristics of the light wave. The vector is composed of four parameters:
S = I Q U V = I 0 ° + I 90 ° I 0 ° I 90 ° I 45 ° I 135 ° I r I l .
S represents the Stokes vector, which is a 4 × 1 column vector composed of four parameters: I, Q, U, and V. I represents the total light intensity received by the detector, Q represents the light intensity difference between the 0° and 90° polarization components I and I90°, U represents the light intensity difference between the 45° and 135° polarization components I45° and I135°, and V represents the intensity difference between the left- and right-handed circularly polarized components Ir and Il. The emergent light S′ = [I′, Q′, U′, V′]T can be obtained by the Mueller matrix:
S = I Q U V = 1 2 1 cos 2 θ sin 2 θ 0 cos 2 θ cos 2 2 θ cos 2 θ sin 2 θ 0 sin 2 θ cos 2 θ sin 2 θ sin 2 2 θ 0 0 0 0 0 I Q U V .
θ is the included angle between the main optical axis and the reference line. S′ represents the outgoing light with angle θ. According to Equation (8), the intensity of the outgoing light with angle θ can be obtained as follows:
I θ = 1 2 I + Q cos 2 θ + U sin 2 θ .
The polarization camera can obtain the light intensity image of the polarization directions of 0°, 45°, 90°, and 135° because each pixel of the CMOS sensor is placed with four polarizers of different angles (0°, 45°, 90°, and 135°), as shown in Figure 2. Every four pixels is set as a computing unit. Then, the light intensity of 0°, 45°, 90°, and 135° and the Stokes vector of the light can be obtained simultaneously. The Stokes vector can be used to further calculate the DoLP and angle of polarization of the incident light:
D o L P = Q 2 + U 2 I ,
φ = 1 2 arctan U Q .
DoLP represents the proportion of linearly polarized components in the total light intensity. The angle of polarization refers to the dominant polarization direction of the incident light.

3. Deep Learning Method

3.1. Network Architecture

The network structure adopted in this paper is shown in Figure 3, which mainly consists of three modules: feature extraction, feature fusion, and image reconstruction. First, in the feature extraction module, the light intensity image and polarization image are input through two channels. The first layer is the convolution layer containing the 3 × 3 convolution kernel and activation function ReLU (rectified linear unit), which is used to extract low-level features. The second layer is the DenseBlock module containing 3 convolution layers to extract high-level features, in which each convolution layer also uses a 3 × 3 convolution kernel. The operation step of the convolution kernel is 1. Before the convolution operation, there are the BN (batch normalization) layer and ReLU activation function. This sort can speed up the training speed of the network. The two input channels of light intensity image and polarization image share the same weight, which can reduce the computational complexity of the network. This is followed by the attention unit (see Section 3.2), which takes the feature map of the previous layer as input. It is able to capture the global relationships in the data and guide the network to learn the distribution of the feature map. Second, in the feature fusion module, the feature map output by the feature extraction module is superimposed. The channel size of the two feature maps is 128, and the channel size of the fused feature map after being superimposed is 256. Finally, the output of the feature fusion module is used as the input of the image reconstruction module. The image reconstruction module includes 5 transposed convolution layers, and the convolution kernel size of each transposed convolution layer is also 3 × 3. The fusion results are reconstructed from the fusion features through these 5 transposed convolution layers. A more detailed network architecture is shown in Table 1.

3.2. Attention Mechanism

The attention unit combines channel attention and spatial attention. Channel attention enables the network to learn the importance of features in the channel domain and give different weights to the feature map, so as to achieve the selective combination of light intensity image and polarization image in the channel domain. Spatial attention focuses on learning the effective information distribution of each layer of the feature map to improve the transmission of salient features. The attention unit includes a global mean pooling layer, a convolution layer, an activation layer, and a splicing layer, and its detailed structure is shown in Figure 4. Given X R H × W × C and X R H × W × C as the input and output of the attention unit, the calculation process of the attention unit is
X = σ F c X F s X X + X ,
where σ is the sigmoid activation function, Fc is the channel attention branch, Fs is the space attention branch, is the broadcast addition operation, and is the element-by-element multiplication operation.
When the input feature map X R H × W × C passes through the channel attention branch, the channel feature X c R 1 × 1 × C is obtained through the global average pooling layer, and then the channel feature size 1 × 1 × C r obtained by point-by-point convolution of PWConv1, BN layer, and ReLU activation function. The channel attention feature map X c of size 1 × 1 × C is obtained by point-by-point convolution of PWConv2 and BN layer. F c is expressed as
F c X = B N P W C o n v 2 δ B N P W C o n v 1 G A P X ,
where δ is the ReLU activation function, and GAP is the global average pooling. Similar to the channel attention branch, when passing through the spatial attention branch, 3 × 3 convolution Conv1, BN layer, and ReLU activation function were first used to obtain the feature map of the size H × W × C r . To obtain the spatial attention feature map of size H × W × C 1 × 1 , convolution PWConv2 and BN layer were used. F s can be expressed as:
F s X = B N P W C o n v 2 δ B N C o n v 1 X .

3.3. Loss Function

The loss function in this paper adopts the globally weighted SSIM (structural similarity) loss function, which is a multiscale and weighted SSIM (MSWSSIM) [17]:
L o s s M S W S S I M = 1 1 5 ω 3 , 5 , 7 , 9 , 11 γ ω L o s s S S I M I S 0 , I f ; ω + 1 γ ω L o s s S S I M I D o L P , I f ; ω .
Loss SSIM ( x ,   y ;   ω ) is a loss function based on the SSIM, which represents the structural similarity of the image x and y on window ω:
L o s s S S I M x , y ; ω = 2 ω ¯ x ω ¯ y + C 1 2 σ ω x ω y + C 2 ω ¯ x 2 + ω ¯ y 2 + C 1 σ ω x 2 + σ ω y 2 + C 2 .
ω x is the region of the image within the window ω , and ω ¯ x is the mean of ω x . The variables σ ω x 2 and σ ω x ω y are the variance of ω x and the covariance of ω x and ω y , respectively. The remaining ones, ω y , ω ¯ y , and σ ω y 2 , represent the corresponding meanings. C 1 and C 2 are constants to avoid instability when ω ¯ x + ω ¯ y and σ ω x + σ ω y are very close to zero, respectively.
The multiwindow SSIM is proposed in the loss function to solve the problem of image detail at different scales. The window sizes used include 3, 5, 7, 9, and 11. Different windows can extract features of different scales. In addition, Loss SSIM ( I S 0 ,   I f ;   ω ) and Loss SSIM ( I DoLP ,   I f ;   ω ) use the weight coefficient, which is based on σ ω S 0 2 and σ ω D o L P 2 , defined as shown in Equation (17). When the window ω of the intensity image S 0 variance is greater than the corresponding DoLP image, it indicates that the local region of S 0 has more image details; that is, the weight coefficient γ ω corresponding to the image of S 0 should be larger.
γ ω = g σ ω S 0 2 g σ ω S 0 2 + g σ ω D o L P 2 .
σ ω S 0 2 is the variance of the intensity image S 0 within the window ω ; σ ω D o L P 2 is the variance of the DoLP image within the window ω . g x = max x ,   0.0001 is a correction function to increase the robustness of the solution.
In addition, MSWSSIM can retain high-frequency information, but it is insensitive to uniform deviation, which can easily lead to changes in brightness. Therefore, the integration of MSWSSIM with the L1 norm loss function can ensure the brightness of fusion results. The L1 norm loss function can be expressed as
L o s s L 1 = 1 M N I a v g I f 1 ,
where M and N are the height and width of the image, respectively. I a v g is the average value of I S 0 and I D o L P . I f is the fused image. Then the final loss function can be expressed as
L M i x = L o s s M S W S S I M + α L o s s L 1 .
where α is a balance parameter.

4. Experiment

In order to obtain the dataset, the underwater imaging experiment was conducted. The experimental device and layout are shown in Figure 1, which mainly includes a polarization camera, glass water tank, polarization light source, and target object. The resolution of the polarizing camera (PHX0550S-P) is 2448 × 2048, and the number of bits is 12. It adopts focal-plane polarizing imaging and can take four images of linear polarized light intensity with polarization angles of 0°, 45°, 90°, and 135° at one time. The focal length of the lens is 10.5 mm. The polarized light source consists of an LED light source and a linear polarizer. A water tank (500 mm × 250 mm × 250 mm) was used as a container and the inside was covered with black flannel to avoid interference from ambient light. The target was placed in the water tank filled with water. The milk was prepared in the water tank to simulate the underwater situation with suspending particles. The light intensity and polarization images were obtained. Finally, a dataset containing 150 sets of images was constructed. Each set of images was composed of corresponding light intensity and polarization images, whose size was 1224 × 1024. A total of 100 groups were used as the training set, and the remaining 50 groups were used as the verification and test sets. In addition, the image of the dataset was flipped and trimmed to 80 × 80 as the input of network training. The training process was carried out on the server, whose graphics card was NVIDIA GeForce RTX 2080 Ti, whose processor was i9-9600X, and whose memory was 128 G. After weight initialization, the optimization was performed using the Adam optimizer with a minibatch size of 128. The learning rate was initially set to 0.0001 and decayed exponentially at a rate of 0.99, with a maximum epoch set to 200.

5. Results

5.1. Image Enhancement

Based on the above method, the network was trained and the performance of underwater image fusion was tested. Information entropy (IE), standard deviation (SD), mutual information (MI), and SSIM were adopted to measure the quality of the fusion image objectively. IE represents the average information amount of an image, as shown in Equation (20). The more information amount, the greater IE. Image fusion will result in the increase in image information, and IE can reflect the degree of change.
H ( X ) = E [ log 1 p ( a i ) ] = i = 1 n p ( a i ) log p ( a i ) .
E is the statistical average, and p ( a i ) represents the probability of the gray value a i appearing.
SD refers to the dispersion degree of the image pixel gray value relative to the mean ( μ ) . If the SD is larger, it indicates that the gray levels in the image are more dispersed and the image quality is better. The calculation formula is as follows:
S D = 1 M N i = 1 M j = 1 N I x , y μ 2 , μ = 1 M N i = 1 M j = 1 N I x , y .
MI can measure the degree of similarity between two images, that is, the amount of information of the original image obtained by the fusion of images. The larger the MI is, the more source image information is retained and the better the quality is. MI is based on the IE H(A) and the joint IE H(A,B) of the image:
M I ( A , B ) = H ( A ) + H ( B ) H ( A , B ) .
The calculation formula adopted in this paper is
M I = 1 2 M I ( I S 0 , I f ) + 1 2 M I ( I D o L P , I f ) .
SSIM is a widely used image quality evaluation index. It is based on the assumption that structured information will be extracted when human eyes watch an image. The SSIM value ranges from −1 to 1. The closer it is to 1, the higher the similarity is and the better the fusion quality is. The calculation formula adopted in this paper is
S S I M = 1 2 S S I M ( I S 0 , I f ) + 1 2 S S I M ( I D o L P , I f ) .
The network training test results are shown in Figure 5. According to the image fusion results, it can be found that the quality of the light intensity image S0 is poor and the scene details are degraded seriously. However, after adding the polarization image for fusion, the target becomes clearer and the texture outline of the key can be clearly identified. According to the image index used for evaluation, the IE and SD after fusion increase by 24.48% and 139%, indicating that the proposed method can improve the quality of underwater images. In addition, the fusion image obtained by this method is compared with several other image fusion methods, including the curvelet transform (CVT) [10], the gradient transfer fusion (GTF) [21], the multiresolution singular value decomposition (MSVD) [22], the ratio of low-pass pyramid (RP) [7], and the discrete wavelet transform (DWT) [8]. As can be seen from Figure 5, the result of RP has poor visual quality. Artifacts are generated to a certain extent on the edge of the key and the shaded part, and there are more noises. The results of CVT, DWT, and MSVD have a certain degree of graininess, and the processing ability of shadows is poor. GTF results have high contrast, but the texture details of the key are not clear enough. However, the method presented in this paper has a relatively real visual effect without obvious artifacts and distortions and has a good processing effect on shadows. To further improve the image-enhancing effect, we also tested different network configurations, but the image quality is not improved much and the corresponding configurations are not valuable. Thus, we choose to only show the current configuration of the network, which already meets the actual requirements.
In order to objectively evaluate the performance of the method, the four image evaluation indexes introduced previously were used to evaluate the images in the test set, and the final results were averaged, as shown in Table 2. The method has better performance in the image evaluation indexes of IE, SD, MI, and SSIM, which further proves the effectiveness of the method.
Finally, the running time was evaluated on a server configured with NVIDIA GeForce RTX 2080 Ti, 3.1 GHz Intel Core i9-9600X, and 128 G RAM. The results are shown in Table 3. All methods were implemented using the average value of multiple groups of images in Python language. It can be seen that the processing speed of the proposed method is faster than other methods.

5.2. Target Segmentation

Most of the existing underwater image segmentation algorithms are based on one of the light intensity information, spectral information, and polarization information, which has great limitations. It is necessary to make reasonable comprehensive use of light intensity and polarization information. In the study, we use the improved U-net network to extract features from the fusion image for image segmentation [23,24]. The framework and corresponding configuration of the image segmentation are similar to the image enhancement section of the previous part. We simulated different turbidities underwater by adding different volumes of milk (0, 1, 2, and 3 mL) into the water tank. Figure 6a,b are original images and segmentation results, respectively, at different water turbidities. With the increase in turbidity, the intensity of backscattered light increases, the clarity of the original image decreases, and the noise increases. When the milk volume is 3 mL, the image quality decreases obviously, but the general outline can still be detected in the segmentation results.
We used pixel accuracy (PA) and mean intersection over union (MIoU) to measure the accuracy of image segmentation. PA represents the ratio of correctly predicted pixels to all pixels in the image. The physical meaning of MIoU is that the overlap of the predicted and labeled region is divided by the combination of the predicted and labeled region. The evaluation index of target segmentation results with different water turbidities is shown in Figure 7. After adding polarization information, PA and MIoU are improved. The results show that the target segmentation based on the proposed method is feasible under turbid water.

6. Conclusions

Aiming at the problem of poor quality of underwater optical imaging, this paper proposes a method of applying a deep fusion network to underwater polarization images. By analyzing the underwater active polarization imaging model, we set up an experimental device to obtain the underwater polarization image and construct the training dataset. We establish an end-to-end network model based on unsupervised learning and attention mechanism guidance and design the loss function. The experimental results show that the method can improve the visual quality of the image and is superior to other methods. The processing speed of this method is faster than that of other methods, which shows the potential of the method to meet the requirements of real-time underwater video processing. Next, we improve the U-net network structure to extract features for image segmentation. The results show that the target segmentation based on the proposed method is feasible under turbid water. Future research includes building a more comprehensive dataset and improving the loss function and network module to further improve the quality of fusion image and meet the requirements of practical applications. At the same time, the processing efficiency of the algorithm will be improved to reduce the operation time and realize the real-time detection of underwater targets.

Author Contributions

Methodology, H.C.; Writing—review & editing, D.Z., J.Z. and H.Y.; Funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC) (52175265).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We thank the anonymous reviewers for their valuable recommendations, which enabled us to improve this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An underwater image enhancement benchmark dataset and beyond. IEEE Trans. Image Process. 2019, 29, 4376–4389. [Google Scholar] [CrossRef] [Green Version]
  2. Miller, P.A.; Farrell, J.A.; Zhao, Y.; Djapic, V. Autonomous underwater vehicle navigation. IEEE J. Ocean. Eng. 2010, 35, 663–678. [Google Scholar] [CrossRef]
  3. Cheng, H.; Chu, J.; Zhang, R.; Tian, L.; Gui, X. Underwater polarization patterns considering single Rayleigh scattering of water molecules. Int. J. Remote Sens. 2020, 41, 4947–4962. [Google Scholar] [CrossRef]
  4. Cheng, H.; Chu, J.; Zhang, R.; Gui, X.; Tian, L. Real-time position and attitude estimation for homing and docking of an autonomous underwater vehicle based on bionic polarized optical guidance. J. Ocean. Univ. China 2020, 19, 1042–1050. [Google Scholar] [CrossRef]
  5. Huang, B.; Liu, T.; Hu, H.; Han, J.; Yu, M. Underwater image recovery considering polarization effects of objects. Opt. Express 2016, 24, 9826–9838. [Google Scholar] [CrossRef] [PubMed]
  6. Treibitz, T.; Schechner, Y.Y. Active polarization descattering. IEEE Trans. Pattern Anal. Mach. Intell. 2008, 31, 385–399. [Google Scholar] [CrossRef] [Green Version]
  7. Toet, A. Image fusion by a ratio of low-pass pyramid. Pattern Recognit. Lett. 1989, 9, 245–253. [Google Scholar] [CrossRef]
  8. Pajares, G.; Cruz, J. A wavelet-based image fusion tutorial. Pattern Recognit. 2004, 37, 1855–1872. [Google Scholar] [CrossRef]
  9. Candes, E.J. Ridgelets: Theory and Applications; Stanford University: Stanford, CA, USA, 1998. [Google Scholar]
  10. Nencini, F.; Garzelli, A.; Baronti, S.; Alparone, L. Remote sensing image fusion using the curvelet transform. Inf. Fusion 2007, 8, 143–156. [Google Scholar] [CrossRef]
  11. Rahmani, S.; Strait, M.; Merkurjev, D.; Moeller, M.; Wittman, T. An adaptive IHS pan-sharpening method. IEEE Geosci. Remote Sens. Lett. 2010, 7, 746–750. [Google Scholar] [CrossRef] [Green Version]
  12. Jolliffe, I.T. Principal Component Analysis. J. Mark. Res. 2002, 87, 513. [Google Scholar]
  13. Liu, Y.; Chen, X.; Peng, H.; Wang, Z. Multi-focus image fusion with a deep convolutional neural network. Inf. Fusion 2017, 36, 191–207. [Google Scholar] [CrossRef]
  14. Kalantari, N.K.; Ramamoorthi, R. Deep high dynamic range imaging of dynamic scenes. ACM Trans. Graph. 2017, 36, 1–12. [Google Scholar] [CrossRef] [Green Version]
  15. Ma, J.; Ma, Y.; Li, C. Infrared and visible image fusion methods and applications: A survey. Inf. Fusion 2019, 45, 153–178. [Google Scholar] [CrossRef]
  16. Elzeki, O.M.; Abd Elfattah, M.; Salem, H.; Hassanien, A.E.; Shams, M. A novel perceptual two layer image fusion using deep learning for imbalanced COVID-19 dataset. PeerJ Comput. Sci. 2021, 7, e364. [Google Scholar] [CrossRef] [PubMed]
  17. Zhang, J.; Shao, J.; Chen, J.; Yang, D.; Liang, B.; Liang, R. PFNet: An unsupervised deep network for polarization image fusion. Opt. Lett. 2020, 45, 1507–1510. [Google Scholar] [CrossRef] [PubMed]
  18. Liu, X.; Liu, Q.; Wang, Y. Remote sensing image fusion based on two-stream fusion network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef] [Green Version]
  19. McGlamery, B.L. A computer model for underwater camera systems. Ocean. Opt. VI 1980, 208, 221–231. [Google Scholar]
  20. Jaffe, J.S. Computer modeling and the design of optimal underwater imaging systems. IEEE J. Ocean. Eng. 1990, 15, 101–111. [Google Scholar] [CrossRef]
  21. Ma, J.; Chen, C.; Li, C.; Huang, J. Infrared and visible image fusion via gradient transfer and total variation minimization. Inf. Fusion 2016, 31, 100–109. [Google Scholar] [CrossRef]
  22. Naidu, V.P.S. Image fusion technique using multi-resolution singular value decomposition. Def. Sci. J. 2011, 61, 479. [Google Scholar] [CrossRef] [Green Version]
  23. Lee, G.; Jhang, K. Neural network analysis for microplastic segmentation. Sensors 2021, 21, 7030. [Google Scholar] [CrossRef] [PubMed]
  24. Zunair, H.; Hamza, A.B. Sharp U-Net: Depthwise Convolutional Network for Biomedical Image Segmentation. Comput. Biol. Med. 2021, 136, 104699. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Underwater imaging model.
Figure 1. Underwater imaging model.
Sensors 23 05594 g001
Figure 2. Polarization imaging model.
Figure 2. Polarization imaging model.
Sensors 23 05594 g002
Figure 3. The network structure.
Figure 3. The network structure.
Sensors 23 05594 g003
Figure 4. Attention unit structure.
Figure 4. Attention unit structure.
Sensors 23 05594 g004
Figure 5. Comparison of fusion results of different methods for the partial test set.
Figure 5. Comparison of fusion results of different methods for the partial test set.
Sensors 23 05594 g005
Figure 6. Results of target segmentation. (a) Original images. (b) Segmentation results.
Figure 6. Results of target segmentation. (a) Original images. (b) Segmentation results.
Sensors 23 05594 g006
Figure 7. Quantitative analysis of target segmentation.
Figure 7. Quantitative analysis of target segmentation.
Sensors 23 05594 g007
Table 1. Network structure configuration.
Table 1. Network structure configuration.
ModuleInputLayerConvolution Kernel SizeInput
Channel
Output ChannelActivation
Function
Output
Feature extraction S 0 Conv3 × 3116ReLU
Conv3 × 31616ReLU
Conv
Conv
3 × 3
3 × 3
32
48
16
16
ReLU
ReLU
AttentionBlock-64128ReLU F S 0
DoLPConv3 × 3116ReLU
Conv3 × 31616ReLU
Conv3 × 33216ReLU
Conv3 × 34816ReLU
AttentionBlock-64128ReLU F D o L P
Feature fusion F S 0 ,     F D o L P Concat-128 × 2256- F f u s e d
Reconstruction F f u s e d Conv3 × 3256128ReLU
Conv3 × 312864ReLU
Conv3 × 36432ReLU
Conv3 × 33216ReLU
Conv3 × 3161- I f
Table 2. Quantitative results of polarization image fusion by different methods.
Table 2. Quantitative results of polarization image fusion by different methods.
CVTGTFMSVDRPDWTOurs
IE6.11236.98186.05916.48156.08347.1073
SD17.8742.6217.1121.9318.2140.46
MI0.73741.27250.75740.361.47171.6808
SSIM0.49530.35420.49590.24090.50490.5186
Table 3. Operating time of different methods.
Table 3. Operating time of different methods.
CVTGTFMSVDRPDWTOurs
time/s1.0420.0680.35680.0820.50860.051
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, H.; Zhang, D.; Zhu, J.; Yu, H.; Chu, J. Underwater Target Detection Utilizing Polarization Image Fusion Algorithm Based on Unsupervised Learning and Attention Mechanism. Sensors 2023, 23, 5594. https://doi.org/10.3390/s23125594

AMA Style

Cheng H, Zhang D, Zhu J, Yu H, Chu J. Underwater Target Detection Utilizing Polarization Image Fusion Algorithm Based on Unsupervised Learning and Attention Mechanism. Sensors. 2023; 23(12):5594. https://doi.org/10.3390/s23125594

Chicago/Turabian Style

Cheng, Haoyuan, Deqing Zhang, Jinchi Zhu, Hao Yu, and Jinkui Chu. 2023. "Underwater Target Detection Utilizing Polarization Image Fusion Algorithm Based on Unsupervised Learning and Attention Mechanism" Sensors 23, no. 12: 5594. https://doi.org/10.3390/s23125594

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop