Next Article in Journal
Extreme Wind Wave Climate off Jeddah Coast, the Red Sea
Previous Article in Journal
Underwater Chatter for the Win: A First Assessment of Underwater Soundscapes in Two Bays along the Eastern Cape Coast of South Africa
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Experimental Research on Overwater and Underwater Visual Image Stitching and Fusion Technology of Offshore Operation and Maintenance of Unmanned Ship

1
School of Civil Engineering & Transportation, South China University of Technology, Guangzhou 510641, China
2
School of Mechanical & Automotive Engineering, South China University of Technology, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
J. Mar. Sci. Eng. 2022, 10(6), 747; https://doi.org/10.3390/jmse10060747
Submission received: 8 April 2022 / Revised: 26 May 2022 / Accepted: 27 May 2022 / Published: 29 May 2022
(This article belongs to the Section Ocean Engineering)

Abstract

:
The new way of offshore operation and maintenance based on unmanned ships has outstanding advantages. Aiming at the problem of lack of overall understanding of the complex environment above and under the water surface during the operation and maintenance of unmanned ships, a stitching and fusion technology of overwater and underwater visual images for unmanned ships is proposed. The software and hardware framework of the overwater and underwater visual image fusion system is constructed, the image processing methods in different environments are defined, and the accurate acquisition of obstacle information is realized. In the two experimental scenarios, the stitching accuracy of the obstacle model based on an extended neighborhood method can reach more than 85% within the obstacle distance of 8 m and more than 80% within the obstacle distance of 14 m. An image-driven Frustum–PointNets detection algorithm is proposed to obtain comprehensive obstacle avoidance information. In addition, the average accuracy of the three-dimensional detection of the algorithm is up to 91.40%. These results are significant and have a good reference value, as it demonstrates that the stitching and fusion method can effectively obtain the comprehensive information of overwater and underwater objects for unmanned ship.

1. Introduction

The ocean is rich in oil and gas, energy, and biological resources, and ocean development has become the strategy of various countries [1]. With the development of the ocean, more and more offshore structures have been put into use, such as offshore wind power platforms and oil production platforms [2,3]. The safety maintenance of offshore platforms is very important. However, the operation and maintenance of conventional ships require a lot of manpower and financial resources, and there are still great risks [4]. At present, the new way of maritime operation and maintenance based on unmanned ships is a development trend.
With the continuous development and improvement of surface unmanned ship technology, the potential business model of unmanned ships has been further explored [5]. The application scenarios of unmanned ships are very broad, involving fields such as surveying and mapping, environmental monitoring, search and rescue, and security patrol [6]. Autonomous collision avoidance in various scenarios is a current hotspot of unmanned ship technology [7,8]. Accurate perception of various obstacles in water in complex environments and obtaining the full picture of collision avoidance objects are the key to autonomous collision avoidance for unmanned ships [9,10].
According to the position of obstacles relative to the water surface, they can be divided into obstacles above the water, obstacles under the water, and obstacles across the water surface. Among them, the shape of obstacles across the water surface extends from underwater to above the water surface, such as icebergs, reefs, and most marine structures. The overall appearance of such obstacles is usually difficult to be obtained by a single sensor. During the navigation of unmanned ships, marine radar and underwater sonar can be used to quickly perceive long-distance obstacles, make collision avoidance decisions, and stay away from the area where the obstacles are located [11]. However, when unmanned ships perform offshore operation and maintenance tasks, they inevitably need to get close to offshore platforms and structures. These offshore structures usually belong to obstacles across the water surface. The irregular shape and different degrees of surface unevenness of their overwater and underwater parts will also pose a threat to the safety of unmanned ships [12]. Therefore, it is of great significance to use visual perception technology and to study image stitching and fusion technology of overwater and underwater images of obstacles which are across the water surface, so as to obtain the full picture of the obstacles and provide comprehensive environmental information for unmanned ships.
In terms of obstacle perception technology for unmanned ships, Szpak Z.L. et al. [13] proposed a method of background subtraction to establish a single Gaussian background model in order to identify and segment the target and the water surface background in the image obtained by a camera, so as to obtain the contour of the target. In order to realize the omnidirectional perception of obstacles, Braginsky B. et al. [14] installed a collision avoidance sonar in the front and on the side of the underwater robot, respectively. Combined with the obstacle perception information of the two sonars, an OAA double-layer obstacle avoidance method was used to realize effective avoidance of the obstacles. On the basis of visual image perception of obstacles, Fang M. et al. [15] proposed research on the control technology of the navigation direction of the vehicle by using the position co-ordinates of the target and the vehicle. This control can select the optimal obstacle avoidance decision path to achieve obstacle avoidance. Gu J. et al. [16] used forward-looking sonar to image the underwater obstacle target. The characteristics of the target were obtained through the analysis and preprocessing of the image, and the detection of the obstacle target was realized according to the main feature matching. After acquiring the sonar image of the underwater target, Williams D.P. [17] used an unsupervised target detection method to achieve accurate measurement of the target’s distance, orientation, and other information. Although there are corresponding recognition technologies for obstacles above the water and obstacles under the water at present, there is a lack of effective overall perception methods for obstacles across the water surface. In particular, there is little research on reconstruction and image stitching and fusion technologies of the above water and underwater parts of the obstacles.
In terms of underwater image detection methods, Raihan A J. et al. [18] considered the distortion of underwater images due to absorption, scattering, polarization, diffraction, and low natural light penetration, and proposed an effective method to restore underwater images. This method can detect more edges and more SURF points by considering the wavelength, attenuation, and backscattering coefficients of red, blue, and green light.
In terms of visual image stitching technology, Berus L. et al. [19] proposed a new edge detection and image stitching method. The results show that, to a certain extent, only the Canny edge detection filter can adequately distinguish the rough boundaries of low-resolution images. Xiu C. et al. [20] proposed a stitching method based on adaptive weighted fusion, which uses an adaptive weighting method to fuse the brightness difference of overlapping areas, realizes smooth transition of stitched images, and eliminates stitching seams. Patel K. et al. [21] proposed a key frame extraction method based on block histogram difference and edge matching rate. The histogram difference value of each frame is calculated, and edge matching is performed on the adjacent images of the acquired video image through the edge detection method of the Prewitt operator. If more than half of them can be successfully matched, the system considers that the selected frame does not meet the key frame requirements. Lu J. et al. [22] proposed an image stitching method based on fuzzy inference, using contrast-limited adaptive histogram equalization to increase the matching points of target surfaces in low-contrast images, reducing ghosting and improving stitching quality. Pollard S. et al. [23] proposed a view synthesis method based on edge matching and transmission of a trinocular vision system. This method does not rely on dense correspondence and has high computational efficiency. Through edge matching, interpolation, occlusion detection, and correction, the automatic view synthesis of three groups of images is finally carried out. In terms of image fusion technology, Li J. et al. [24] proposed a new method for accurate stitching of images acquired by multiple pushbroom imagers, established a relative orientation model based on homophonic points, and then used a strict geometric imaging model to generate seamless stitched images, which improved the precision. Selvaraj A. et al. [25] obtained the edge curve features with better effect through multi-scale decomposition of the visible light image and infrared image obtained in the experiment, and fused the high-frequency coefficient features by using the consistency verification method, which not only suppressed the noise of the image, but also retained the detailed information of the image. Ch M.M.I. et al. [26] proposed a multifocal image fusion method based on cross bilateral filtering and non-subsampling contourlet transform. In this method, the obtained original image and approximate components are filtered bilaterally to obtain the approximate weight graph, the details in the image components are combined by the weighted average method to obtain the detail weight graph, and the detail weight graph is combined with the original image to realize the image fusion processing. Hayat N. et al. [27] proposed a multi-exposure image fusion (MEF) technology that does not require manual operation. This technology generates high-quality output images by acquiring multi-exposure input images. This method obtains the weight graph of the original image by analyzing the information of image contrast, saturation, and exposure, and uses the multi-resolution pyramid decomposition method to mix and generate the final fused image. Paramanandham N. et al. [28] proposed a new image fusion technology based on the similarity of continuous pixel intensity. The noise-free and noisy parts of the source image are detected, and the optimized particle swarm algorithm is used for image fusion. Experiments were carried out on the images affected by Gaussian noise, Poisson noise, and speckle noise, and high-quality fused images were obtained. At present, most edge stitching technologies are applied in the fields of file restoration, broken coins stitching, aerial images, etc. There is little research on overwater and underwater image stitching. Most of the visual image fusion techniques are researched on two or more images with overlapping areas, while there is less research on segmented images without overlapping areas. In addition, the research on the fusion technology of overwater images and underwater images of navigation obstacles of unmanned ships is even less.
Therefore, in view of the difficulty in obtaining the overall information of obstacles across the water surface and the overall picture of the environment during the operation and maintenance of the unmanned ship at sea, in this paper, the overwater and underwater visual perception technology is used to obtain the obstacle images. By expanding the neighborhood of the edge of the obstacle image, combining the characteristic information of the neighborhood with the outer edge contour, and stitching and fusing the overwater and underwater images of the obstacle, the overall panoramic information of the obstacle is finally obtained, so as to provide useful information for unmanned ship operation.

2. Features and Framework of Unmanned Ship Visual Image Fusion System

The fusion characteristics of the overwater and underwater images of unmanned ship obstacles are analyzed, and then the hardware and software frameworks of the unmanned ship visual image stitching and fusion system are established.

2.1. Analysis of Visual Image Fusion Characteristics

For the overwater image of obstacles, due to the mirror effect of the water surface, the image of obstacles will produce reflection. In low sea conditions, the characteristics of the obstacle reflection are similar to the actual image of the obstacle, so the symmetry can be used to remove the reflection to obtain the actual image of the obstacle. In high sea conditions, the reflection of the obstacle is quite different from the actual image. At this time, the reflection can be regarded as the image background. Therefore, the actual image of the obstacle can be obtained by removing the background. After obtaining the overwater image of the obstacle, in order to reduce the huge amount of data processed by the system and reduce the complexity of feature extraction, it is necessary to perform edge detection on the overwater obstacle image to reduce the redundant information of the pixels and to obtain the fusion features of the main edges and their extended neighborhoods.
For the underwater images of obstacles, there will be a certain fogging phenomenon due to the refraction, reflection, and absorption of light waves. In addition, impurities in water will also produce noise interference to the images. In low sea conditions, the underwater part of the obstacle will form a mirror image with the water surface as the symmetry axis. In this case, the symmetry can be used to remove the mirror image and segment the underwater part image of the obstacle. In high sea conditions, the mirror image of the underwater part of the obstacle begins to be distorted and blurred, so it can be removed as the background without affecting the segmentation and acquisition of the underwater part of the obstacle.
In the process of unmanned ship operation, the image, contour, distance, orientation, and other information of the overwater part of the obstacle are obtained through the overwater camera and lidar, and the same information of the underwater part of the obstacle are obtained through the underwater camera and sonar ranging system. Then, the overwater image and the underwater image are fused based on the edge features of the two images. In order to reduce the amount of calculation of image processing and improve the robustness of image fusion, this paper carries out image edge detection on both the overwater image and the underwater image, and completes the image stitching of the obstacle through the characteristics of the pixel points of the image edge and its extended neighborhood.

2.2. Hardware Framework of Visual Image Fusion System

Obtaining more accurate information of the overwater and underwater parts of the obstacle is very important for the final visual image fusion. After determining the basic structure of the hull and the power system of the unmanned ship, this paper designs a reliable and efficient visual image fusion system for the unmanned ship obstacles.
The hardware framework of the visual image fusion system is shown in Figure 1. It consists of two parts: information display and parameter control system, and visual image fusion and collision avoidance decision system.
The information display and parameter control system is realized by the host computer on shore, and its function module mainly includes two parts: ➀ the obstacle information display module, which can display the size, distance, and orientation of the obstacles, and ➁ the control parameters display and input module, of which the control parameters mainly include emergency stop, remote control, path planning, and other instructions. Through the information transmission system, the host computer on shore can receive the environmental information obtained by each environmental sensing device in real time and display the data processed by the system on the interface of the host computer.
The visual image fusion and collision avoidance decision system is divided into the following five modules: ➀ overwater obstacle perception system, which obtains overwater information of the obstacles, ➁ underwater obstacle perception system, which obtains underwater information of the obstacles, ➂ navigation and positioning system, which obtains real-time position and attitude information of the unmanned ship, ➃ energy power system, which provides power supply of unmanned ship, and ⑤ information transmission system, which performs remote wireless transmission of information of the unmanned ship.

2.3. Software Framework of Visual Image Fusion System

The software framework designed in this paper is shown in Figure 2, which is mainly divided into two modules: ➀ human computer interaction module, which is mainly for the convenience of shore-based staff to manually intervene and control the unmanned ship when necessary and to timely grasp the current heading and hull condition of the unmanned ship, and ➁ image stitching and fusion module, which obtains the information of the overwater part of the obstacle through the underwater camera and lidar, obtains the information of the underwater part of the obstacle through the underwater camera and sonar ranging system, and, finally, carries out image fusion to obtain the full picture of the obstacle.

3. Fusion Technology of Overwater and Underwater Images

After the overwater and underwater images of the obstacles of the unmanned ship are obtained, the images are first preprocessed, then obstacle segmentation and edge detection are performed, and then the overwater and underwater images are spliced, and, finally, the obstacle information is fused.

3.1. Image Preprocessing

Technical analysis and method realization are carried out, respectively, for overwater image defogging, underwater image clarity, and reflection detection.

3.1.1. Defogging Processing of Overwater Image

In view of the problems of insufficient transmittance estimation, slow processing speed, and high complexity in the traditional defogging technology in complex environments, and in order to meet the requirements of dynamic adaptability, processing speed, and image detail enhancement of image defogging during the navigation of the unmanned ship, transmittance optimization and fast estimation of atmospheric light values are conducted in this paper on the basis of the dark channel prior principle [29].

3.1.2. Clarity Processing of Underwater Image

The factors causing overwater image fogging and underwater image fogging are different. The fogging of the overwater image is mainly due to the influence of water vapor or fog, while that of the underwater image is mainly caused by the absorption and scattering of light by the water body. In this paper, an improved relative global histogram stretching method is used to restore and enhance the underwater image of the obstacle.

3.1.3. Obstacle Reflection Detection

Under low sea conditions, especially in the case of sufficient lighting, the actual image of the obstacle and its reflection have little difference in color and brightness, forming a strong reflection. In this paper, the underlying descriptor differential feature extraction method is used to extract the distance feature and intensity feature of the pixel point. Then, the symmetry axis detector based on multi-instance learning is used to detect the symmetry axis between the actual image and the reflection of the obstacle.

3.2. Image Segmentation and Edge Detection

After image preprocessing, the obstacles in the overwater image and underwater image need to be segmented and edge detected. In this paper, an obstacle image segmentation method based on an automatic multi-seed region growing algorithm is adopted, and then the edge detection method based on Canny algorithm is used to detect the contour of the obstacle.

3.2.1. Image Segmentation

The image to be detected includes obstacle region M and background region N. The ideal image segmentation is to accurately distinguish the obstacle region M from the background region N. However, in the actual image, there is a blurred transition region between M and N, which makes it difficult to segment them accurately. In order to achieve effective segmentation and better clarify the segmentation algorithm, the obstacle region M is divided into strong obstacle region P1 and weak obstacle region P2, and the background region N is divided into weak background region P3 and strong background region P4. The distribution of the four regions is shown in Figure 3.
According to the above division method, the following relationship exists:
{ M = P 1 + P 2 N = P 3 + P 4
Among them, the strong obstacle region P1 and the strong background region P4 are considered to have obvious image features, which are easy to be segmented and rarely produce segmentation errors, while the weak obstacle region P2 and the weak background region P3 are in the blurred transition region, which is difficult to accurately segment, and it is considered that there will be segmentation errors in general.
The segmentation idea of this paper is to find the strong background region P4, then set seed points on P4, and execute the region growth algorithm to realize the final background segmentation of the image.
The strong background region P4 can be determined by the following formula.
P 4 = N 1 N 2 N n   ( n     2 )
where N1, N2, … and Nn are the background area of the image obtained by segmentation method 1, segmentation method 2, … and segmentation method n, respectively.
The specific method of this paper is to obtain the background region N1 through brightness threshold segmentation, and then obtain the background region N2 through saturation threshold segmentation, and then use the common region of N1 and N2 as the strong background region P4. It should be noted here that, in the strong background area P4, due to the fluctuation of water surface, P4 may be one connected area or several unconnected areas. Selecting seed points from each area can avoid false detection and missing detection. The flow chart is shown in Figure 4.

3.2.2. Image Edge Detection

After image preprocessing, the edge detection method based on Canny algorithm can be directly used for obstacle contour detection. However, for the application scenario of unmanned ships, the image information obtained may be different in resolution and noise pollution. The gamma transform is used to enhance the detail information of the dark part of the image, and then the median filter is used to denoise the image. The traditional Canny algorithm edge detection method cannot adapt to the edge detection of obstacle images under all sea conditions. This paper improves the above problem by introducing an adaptive threshold. The flowchart of the improved algorithm is shown in Figure 5.
In this paper, the original 2 × 2 Roberts operator of the Canny algorithm is replaced by a 3 × 3 Sobel operator. On this basis, two gradient directions, 45° and 135°, respectively, are added to obtain the gradient size more accurately. The calculated gradient size and angle after increasing the gradient direction are:
G ( x , y ) = G x 2 ( x , y ) + G y 2 ( x , y ) + G 45 0 2 ( x , y ) + G 135 0 2 ( x , y )
θ ( x , y ) = tan 1 ( G y ( x , y ) / G x ( x , y ) )
The gradient size calculation method by increasing the gradient direction of 45° and 135° can obtain more accurate edge detection results and reduce the interference of redundant and complex information.
In this paper, by comparing and analyzing the statistical information of pixel intensity in the obstacle image, the optimal threshold size is automatically determined, and an adaptive threshold method based on the calculation of the gradient intensity statistical data is realized. The specific implementation steps of this method are to count the gradient size of all pixels in the obstacle image, obtain the maximum Vmax, minimum Vmin, and median values Vmed of single channel pixel intensity, and calculate its average value according to the following formula.
V = (Vmax + Vmed + Vmin)/3
Then, the high threshold and low threshold of the adaptive threshold are adjusted through the value of a parameter σ . The algorithm is as follows:
l o w e r = max ( 0 , ( 1 σ ) × V )
u p p e r = min ( 255 , ( 1 + σ ) × V )
After many experiments, it is found that, when σ = 0.32, the processing effect of the algorithm is the best. The method can quickly obtain the high threshold and low threshold of the double-threshold detection and reduce the time consumption of the algorithm.

3.3. Image Stitching and Fusion

This part realizes the stitching of the overwater image and underwater image of the obstacle through the matching and calculation of image coding points. In addition, the 3D fusion detection of the obstacle is carried out by using the overall contour of the stitched obstacle as the image driver.

3.3.1. Image Edge Stitching

The representation of edge contour can be described by chain codes, which describe the edge characteristics of the image by obtaining the coding of the image boundary. The eight-way code direction symbol used in this paper is shown in Figure 6. The eight numbers 0–7 in the eight-way chain code represent eight directions, respectively. In the eight-way chain code rule, the boundary of a pixel in the image is shown in Figure 7.
The principle of fast stitching and matching of the overwater and underwater edge detection binarized images of unmanned ship obstacles by chain code is as follows: first, obtain the characteristic original code of the region of interest (ROI) of edge detection of the overwater image and underwater image, respectively. Then, judge the original codes of the two images. If the original codes are the same, they will be matched. Otherwise, they will not be matched. Finally, the ROI of the edge detection binarized image of the overwater and underwater obstacle is stitched. According to the original code of the matched image obtained in the above steps, the corresponding subcode space is established, and the subcode is matched through the complementary relationship between the subcodes, which is established by the original code to be matched. If the subcode can be matched, it is considered that the pixels of the edge detection binary image of the overwater obstacle and the underwater obstacle corresponding to the subcode can be stitched; otherwise, it will not be stitched.
The complementary relationship of the subcodes of the eight-way chain code is shown in Table 1. In order to reduce the calculation of unnecessary pixels, only the underwater images of obstacles and the pixels of the edges to be stitched and their neighborhoods of the underwater images are coded. For example, if the chain code is shown in Figure 8, its original code is 10276534 and its complement is 54632170.
In order to minimize the excessive coding of the image and the possible noise influence, this paper takes the image to be stitched as the benchmark, uses the disk model to define the characteristics of each pixel, and determines the starting point of coding and the coding principle of the image edge. There are two types of points that need to be encoded in the edge detection binarized image of the obstacle water image and the underwater image: one is the feature point less than the radius of the disc, and the other is the longest end point of the encoded line segment.
The detection criteria for feature points smaller than the radius are:
R 1 = λ i i = 1 n k ( i = 1 n λ i ) 2
| λ i E M | = 0
M = w ( r i sin θ i , r i cos θ i ) × r 2 ( sin 2 θ i sin θ i cos θ i sin θ i cos θ i cos 2 θ i )
ω ( r sin θ , r cos θ ) = R ω ( n )
R ω ( n ) = ε ( n ω + 1 ) ε ( n )
f ( r i , θ i ) f ( r j , θ j ) r i r j = k
where R 1 represents the code point detection function, M represents the function matrix of corner scanning, ω and R ω both refer to the rectangular window function, and f ( r , θ ) is the curve-fitting function for the contour.
The detection standard for the longest end point of the encoded line segment is:
I ( r , θ ) = f ( r , θ ) + δ × g ( r , θ )
| I ( r , θ ) f ( r i , θ i ) | 0
where θ i , θ j [ 0 , 2 π ] , g ( r , θ ) function represents the direction, and δ represents the function that can be fine-tuned.
The specific process is: firstly, the disc model is established for the edge detection images of the overwater obstacle and the underwater obstacle. Then, set the edge of the images to be stitched and their extended neighborhood as the coded ROI area. Take the disc point O determined by the stitching model as the starting point, and take the disc radius R as the maximum coding length to determine the first coding point. Then, take this point as a new starting point and set the coding direction within ± 90 ° and the coding length to be less than or equal to R to select the next coding point. According to the characteristics of the obstacle images, when selecting the next coding point, first determine the direction, and then determine the coding length. According to the coding rules, select the next coding point again until all the coding points suitable for coding in the disc model are selected. The flowchart is shown in Figure 9.
It is assumed that all coding points have a total of n decision directions, and the ordering and coding length are determined according to the coding points. Then, the coding direction of each coding point is α = 2 π / n , and the coding length is L i = ( R r i ) × ( θ i θ i 1 ) . The Euclidean distances of R/4, R/2, and R for the coding length L i are calculated, and then the coding length corresponding to the shortest Euclidean distance calculated is selected as the final decision coding length, so as to obtain the original code of the obstacle images. The formulas for solving the n-direction normalized differential code s i and the n-direction complement code t i are as follows:
{ S i = ( a i + 1 ) n ( a i ) n S n = ( a 0 ) n ( a n ) n t i = ( S i ) o + ( n + 1 ) / 2
where ( a i ) n refers to the original code value of the i-th point of the n-direction, and ( S i ) o refers to the code value of the i-th point after the normalized difference.
Next, the subcode space is established to obtain the subcode. First, three space sets are defined. E n , S n , and T n , respectively, are used to represent the space sets of the original code, the normalized differential code, and the complement code. The corresponding relationship between the three is as follows:
{ E n f ( s ) S n S n f ( t ) T n
where the subcode space S i and T i are subsets of sets S n and T n , respectively.
The test takes the underwater image of an iceberg model as an example to calculate its normalized differential code. Figure 10 is the coding model of the underwater image of the iceberg model, where 1 refers to the relative position, 2 refers to the stage radius, and 3 refers to the concentric angle. Because the stitching of the overwater image and the underwater image of the obstacle is based on the edge feature, this paper selects the edge and its extended neighborhood as the coding point. The final selection of the coding point for the underwater image of the obstacle is shown in Figure 11. From this, it can be concluded that the original code of the center pole code of the edge of the underwater image of the obstacle is 4541474345435414541460, and the corresponding complement code is 0105030701071050105024.
Normalized chain code is conducive to the unique representation of a code point, but it does not have rotation invariance. In the process of image stitching, it is required that the chain code belonging to a certain edge is unique and does not change after moving and rotating. Normalized differential code has this characteristic. Therefore, in order to obtain the chain code that meets the stitching requirements, the original code needs to be differentiated and then normalized to obtain the normalized differential code. Next, taking the original code of the underwater image of the obstacle as an example, the solution of the normalized differential code is carried out, as shown in Figure 12.
The normalized differential code of the underwater image of the obstacle can be calculated as 1177275317532241753357, while that of the overwater image of the obstacle can be calculated as 1177275317462241753357 by using the same calculation method.
It can be seen that, because the acquisition and processing of the image are affected by the water environment, perception equipment, and processing algorithm, the normalized differential codes of the edge of the overwater image and the underwater image are not exactly the same. However, the matching ratio of the two codes has reached more than 90%, which meets the conditions of edge matching within the allowable range of error.
After obtaining the normalized differential codes of the extended neighborhood edge of the overwater image and the underwater image of the obstacle, the fast matching and stitching of the images can be carried out. Feature extraction and matching calculation must meet the following conditions:
m , S m S i , S m S j
while edge stitching must meet the following conditions:
a , b , S a S i , t b t i , S a = t b
where S i and S j , respectively, refer to the set of edge normalized differential codes to be matched between the overwater image and the underwater image of the obstacle, S m and S a are their subcodes, t i is the set of complementary codes to be stitched, and t b is the subcode of t i . Due to the existence of errors, it is impossible to obtain a completely consistent normalized differential code of the edge to be stitched of the overwater image and the underwater image of the obstacle. Therefore, in the stitching process, the subcode space is used for matching and stitching. The following takes the 5-bit normalized differential code 01234 as an example to illustrate the subcode space establishment, which is shown in Table 2.
It can be seen from the table that the establishment rule of subcode space is to move cyclically according to the sequence of coding points, so that the continuity and accuracy of each subcode during stitching can be ensured. In the matching process, the complexity of the calculation method of the extracted features also directly affects the final stitching effect. The more complex the calculation method is, the more accurate the feature representation will be, which can further reduce the stitching error rate, but it will also reduce the speed of matching and stitching.

3.3.2. Fusion Algorithm

The perception of obstacles based on visible images only obtains the obstacles’ contour and texture, and lacks the depth information of the obstacles, while the point cloud information obtained by radar and sonar contains the depth information of the obstacles. Image information can help radar and sonar to extract and segment targets faster, and images incorporating depth information are given information such as obstacle size and position. Therefore, more comprehensive information of obstacles can be obtained by fusing the image information with the point cloud information obtained by radar and sonar.
PointNet [30] is a classic architecture that uses neural networks to process the target point clouds. The architecture applies the maximum pooling function to extract the features from the target point clouds. In order to effectively utilize the spatial relationship between point clouds, PointNet also connects the point-by-point features and global features of the target, and uses the obtained aggregated information to classify and segment the target. It directly processes the target point cloud, which greatly reduces the loss of information. However, since PointNet uses a multi-layer perception method to extract the target point cloud point by point, the relationship between points is not fully considered, which can easily cause misclassification and mis-segmentation. Therefore, according to the needs of this paper, an image-driven Frustum–PointNet algorithm is proposed.
In the obtained point cloud data, the redundant water surface, background, and other information are nonimportant information. In order to make the information processing of the point cloud network more inclined to the target area, this paper also introduces an attention module, which can make the network learn the location of the target and enhance the expression of the area. The point cloud target type of obstacles is n × 3, which has two dimensions: point and channel. It is necessary to add the attention mechanism to these two dimensions. Figure 13 is the schematic diagram of the channel attention module, and Figure 14 is the schematic diagram of the point attention module.
For the channel attention module, first the obtained point cloud features are input, in which the channel features of the points are solved by maximum pooling and average pooling, and then multi-layer perceptron learning is carried out to obtain the distribution attention of the channel.
The processing process of the point attention module has something in common with the channel. The acquired attention channel features are sent to the multi-layer perceptron for activation to obtain the attention weight of the point attention feature, and then multiplied with the input feature to obtain the point attention features. In this paper, the feature of the target area is obtained first through the channel attention module, and the important points of the target area are obtained through the point attention module, so as to suppress invalid information and enhance the role of effective information.
Compared with the view projection or voxelization method, PointNet point cloud network can minimize the loss of target information, so that the learned 3D information of the target is more complete, especially for the small obstacles; it also has better detection performance.
The Frustum–PointNet algorithm can utilize both 2D image and point cloud information for obstacle detection. Firstly, the 2D image of the obstacle is upgraded to the 3D space through the ascending dimension projection, and then the view cone point cloud is extracted. The 3D segmentation of the input view cone point cloud is performed through the Frustum–PointNet network to obtain the target area of the view cone point cloud, the co-ordinate transformation of the obtained target point cloud is carried out through t-net, and, finally, the 3D frame regression is carried out through PointNet to obtain the specific parameters of the obstacle. The stitched image of obstacles is used to drive the obstacle detection of the 3D point cloud, which can realize accurate positioning of the point cloud. The specific algorithm flowchart is shown in Figure 15.

4. Experiments of the Overwater and Underwater Image Fusion System

The experimental platform is set up, the fusion technology experiment of the overwater and underwater visual images of the unmanned ship is carried out, and, finally, the results are analyzed.

4.1. Layout of the Experiments

The unmanned ship of the experimental platform is independently developed by the research group, which is equipped with various sensors, industrial computers, and upper computers required for the experiment. It can realize the basic functions of target detection, path planning, autonomous berthing, and static and dynamic collision avoidance under specific conditions. The equipment models and characteristics used in this paper are shown in Table 3. On this basis, this paper studies the fusion method of the overwater and underwater visual images of obstacles for the unmanned ship in specific scenes, so as to realize the effective collision avoidance of the unmanned ship against the obstacles with both overwater and underwater bodies. In this paper, the iceberg model and pontoon model are selected as obstacles to carry out experimental research in a wave pool, as shown in Figure 16.
The running environment of the experiment is shown in Table 4. The experimental environment is configured on the laboratory host computer, and the model is built and trained on the basis of the Pytorch framework. The collection of the data set is carried out under the conditions of the laboratory. A total of 500 training samples and 100 test samples are collected, and the test samples include obstacle images and their corresponding point clouds. The data set contains the calibration files between each sensor, and the correspondence between the obstacle images and the radar point cloud space and sonar point cloud space can be obtained through the calibration files.
The system interface of information display and parameter control of the experimental platform is shown in Figure 17, which can display the real-time navigation data of the unmanned ship and the obstacle avoidance information of obstacles in real time, and conduct human–vessel interaction with the unmanned ship in real time.

4.2. Image Segmentation and Obstacle Edge Detection

Figure 18 shows the segmentation results of an obstacle. It can be seen that the method in this paper can eliminate the reflection and background, and retain the obstacle information to the greatest extent. The experimental results show that the image segmentation method based on the automatic multi-seed region growing algorithm in this paper can not only keep the segmentation accuracy of the region growing image, but also avoid the problem that the traditional region growing algorithm has great dependence on the seed point and seed growth order. The selection of seed points does not need artificial participation, which improves the automation performance of the image processing.
For the edge detection of the overwater and underwater images of obstacles, the results of the adaptive threshold algorithm proposed in this paper are compared with the results of the traditional Canny algorithm where the threshold is set manually, as shown in Figure 19. It can be seen that the false edges of the obstacles detected by the algorithm in this paper are greatly reduced, the loss of the true edges is relatively less, and the obtained edges are smoother. The algorithm retains the important edges of the images to the greatest extent, which lays a good foundation for the subsequent image stitching and fusion processing.
In order to further verify that the adaptive threshold algorithms in this paper have a good edge detection effect while maintaining high anti-noise characteristics, this paper makes a comparative study on the detection effect of the common threshold algorithms. The comparison index is the average peak signal-to-noise ratio (PSNR) [31], for which the mathematical expression is:
{ P S N R = 10 × log 10 ( 255 2 M S E ) M S E = 1 m n i = 1 m j = 1 n ( I ( i , j ) K ( i , j ) ) 2
where MSE is the overall mean-square error of the two images before and after denoising, m and n are the length and width of the image, i and j represent the (i, j) pixel of the image, and I(i, j) and K(i, j) represent the pixel value of the point (i, j) before and after denoising respectively.
The comparison results are shown in Figure 20. In addition, the larger the PSNR, the better the image quality. Among the five algorithms, the peak signal-to-noise ratio of the algorithm in this paper is higher, and the average signal-to-noise ratio can reach 7.05. Combining with Figure 21, it can be seen that the algorithm in this paper has a better enhancement effect on the obstacle image and can retain the useful information of the image.

4.3. Quick Stitching of Images

Figure 21 shows the comparison between the stitching results of the overwater image and the underwater image of the obstacles and the real object. Due to the influence of objective environmental conditions, the detection errors of the system and shipborne sensors, and the light of the overwater and underwater environment, there are some errors in the stitching results. However, the errors are within the allowable range and the stitching results have achieved the desired effect.
The quality of the experimental results is measured by the edge description integrity error of the normalized differential code and the stitching error in this paper. The edge description integrity error is calculated by the following formula:
η 1 = | 1 S C |
S = i = 1 n a i , a i [ 0 , R ]
where S is the total length of the normalized differential codes of the coding points, C is the fitting boundary length of the edge to be stitched in the obstacle image, and a i is the code length of the normalized differential code. The edge description integrity error in this paper is within 10%, which is an acceptable error range.
The calculation formula of the stitching error is:
η 2 = E a + E b S a + S b
where Ea and Eb are, respectively, the lengths of the normalized differential codes of the code points that are not stitched in the two images to be stitched. Sa and Sb are, respectively, the total length of the normalized differential codes of the coding points of the two images.
Four wave conditions are set in the experiment, which are no wave, calm wave, smooth wave, and slight wave. For each wave condition, seven groups of experiments were carried out when the obstacle was 2 m, 4 m, 6 m, 8 m, 10 m, 12 m, and 14 m away from the bow of the unmanned ship. The results are shown in Figure 22.
Due to the systematic errors and the complexity of the environment, some points with large errors appear in the experiment, so it is necessary to eliminate these abnormal points. It can be seen from the experimental results that the better the water environment and the closer the distance between the obstacle and the unmanned ship, the higher the image acquisition quality, the clearer the texture, and the higher the image stitching accuracy. For the experimental scene of the obstacle 1, the image stitching accuracy is close to 90% when the obstacle distance is less than 8 m, and the accuracy is more than 80% when the obstacle distance is less than 14 m. For the experimental scene of the obstacle 2, the image stitching accuracy is close to 85% when the obstacle distance is less than 8 m, and the accuracy is also more than 80% when the obstacle distance is less than 14 m. The image stitching accuracy can meet the accuracy requirement of obstacle avoidance of the unmanned ship in the test. In the obstacle 1 and the clam wave environment, the method in this paper is compared with the methods of SITF [32], Harris [33], and Shape-Optimizing [34], and the results are shown in Figure 23. It can be seen that, with the increase in distance, due to the weakening of some edge features of obstacles in the image, the stitching accuracy of all methods gradually decreases, but the accuracy of this paper is always the best.
The stitching efficiency of the overwater and underwater obstacle images is mainly reflected in two aspects: one is the chain code acquisition time, and the other is the matching completion time. In this paper, the stitching efficiency of the normalized differential code and the traditional Freeman chain code are compared experimentally. The comparisons of their chain code acquisition time and matching completion time are shown in Table 5 and Table 6, respectively.
According to the data in the above two tables, it can be seen that the chain code acquisition time of the normalized differential code is longer than that of the Freeman chain code, but the time to complete a stitching matching of the normalized differential code is far less than that of the Freeman chain code. From the overall analysis of the stitching process, the processing speed of the normalized difference code method has great advantages. It can reduce some false matching and reduce the matching time, which is more suitable for the image stitching of the unmanned ship in this paper.

4.4. Experiment Results

Figure 24 shows the fusion detection results of the iceberg model obstacle using the image-driven Frustum–PointNet detection algorithm. The stitched image of the obstacle is used as the image driver and the point channel attention module and the channel attention module are added to ensure the overwater and underwater point cloud images can be quickly and accurately identified. In addition, it is fused with the obstacle edge contour image to obtain the 3D regression border of the obstacle, so as to realize the overall accurate positioning of the obstacle. Then, the obtained information, such as the direction, position, and size of the obstacle, is sent to the information display and parameter control system of the unmanned ship to realize the obstacle perception and obstacle avoidance control of the unmanned ship.
In order to verify the performance of the image-driven Frustum–PointNet algorithm proposed in this paper, the algorithm is compared with several other classical detection algorithms, including the MV3D algorithm [35], the AVOD algorithm [36], and the Frustum–PointNet algorithm [37] in the same experimental environment. Table 7 shows the comparison of their 3D detection accuracy. Figure 25 shows the relationship between accuracy and distance. It can be seen that the accuracy of each method decreases as the distance increases, while our method is still at a high level.
It can be seen from the data in the above table that, except the MV3D algorithm, the 3D detection accuracy of the remaining algorithms is greater than 80%. In addition, the detection accuracy of the image-driven Frustum–PointNet algorithms is superior to other algorithms. In addition, when the obstacle distance increases, the detection accuracy of each algorithm decreases. The reason is that, the farther the distance is, the systematic errors and random errors of each sensor in acquiring information will increase due to the influence of the environment, resulting in the decline in the performance of the algorithm. However, the algorithm in this paper can still maintain its superiority, and the performance index can also meet the experimental requirements.
The running speed of each fusion algorithm is also compared in this experiment, and the results are shown in Table 8.
As can be seen from the above table, the average inference time of the image-driven Frustum–PointNet algorithm proposed in this paper is 75 ms, and the running speed is 12.5 frames per second, which is better than other algorithms. The image-driven Frustum–PointNet algorithm not only improves the performance of obstacle detection, but also reduces the time consumption and improves the robustness of the algorithm by adding the attention module.

5. Conclusions

Aiming at the problem of lack of overall understanding of the complex environment above and under the water surface during the operation and maintenance of unmanned ships, a fusion technology of overwater and underwater visual images for unmanned ships is proposed in this paper. The software and hardware framework of the overwater and underwater visual image fusion system is constructed, the image processing methods in different environments are defined, and experimental research is carried out. The following conclusions are obtained in the experiment:
(1)
For the edge detection of the overwater and underwater images of obstacles, the false edges of the obstacles detected by the adaptive threshold algorithm proposed in this paper are greatly reduced, the loss of the true edges is relatively less, and the obtained edges are smoother. The algorithm retains the important edges of the images to the greatest extent, which lays a good foundation for the subsequent image stitching and fusion processing.
(2)
The image stitching accuracy of the obstacle model based on the extended neighborhood method can reach more than 85% within the obstacle distance of 8 m, and more than 80% within the obstacle distance of 14 m, which can meet the accuracy requirements of obstacle avoidance of unmanned ships.
(3)
The accuracy of the three-dimensional detection performance and three-dimensional positioning performance of the image-driven Frustum–PointNets algorithm proposed in this paper can reach more than 90%. The algorithm not only improves the performance of obstacle detection, but also reduces the time consumption and improves the robustness of the algorithm by adding the attention module.
(4)
In this paper, an overwater and underwater visual image stitching and fusion scheme is proposed, and the experimental verification research is carried out, which has achieved good results. The results are significant and have good reference value as they demonstrate that the stitching and fusion method can effectively obtain the comprehensive information of overwater and underwater objects for unmanned ships. However, because all research is based on laboratory conditions, it is often easy to ignore some influencing factors of actual sea conditions. Therefore, this algorithm may need to consider more actual scene conditions and carry out more detailed and in-depth research.

Author Contributions

Conceptualization, Y.C. and X.H.; Formal analysis, Y.C. and T.F.; Funding acquisition, X.H.; Investigation, W.C. and H.W.; Methodology, Y.C. and X.H.; Project administration, X.H.; Software, H.W. and W.C.; Supervision, X.H.; Validation, W.C. and H.W.; Writing—original draft, Y.C., H.W. and W.C.; Writing—review and editing, Y.C. and T.F. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Guangdong province basic and applied basic research fund project under grant 2019B151502057 and the national key research and development program of China under grant 2019YFB1804200.

Institutional Review Board Statement

All the agencies supported the work.

Informed Consent Statement

All authors agreed to submit the report for publication.

Data Availability Statement

We have obtained all the necessary permission.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sherman, K. The large marine ecosystem concept: Research and management strategy for living marine resources. Ecol. Appl. 1991, 1, 349–360. [Google Scholar] [CrossRef] [PubMed]
  2. Thirugnana, S.T.; Jaafar, A.B.; Yasunaga, T.; Nakaoka, T.; Ikegami, Y.; Su, S. Estimation of Ocean Thermal Energy Conversion Resources in the East of Malaysia. J. Mar. Sci. Eng. 2021, 9, 22. [Google Scholar] [CrossRef]
  3. Wang, Z.; Jiang, C.; Ai, Q.; Wang, C. The key technology of offshore wind farm and its new development in China. Renew. Sust. Energ. 2009, 13, 216–222. [Google Scholar]
  4. Abbassi, R.; Khan, F.; Garaniya, V.; Chai, S.; Chin, C.; Hossain, K.A. An integrated method for human error probability assessment during the maintenance of offshore facilities. Process. Saf. Environ. Prot. 2015, 94, 172–179. [Google Scholar] [CrossRef]
  5. Munim, Z.H. Autonomous ships: A review, innovative applications and future maritime business models. Supply Chain Forum An. Int. J. 2019, 20, 1–14. [Google Scholar] [CrossRef]
  6. Chen, J. The application of unmanned ships in smart water control. Mod. Inf. Technol. 2020, 4, 137–139. [Google Scholar]
  7. Wang, C.; Zhang, X.; Cong, L.; Li, J.; Zhang, J. Research on intelligent collision avoidance decision-making of unmanned ship in unknown environments. Evol. Syst. 2019, 10, 649–658. [Google Scholar] [CrossRef]
  8. Huang, Y.; Chen, L.; Chen, P.; Negenborn, R.R.; Gelder, P.H.A.J.M. Ship collision avoidance methods: State-of-the-art. Saf. Sci. 2020, 121, 451–473. [Google Scholar] [CrossRef]
  9. Xiao, G.; Ren, B.; Tong, C.; Hong, X. A Quantitative Evaluation Method for Obstacle Avoidance Performance of Unmanned Ship. J. Mar. Sci. Eng. 2021, 9, 1127. [Google Scholar] [CrossRef]
  10. Zhang, W.; Yang, C.; Jiang, F.; Gao, X.; Yang, K. A Review of Research on Light Visual Perception of Unmanned Surface Vehicles. J. Phys. Conf. Ser. 2020, 1606, 012022. [Google Scholar] [CrossRef]
  11. Zhuang, J.; Zhang, L.; Zhao, S.; Cao, J.; Wang, B.; Sun, H. Radar-based collision avoidance for unmanned surface vehicles. China Ocean Eng. 2016, 30, 867–883. [Google Scholar] [CrossRef]
  12. Abu-Tair, M.; Naeem, W. A decision support framework for collision avoidance of unmanned maritime vehicles. Commun. Comput. Info. Sci. 2013, 355, 549–557. [Google Scholar]
  13. Szpak, Z.L.; Tapamo, J.R. Maritime surveillance: Tracking ships inside a dynamic background using a fast level-set. Expert Sys. Appl. 2011, 38, 6669–6680. [Google Scholar] [CrossRef]
  14. Braginsky, B.; Guterman, H. Obstacle Avoidance Approaches for Autonomous Underwater Vehicle: Simulation and Experimental results. IEEE J. Ocean. Eng. 2016, 41, 882–892. [Google Scholar] [CrossRef]
  15. Fang, M.; Wang, S.; Wu, M.; Lin, Y. Applying the self-tuning fuzzy control with the image detection technique on the Obstacle-Avoidance for Autonomous Underwater Vehicles. Ocean Eng. 2015, 93, 11–24. [Google Scholar] [CrossRef]
  16. Gu, J.; Pyo, J.; Joe, H.; Kim, B.; Kim, J.; Cho, H.; Yu, S.C. A method for automatic detection of underwater objects using forward-looking imaging sonar. In Proceedings of the OCEANS 2015—MTS/IEEE Washington, Washington, DC, USA, 19–22 October 2015. [Google Scholar]
  17. Williams, D.P. Fast Target Detection in Synthetic Aperture Sonar Imagery: A New Algorithm and Large-Scale Performance Analysis. IEEE J. Ocean. Eng. 2015, 40, 71–92. [Google Scholar] [CrossRef]
  18. Raihan, A.J.; Abas, P.E.; De Silva, L.C. Role of Restored Underwater Images in Underwater Imaging Applications. Appl. Syst. Innov. 2021, 4, 96. [Google Scholar] [CrossRef]
  19. Berus, L.; Skakun, P.; Rajnovic, D.; Janjatovic, P.; Sidjanin, L.; Ficko, M. Determination of the Grain Size in Single-Phase Materials by Edge Detection and Concatenation. Metals 2020, 10, 1381. [Google Scholar] [CrossRef]
  20. Xiu, C.; Fang, J.; Zhang, J. Image Stitching Method Based on Adaptive Weighted Fusion. In Proceedings of the CCDC 2021: 33rd Chinese Control and Decision Conference, Kunming, China, 22–24 May 2021. [Google Scholar]
  21. Patel, K. Key Frame Extraction Based on Block Based Histogram Difference and Edge Matching Rate. Int. J. Sci. Eng. Technol. 2011, 1, 23–30. [Google Scholar]
  22. Lu, J.; Huo, G.; Cheng, J. Research on image stitching method based on fuzzy inference. Multimedia Tools Appl. 2022, 81, 1–12. [Google Scholar] [CrossRef]
  23. Pollard, S.; Pilu, M.; Hayes, S.; Lorusso, A. View synthesis by trinocular edge matching and transfer. Image Vis. Comput. 2000, 18, 749–757. [Google Scholar] [CrossRef]
  24. Li, J.; Ma, L.; Fan, Y.; Wang, N.; Duan, K.; Han, Q.; Zhang, X.; Su, G.; Li, C.; Tang, L. An Image Stitching Method for Airborne Wide-Swath HyperSpectral Imaging System Equipped with Multiple Imagers. Remote Sens. 2021, 13, 1001. [Google Scholar] [CrossRef]
  25. Selvaraj, A.; Ganesan, P. Infrared and visible image fusion using multiscale NSCT and rolling-guidance filter. IET Image Proc. 2020, 14, 1–10. [Google Scholar] [CrossRef]
  26. Ch, M.M.I.; Riaz, M.M.; Iltaf, N.; Ghafoor, A.; Ahmad, A. Weighted image fusion using cross bilateral filter and non-subsampled contourlet transform. Multidimens. Syst. Signal Proc. 2019, 30, 2199–2210. [Google Scholar] [CrossRef]
  27. Hayat, N.; Imran, M. Multi-exposure image fusion technique using multi-resolution blending. IET Image Proc. 2019, 13, 2554–2561. [Google Scholar] [CrossRef]
  28. Paramanandham, N.; Rajendiran, K. Swarm intelligence based image fusion for noisy images using consecutive pixel intensity. Multimed. Tools Appl. 2018, 77, 32133–32151. [Google Scholar] [CrossRef]
  29. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  30. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the CVPR 2017: 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  31. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the ICPR 2010: 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010. [Google Scholar]
  32. Cao, W. Applying image registration algorithm combined with CNN model to video image stitching. J. Supercomput. 2021, 77, 13879–13896. [Google Scholar] [CrossRef]
  33. Chen, K.; Wang, M. Image stitching algorithm research based on OpenCV. In Proceedings of the CCC 2014: 33rd Chinese Control Conference, Nanjing, China, 28–30 July 2014. [Google Scholar]
  34. Liu, S.; Chai, Q. Shape-optimizing and illumination-smoothing image stitching. IEEE Trans. Multimed. 2019, 21, 690–703. [Google Scholar] [CrossRef]
  35. Chen, X.; Ma, H.; Wan, J.; Li, B.; Xia, T. Multi-view 3D object detection network for autonomous driving. In Proceedings of the CVPR 2017: 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  36. Ku, J.; Mozifian, M.; Lee, J.; Harakeh, A.; Waslander, S.L. Joint 3D proposal generation and object detection from view aggregation. In Proceedings of the IROS 2018: IEEE International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018. [Google Scholar]
  37. Qi, C.R.; Liu, W.; Wu, C.; Su, H.; Guibas, L.J. Frustum PointNets for 3D Object Detection from RGB-D Data. In Proceedings of the CVPR 2018: 31st IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
Figure 1. The overall hardware framework of the visual image fusion system.
Figure 1. The overall hardware framework of the visual image fusion system.
Jmse 10 00747 g001
Figure 2. The overall software framework of the visual image fusion system.
Figure 2. The overall software framework of the visual image fusion system.
Jmse 10 00747 g002
Figure 3. Schematic diagram of image region distribution.
Figure 3. Schematic diagram of image region distribution.
Jmse 10 00747 g003
Figure 4. Flowchart of obstacle segmentation.
Figure 4. Flowchart of obstacle segmentation.
Jmse 10 00747 g004
Figure 5. Flowchart of the improved algorithm of image edge detection.
Figure 5. Flowchart of the improved algorithm of image edge detection.
Jmse 10 00747 g005
Figure 6. The eight-way code direction symbol.
Figure 6. The eight-way code direction symbol.
Jmse 10 00747 g006
Figure 7. Eight-way chain code direction representation.
Figure 7. Eight-way chain code direction representation.
Jmse 10 00747 g007
Figure 8. Eight-way chain code diagram (s is the starting point).
Figure 8. Eight-way chain code diagram (s is the starting point).
Jmse 10 00747 g008
Figure 9. Code point acquisition flowchart.
Figure 9. Code point acquisition flowchart.
Jmse 10 00747 g009
Figure 10. The establishment of the underwater image model of obstacles.
Figure 10. The establishment of the underwater image model of obstacles.
Jmse 10 00747 g010
Figure 11. Selection of encoding points for underwater images of obstacles.
Figure 11. Selection of encoding points for underwater images of obstacles.
Jmse 10 00747 g011
Figure 12. Calculation example of the normalized differential code of obstacle underwater image.
Figure 12. Calculation example of the normalized differential code of obstacle underwater image.
Jmse 10 00747 g012
Figure 13. Schematic diagram of channel attention module.
Figure 13. Schematic diagram of channel attention module.
Jmse 10 00747 g013
Figure 14. Schematic diagram of point attention module.
Figure 14. Schematic diagram of point attention module.
Jmse 10 00747 g014
Figure 15. Flowchart of the image-driven Frustum–PointNet detection algorithm.
Figure 15. Flowchart of the image-driven Frustum–PointNet detection algorithm.
Jmse 10 00747 g015
Figure 16. Unmanned ship visual image fusion experimental platform. (a) Obstacle 1: iceberg model. (b) Obstacle 2: pontoon model.
Figure 16. Unmanned ship visual image fusion experimental platform. (a) Obstacle 1: iceberg model. (b) Obstacle 2: pontoon model.
Jmse 10 00747 g016
Figure 17. The control interface of the experiment platform.
Figure 17. The control interface of the experiment platform.
Jmse 10 00747 g017
Figure 18. Obstacle image segmentation results. (a) Overwater image of obstacle. (b) Region growing segmentation result. (c) Automatic multi-seed segmentation result. (d) Underwater image of obstacle. (e) Region growing segmentation result. (f) Automatic multi-seed segmentation result.
Figure 18. Obstacle image segmentation results. (a) Overwater image of obstacle. (b) Region growing segmentation result. (c) Automatic multi-seed segmentation result. (d) Underwater image of obstacle. (e) Region growing segmentation result. (f) Automatic multi-seed segmentation result.
Jmse 10 00747 g018
Figure 19. Comparison of edge detection results of different methods. (a) Edge detection of the obstacle 1 by traditional Canny algorithm. (b) Edge detection of the obstacle 1 by adaptive threshold algorithm of this paper. (c) Edge detection of the obstacle 2 by traditional Canny algorithm. (d) Edge detection of the obstacle 2 by adaptive threshold algorithm of this paper.
Figure 19. Comparison of edge detection results of different methods. (a) Edge detection of the obstacle 1 by traditional Canny algorithm. (b) Edge detection of the obstacle 1 by adaptive threshold algorithm of this paper. (c) Edge detection of the obstacle 2 by traditional Canny algorithm. (d) Edge detection of the obstacle 2 by adaptive threshold algorithm of this paper.
Jmse 10 00747 g019
Figure 20. Comparison of the average peak signal-to-noise ratio of different edge detection threshold algorithms.
Figure 20. Comparison of the average peak signal-to-noise ratio of different edge detection threshold algorithms.
Jmse 10 00747 g020
Figure 21. Comparison of stitching results of the obstacles. (a) Physical picture of the obstacle 1. (b) Encoding points selection of the obstacle 1. (c) Stitching result of the obstacle 1. (d) Physical picture of the obstacle 2. (e) Encoding points selection of the obstacle 2. (f) Stitching result of the obstacle 2.
Figure 21. Comparison of stitching results of the obstacles. (a) Physical picture of the obstacle 1. (b) Encoding points selection of the obstacle 1. (c) Stitching result of the obstacle 1. (d) Physical picture of the obstacle 2. (e) Encoding points selection of the obstacle 2. (f) Stitching result of the obstacle 2.
Jmse 10 00747 g021
Figure 22. Experimental results of image stitching accuracy. (a) Image stitching accuracy of the obstacle 1. (b) Image stitching accuracy of the obstacle 2.
Figure 22. Experimental results of image stitching accuracy. (a) Image stitching accuracy of the obstacle 1. (b) Image stitching accuracy of the obstacle 2.
Jmse 10 00747 g022
Figure 23. Comparison of stitching accuracy of different methods.
Figure 23. Comparison of stitching accuracy of different methods.
Jmse 10 00747 g023
Figure 24. Fusion detection results of the iceberg model obstacle. (a) Overwater point cloud image of the obstacle. (b) Overwater 3D regression border of the obstacle. (c) Underwater point cloud image of the obstacle. (d) Underwater 3D regression border of the obstacle.
Figure 24. Fusion detection results of the iceberg model obstacle. (a) Overwater point cloud image of the obstacle. (b) Overwater 3D regression border of the obstacle. (c) Underwater point cloud image of the obstacle. (d) Underwater 3D regression border of the obstacle.
Jmse 10 00747 g024
Figure 25. Relationship between 3D detection accuracy and distance.
Figure 25. Relationship between 3D detection accuracy and distance.
Jmse 10 00747 g025
Table 1. Complementary relationship of eight-way chain code values.
Table 1. Complementary relationship of eight-way chain code values.
Complementary Relationship
code value012345670
Complementary code value456701234
Table 2. Establishment of subcode space.
Table 2. Establishment of subcode space.
Original Code5-Bit Code Long Subspace4-Bit Code Long Subspace3-Bit Code Long Subspace2-Bit Code Long Subspace1-Bit Code Long Subspace
01234012340123, 1234, 2340, 3401, 4012012, 123, 234, 340, 40101, 12, 23, 34, 400, 1, 2, 3, 4
Table 3. The model and characteristics of the equipment.
Table 3. The model and characteristics of the equipment.
ModelPhysical MapCharacteristics
CameraDS-2CD1211-I3 Jmse 10 00747 i001With infrared sensing function, the lens aperture is 4mm, the horizontal field of view is 81°, the resolution can reach 1920 × 1080@25fps, the current mainstream video transmission protocols can be supported, and the adaptive network port can be switched between 10 M or 100 M.
LidarPandar-40p Jmse 10 00747 i002It has 40 sets of laser transceiver pairs, which can be rotated at all angles to obtain 3D point cloud information of the target in real time, and its static ranging accuracy is less than ±2 cm.
Underwater cameraUC5032 Jmse 10 00747 i003The shell is made of 316 L stainless steel, which is corrosion-resistant and can adapt to complex underwater environments. The maximum pressure-resistant water depth is 60 m. The 2.8–12 zoom lens can capture fast-moving objects at a frame rate of 60. It supports current mainstream protocols.
SonarFarSounder-500 Jmse 10 00747 i004Its horizontal viewing angle can reach 90°, vertical angle can reach 60°, and the operating frequency is 61 kHz. It enables single ping 3D imaging.
GPSBD970 GNSS Jmse 10 00747 i005Obtains the exact latitude and longitude of the drone.
3D Electronic compassSEC345 Jmse 10 00747 i006Obtains the heading angle, roll angle, and pitch angle of the unmanned ship.
Table 4. Experiment running environment information.
Table 4. Experiment running environment information.
CategoryNameModelPerformance
hardwareCPUIntel I7-6700k4 GHz × 4
RAMDDR42400 MHZ 32 GB
GPUNVIDIA GTX 1060(6G)4.4TFLOPs of single-precision floating-point operations
softwaresystemWindows 10
environmentPython3.6
DL frameworkPytorch1.2.0
Table 5. Chain code acquisition time comparison.
Table 5. Chain code acquisition time comparison.
Environmental ConditionsFreeman Chain CodeNormalized Differential Code
No wave326 ms351 ms
Calm wave341 ms380 ms
Smooth wave369 ms404 ms
Slight wave390 ms426 ms
Table 6. Comparison of the average time to complete a stitching matching.
Table 6. Comparison of the average time to complete a stitching matching.
OperationFreeman Chain CodeNormalized Differential Code
Matching search336 ms121 ms
Stitching completed379 ms145 ms
Similarity search235 ms101 ms
Table 7. Comparison of 3D detection accuracy of different fusion algorithms.
Table 7. Comparison of 3D detection accuracy of different fusion algorithms.
Fusion Algorithm3D Detection Accuracy When Obstacle Distance Is 4 m3D Detection Accuracy When Obstacle Distance Is 8 m3D Detection Accuracy When Obstacle Distance Is 12 m
MV3D75.98%74.58%72.98%
AVOD85.21%84.71%82.56%
Frustum–PointNet84.39%83.01%81.59%
Image-driven Frustum–PointNet92.12%91.90%90.17%
Table 8. Speed comparison of different fusion algorithms.
Table 8. Speed comparison of different fusion algorithms.
Fusion AlgorithmInference Time (ms)Running Speed (Frames Per Second)
MV3D3202.6
AVOD1309.4
Frustum–PointNet1505.5
Image-driven Frustum–PointNet7512.5
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Chen, Y.; Hong, X.; Chen, W.; Wang, H.; Fan, T. Experimental Research on Overwater and Underwater Visual Image Stitching and Fusion Technology of Offshore Operation and Maintenance of Unmanned Ship. J. Mar. Sci. Eng. 2022, 10, 747. https://doi.org/10.3390/jmse10060747

AMA Style

Chen Y, Hong X, Chen W, Wang H, Fan T. Experimental Research on Overwater and Underwater Visual Image Stitching and Fusion Technology of Offshore Operation and Maintenance of Unmanned Ship. Journal of Marine Science and Engineering. 2022; 10(6):747. https://doi.org/10.3390/jmse10060747

Chicago/Turabian Style

Chen, Yuanming, Xiaobin Hong, Weiguo Chen, Huifang Wang, and Tianhui Fan. 2022. "Experimental Research on Overwater and Underwater Visual Image Stitching and Fusion Technology of Offshore Operation and Maintenance of Unmanned Ship" Journal of Marine Science and Engineering 10, no. 6: 747. https://doi.org/10.3390/jmse10060747

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop