Next Article in Journal
Star Image Prediction and Restoration under Dynamic Conditions
Next Article in Special Issue
Filtered Multitone Modulation Underwater Acoustic Communications Using Low-Complexity Channel-Estimation-Based MMSE Turbo Equalization
Previous Article in Journal
Three-to-One Internal Resonance in MEMS Arch Resonators
Previous Article in Special Issue
Real-Time Monocular Visual Odometry for Turbid and Dynamic Underwater Environments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visual Navigation for Recovering an AUV by Another AUV in Shallow Water

1
State Key Laboratory of Robotics, Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
2
Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(8), 1889; https://doi.org/10.3390/s19081889
Submission received: 28 February 2019 / Revised: 16 April 2019 / Accepted: 18 April 2019 / Published: 20 April 2019

Abstract

:
Autonomous underwater vehicles (AUVs) play very important roles in underwater missions. However, the reliability of the automated recovery of AUVs has still not been well addressed. We propose a vision-based framework for automatically recovering an AUV by another AUV in shallow water. The proposed framework contains a detection phase for the robust detection of underwater landmarks mounted on the docking station in shallow water and a pose-estimation phase for estimating the pose between AUVs and underwater landmarks. We propose a Laplacian-of-Gaussian-based coarse-to-fine blockwise (LCB) method for the detection of underwater landmarks to overcome ambient light and nonuniform spreading, which are the two main problems in shallow water. We propose a novel method for pose estimation in practical cases where landmarks are broken or covered by biofouling. In the experiments, we show that our proposed LCB method outperforms the state-of-the-art method in terms of remote landmark detection. We then combine our proposed vision-based framework with acoustic sensors in field experiments to demonstrate its effectiveness in the automated recovery of AUVs.

1. Introduction

Autonomous underwater vehicles (AUVs) are of great interest to oceanographers and navies for marine scientific research and reconnaissance due to their wide scope of activity and high efficiency. However, long-duration missions struggle with the limited on-board energy and storage capability of AUVs. The autonomous recovery of AUVs is highly desirable for recharging, data transfer, and relaunch. Various recovery systems have been proposed by previous works, and they usually contain a docking station. Docking stations can either be stationary [1,2,3,4], such as landed on the seafloor, or mobile [5,6,7,8,9,10], such as carried by Unmanned Surface Vehicles (USVs) [5] and submarines [7,8,9]. Stationary docking stations are able to perform recharging and data transfer for AUVs, but in fixed locations. Mobile docking stations are more flexible. They can not only recharge and transfer data almost anywhere, but also take AUVs back to home harbors. All recovery systems aim at the precise physical contact between docking stations and AUVs. Precise contact can either be passive or active. Passive contact is to capture AUVs by using custom mechanisms. An AUV was captured by a robotic manipulator when it is in the docking envelope of submarines in [9]. Active contact means that AUVs actively connect to docking stations with the help of custom mechanisms. A common kind of active contact is that AUVs catch a cable [5,8] or pole [11,12,13] with an incorporated custom latch or hook. This form allows AUVs to approach the cable or pole in any direction, but requires additional space to install the mechanism on the AUVs, which limits the wide application to a range of AUVs. The other kind of active contact is using funnel-like docking stations [1,3,14,15,16]. Funnel-like docking stations allow for trajectory errors during docking without much refitting on AUVs. AUVs autonomously and actively move into the funnel-like docking station in this way. In this work, we address the recovery of AUVs by actively docking them with a funnel-like docking station carried by another slow-moving mother AUV.
The position sensing of docking stations is one of the most crucial modules for AUV recovery, no matter what forms of recovery are deployed. Three kinds of sensors are available for the position sensing of docking stations: (1) electromagnetic sensors, (2) acoustic sensors, and (3) optical sensors. The main concerns for AUV recovery are sensing range and resolution. Acoustic sensors have the longest sensing range, up to several kilometers, but the lowest resolution, with practically more than 1.5 m error [13]. Electromagnetic sensors have a better resolution than acoustic sensors, but much less range, about 30–50 m [17]. Optical sensors are superior in resolution, but inferior in propagation range. Their sensing range depends on water-body turbidity. It was verified in [18] that it is possible for optical sensors to guide AUV docking in the range of 10–15 m, even in turbid water. Landmarks are usually leveraged for optical-sensor-based visual navigation in recovering AUVs. They may either be passive or active landmarks. Passive landmarks are patterns generally drawn on boards. Active landmarks, such as light beacons, emit energy, having higher visibility compared to passive landmarks. LED beacons with 460-nm wavelength were used in our work due to their relatively good propagation in water. Positional errors of optical sensors, depending on the algorithms, could be less than 0.01 m [16,18]. Taking advantage of their positional accuracy, optical sensors are usually in charge of the final short-range stage precise positioning and combined with acoustic sensors [13,15,19], which are good in sensing range, but bad at resolution. Following this strategy, in this work, we also combined acoustic and optical sensors for AUV recovery. We focused on the development of visual-based AUV recovery (VBAR) algorithms using optical sensors in shallow water since many acoustic sensors for underwater positioning are now commercially available.
VBAR algorithms consist of a detection phase and pose-estimation phase. The detection phase localizes underwater landmarks in 2D images once landmarks appear in the view of the camera. Detection of landmarks in shallow water severely suffers from ambient light and nonuniform spreading [20]. Ambient light in open shallow water mainly comes from the Sun. The problem of ambient light in shallow water is more acute than in deep water. Previous methods for landmark detection in deep water are no longer effective in shallow water due to severe ambient light. Nonuniform landmark spreading is mainly caused by the camera’s view points. The luminance of landmarks close to the camera is stronger than those away from the camera. The developed methods so far have not dealt with the aforementioned problems particularly well. Some previous attempts took advantage of the luminance feature of landmarks for detection. An empirical fixed threshold was used for detection of landmarks by binarizing the underwater image in [1,21]. Binarizing underwater images with a fixed threshold is vulnerable to ambient light and nonuniform spreading. Underwater images were binarized with a fixed threshold within the region of interest (ROI), computed by a convolutional neural network in [16]. It overcomes ambient light to some extent. A histogram-based adaptive thresholding method was proposed for landmark detection in [20]. We show in Section 4 that this adaptive method is effective at landmark detection for short distances, but fails for remote distances. Remote detection of landmarks can provide AUVs with more time and space to adjust their pose for docking, which vastly increases docking success rate. AUVs cannot enormously adjust their pose within a short distance owing to their inherent mobility. We propose a Laplacian-of-Gaussian-based coarse-to-fine blockwise (LCB) method for the detection of underwater landmarks to overcome ambient light and nonuniform spreading.
The pose-estimation phase recovers the 3D relative position and orientation between landmarks and AUVs from detected landmarks on 2D images. Methods of pose estimation proposed for AUV recovery may be classified as based on points, lines, and curvature features. The pose was estimated by fitting lines and elliptical curves in [13,20], respectively. A points-based feature is the most commonly used due to its flexibility. A landmark is viewed as a feature point. A points-based feature was leveraged in [1,3,4,9,16,19,21] for pose estimation. Perspective-n-points (PnP) algorithms can accurately solve points-based pose estimation, but a drawback of PnP algorithms is that they fail in estimating poses in cases where not all predefined landmarks are detected, namely incomplete observations. However, an incomplete observation is highly possible in long-duration underwater tasks and may be caused by broken landmarks and biofouling coverage on landmarks in practice. In order to solve this problem, we propose a novel pose-estimation method for incomplete observations in AUV recovery. To the best of our knowledge, we are the first to solve the problem of pose estimation for incomplete observations in PnP-based VBAR.
Our contributions are four-fold:
  • Detection of landmarks in shallow water suffers from ambient light and nonuniform spreading. In order to overcome them, we propose an LCB method for the detection of underwater landmarks.
  • Incomplete observations of all predefined landmarks that may be caused by broken landmarks and biofouling coverage comprise a practical problem in pose estimation for recovering AUVs in long-duration missions. A novel pose-estimation method for incomplete observations in AUV recovery is proposed in this work. To the best of our knowledge, the proposed method is the first points-based method for pose estimation in VBAR algorithms.
  • Field experiments were performed to evaluate the proposed LCB method and the pose-estimation method.
  • We provide an underwater active landmark dataset that was collected in our field experiments. The 2D location of each landmark was manually labeled for each image. The dataset can help researchers develop and validate their underwater vision-based recovery or docking algorithms in the absence of related underwater infrastructures.

2. System Overview

In this section, we introduce an overview of our recovery system. We illustrate the process of recovery in Figure 1a,b. The recovery system consisted of a mother vessel and a sub-AUV (SAUV) for recovery. The SAUV was attached to the docking station, physically and autonomously carried by the mother vessel, for recovery by employing underwater docking. The docking station was rigidly fixed to the underbelly of the mother vessel. The entrance of the docking station was funnel-like, 2 m in diameter, and equipped with several active landmarks mounted around its rim, as shown in Figure 1c. These landmarks were used for detecting and computing the position of the docking station, whose details are given in Section 3. Blue LEDs with a 460-nm wavelength were employed as our landmarks due to their good propagation in water. We utilized another AUV as the mother vessel in our system. The mother vessel could also be in other forms, such as unmanned surface vehicles (USVs).
The SAUV was a torpedo-shaped vehicle that was 384 mm in diameter and 5486 mm in length. It was equipped with a control computer, Doppler velocity log (DVL), inertial measurement unit (IMU), GPS, battery units, and motors. An acoustic sensor, Evologics S2C R, was used for rough long-range positioning, and our vision-based underwater-docking module for short-range precision guidance was installed in the SAUV, as shown in Figure 1d. A N a n o S e a C a m ® monocular camera with a frame rate of 20 fps was rigidly mounted on the SAUV head for capturing underwater images. The captured images were fed into an embedded computer, Jetson TX2, for detecting the docking station and computing its pose (Section 3). Jetson TX2 is a power-efficient embedded computer with a 6-core CPU, 8 Gb RAM, and a 256-core GPU. The computed pose information was further transmitted to the SAUV controller for short-range precision navigation.

3. Underwater-Docking Recovery Algorithm

Our proposed underwater-docking algorithm for recovery consisted of two phases. The first phase was landmark detection, and the second was pose estimation. In the detection phase, the detector localizes 2D landmark coordinates on images once the landmarks appear in the field of view of the camera. The pose-estimation phase recovers the relative 3D pose, including position and orientation, between the docking station and SAUV. The estimated pose is then transmitted to the control computer of the SAUV for short-range precision navigation.

3.1. Landmark Detection

Landmarks in our task were LEDs actively emitting energy. The main problems in the detection of landmarks are nonuniform spreading and complex ambient light. In order to overcome these problems, we proposed an LCB method for landmark detection. In the coarse detection phase, we deployed a convolutional neural network, DoNN, proposed in [16], to detect the bounding box of the landmarks. Then, we proposed a Laplacian-of-Gaussian (LoG) filter that utilizes blob features to overcome ambient light. Blockwise segmentation on filtered images was also performed to deal with nonuniform spreading. We show that the proposed detector outperformed state-of-the-art methods in the remote detection of landmarks in Section 4.
Coarse detection phase. We leveraged a convolutional neural network, DoNN, proposed in [16], to detect the landmark bounding box. DoNN showed superior performance in robust bounding box detection of underwater landmarks in various underwater environments in [16]. Hence, DoNN was employed to detect the bounding box of all underwater landmarks, which is also known as coarse detection. We also validated the importance of coarse detection performed by DoNN by comparing the performance in Table 2. Rows 3 and 5 in the table show that the method leveraging coarse detection significantly outperformed the method that did not take advantage of coarse detection. Instead of first detecting the bounding box of all the landmarks, an alternative solution is to detect each individual landmark directly. However, the solution is not as stable and robust as detecting the pattern formed by all the landmarks. There are many noisy luminaries that are similar to landmarks in terms of their appearance in water. Noisy luminaries may come from other light sources from the external environment and mirror images of landmarks reflected from the water surface due to total internal reflection, as given in [1,16]. If detection is performed on individual landmarks, noisy luminaries become potential detection results. Instead, the pattern formed by all landmarks represent the concept of the docking station, which is more stable and robust for detection. For the above reasons, the bounding box of all landmarks was detected in our work.
For the sake of completeness, we briefly introduce DoNN. DoNN contains nine convolution layers and seven pooling layers. It maps captured underwater images to a vector v = { x , y , w , h , s } , D o N N : ( I ) v . I denotes the captured underwater image. { x , y } , { w , h } represent the position and size of the predicted bounding box, respectively. s is the confidence score predicted by DoNN for the bounding box. DoNN first divides the input underwater image into G × G grids and then predicts multiple bounding boxes encoded by { x , y , w , h } along with the confidence score s for each grid. Finally, the bounding box with the highest confidence score s is selected as the final predicted bounding box. To summarize, the phase of coarse detection outputs the predicted confidence score and landmark bounding box.
LoG filter After coarse detection by DoNN, the bounding box of landmarks is localized in the image. The following processes were performed to localize landmarks within the image patch given by the bounding box. LEDs actively emitting energy were designed as our landmarks due to their good propagation in water. Since such landmarks are quite bright, most previous works proposed to detect and extract landmarks by taking advantage of the luminance features of landmarks. Methods based on luminance features suffer from ambient light, and the problem becomes more acute in shallow water, as shown in Figure 2a,b. Besides luminance, landmarks have blob-like structures that were neglected by previous works. Blob features were employed to overcome the interference of strong ambient light in shallow water.
The LoG filter is a differential blob detector. It is defined as:
L o G ( x , y ) 2 G σ ( x , y ) = 2 x 2 G σ ( x , y ) + 2 y 2 G σ ( x , y ) = x 2 + y 2 2 σ 2 π σ 4 e x p ( x 2 + y 2 2 σ 2 ) ,
where σ is a standard deviation of Gaussian filter G σ ( x , y ) , 2 is the Laplacian operator, and ( x , y ) are image coordinates. Images filtered by LoG can be implemented by first convolving the image with a Gaussian filter G σ ( x , y ) and then applying the Laplacian operator 2 to the image filtered by G σ ( x , y ) , since:
f ( x , y ) = 2 ( G σ ( x , y ) f ( x , y ) ) = ( 2 G σ ( x , y ) ) f ( x , y ) = L o G ( x , y ) f ( x , y ) ,
where f ( x , y ) is the pixel value of the image at ( x , y ) and G σ ( x , y ) is a Gaussian kernel defined as:
G σ ( x , y ) = 1 2 π σ 2 e x p ( x 2 + y 2 2 σ 2 ) .
The Laplacian operator is very sensitive to noise since it is a second derivative measurement on the image. To counter this case, a Gaussian filter was leveraged to smooth the image and attenuate high-frequency noise.
LoG filters obtain a strong response for dark blobs of σ extent on a light background or light blobs on a dark background when LoG filters convolve with images. The “ 3 σ ” rule indicates that 99 % of the energy of a Gaussian is concentrated within three standard deviations of its mean. Hence, the standard deviation σ is usually computed as σ = ( s 1 ) / 3 to detect blobs of radius s [22]. Since the radius s of the blobs in our task is in the range of 10–25, the performance of our method with σ { 3 , 5 , 7 , 9 } was analyzed. Figure 3 shows the convolution of LoG filters ( σ { 3 , 5 , 7 , 9 } ) with landmarks in one dimension for ease of understanding. The response obtained its strongest value at the peak of the original signal. The response became smooth with the increment of σ .
Blockwise segmentation Nonuniform spreading is primarily due to the viewpoint of the camera, as illustrated in Figure 2c,d, posing great challenges to the extraction of landmarks from f ( x , y ) . Landmark images away from the camera are less bright than landmarks close to the camera, which results in a weak response in f ( x , y ) . To overcome this problem, we proposed to segment filtered images f ( x , y ) in a blockwise manner. Filtered image f ( x , y ) was first equally divided into four non-overlapping blocks so that the illumination of each block was approximately uniform. Then, we applied a global threshold method to each block for segmentation. Here, Otsu’s global segmentation method [23] was adopted for blockwise segmentation, which can work well since illumination is nearly uniform.

3.2. Pose Estimation

The pose is defined as the relative 3D position, represented by X = ( x , y , z ) , and orientation, represented by Euler angles (yaw, pitch, roll) between landmarks and the SAUV, as shown in Figure 4. Pose estimation is defined as the process of recovering the relative pose between landmarks and the SAUV from the 2D coordinates of landmarks localized by the LCB method. Landmarks in 2D underwater images were detected and localized in Section 3.1. In this section, we first introduce the method to recover the relative pose between landmarks and the SAUV in the case where all predefined landmarks are observed, namely complete observation. Then, our proposed method is provided for pose estimation in the case where not all predefined landmarks can be observed, namely incomplete observation. Pose estimation can be solved by the perspective-n-point (PnP) algorithm for the complete observation of landmarks. However, the PnP algorithm cannot work for the case of an incomplete observation. Incomplete observation of landmarks may be caused by broken landmarks and biofouling coverage on landmarks in practice. Pose estimation in the case of incomplete observations is imperative for the long-term use of unmanned systems. Previous works [1,16,21] that used the PnP algorithm required a complete observation of landmarks for pose estimation. The pose cannot be estimated in incomplete observations in these works. To our best knowledge, we are the first to solve the problem of pose estimation for the case of incomplete observations in points-based underwater recovery algorithms.

3.2.1. Pose Estimation in the Case of Complete Observations

In order to estimate the pose between landmarks and the SAUV, three coordinate frames are first defined, as illustrated in Figure 4. The first coordinate frame is the camera coordinate frame, which is a 3D coordinate frame. Its origin is the optical center of the camera. Another coordinate frame is a reference frame attached to the docking station. Its origin resides in the geometric center of all landmarks. The third coordinate frame is the image coordinate frame, which is a 2D frame established on the image plane. The relative pose between landmarks and the SAUV is described by a transformation between the camera and reference coordinate frame. The relative pose is estimated by determining the transformation.
In the following, we detail the determination of the transformation between the camera and reference coordinate frame. Assume a pinhole camera is used; the transformation between the image and the camera coordinate frame can be expressed as a set of linear equations:
λ u v 1 = Λ X c = ϕ x γ δ x 0 0 ϕ y δ y 0 0 0 1 0 x c y c z c 1 ,
where λ is a scale factor, ( u , v ) R 3 is the coordinates in the image frame, ( x c , y c , z c ) R 3 is the coordinates in the camera frame, and Λ R 3 × 4 is the intrinsic matrix of the inherent parameters of a camera, mapping 3D points in the camera frame to 2D points in the image. Λ contains five parameters. ( δ x , δ y ) R 2 is the principal point, whose unit is pixels. ϕ x R and ϕ y R represent the scaling factor converting space metrics to pixel units. γ is the skew coefficient. The intrinsic matrix Λ of a camera can be calibrated before using the camera for pose estimation.
A matrix, called extrinsic matrix E, relates the camera frame and the reference frame. It is defined by:
x c y c z c 1 = E X r = R T 0 1 x r y r z r 1 ,
where E R 4 × 4 is the extrinsic matrix, ( x r , y r , z r ) R 3 are the coordinates of the reference frame, and R and T are the rotation and translation matrix, respectively. The relative 3D position and orientation of the relative pose can be obtained from T and R, respectively.
Combining Equations (4) and (5), the relationship between the image frame and the reference frame is:
λ x = λ u v 1 = Λ E x r y r z r 1 = Λ E w = Ω w .
Given a known intrinsic matrix Λ , n 3D points w = ( x r , y r , z r ) in the reference frame, and their corresponding projections x = ( u , v ) in the image, rotation matrix R and translation matrix T can be computed. This problem is known as the PnP problem. A unique solution can be obtained when n 4 . Given n point correspondences, Equation (6) can be solved using direct linear method “DLT” [24], which is a noniterative method. Noniterative methods are efficient, but less accurate and stable in the presence of noise. Iterative methods enhance solution accuracy by minimizing a nonlinear error function, such as image space error [25,26] and object space error [27]. However, the computational cost of the iterative methods is usually high. RPnP [28] is a noniterative method that is as accurate as iterative algorithms with much less computational time. Hence, RPnP was employed in our work for pose estimation. The main idea of RPnP consists of three steps.
Firstly, n reference points are divided into ( n 2 ) three-point subsets. According to the law of cosines, each point pair ( Q i , Q j ) in the subset gives constraint:
f i j ( d i , d j ) = d i 2 + d j 2 2 d i d j c o s θ i j d i j 2 = 0 ,
where d i = | | O c Q i | | , d j = | | O c Q j | | , d i j = | | Q i Q j | | and θ i j is the view angle, as illustrated in Figure 5. Q i and Q j are 3D reference points. q i and q j are the projection of Q i and Q j on the image plane, respectively. Based on Equation (7), we obtain the following polynomial system:
f 12 ( d 1 , d 2 ) = 0 , f 13 ( d 1 , d 3 ) = 0 , f 23 ( d 2 , d 3 ) = 0 ,
for each subset. Equation (8) is then converted to a fourth-degree polynomial:
g k ( x ) = a 1 x 4 + a 2 x 3 + a 3 x 2 + a 2 x + a 1 , k = 1 , 2 , ... , ( n 2 )
for the k th subset.
Second, a cost function G = k = 1 n 2 g k 2 ( x ) is defined as the square sum of the polynomials in Equation (9). It was demonstrated in [28] that cost function G has four minima at most.
Finally, the minima of G were explored by finding the roots of its first derivative G = k = 1 n 2 g k ( x ) g k ( x ) = 0 , which is a seven-degree polynomial. Rotation matrix R and translation matrix T were computed from each local minimum, and the result with the least reprojection residual was selected as the final solution.
Rotation matrix R was then further converted to Euler angles represented by (yaw, pitch, roll). Translation matrix T denotes the coordinates of point O r in the camera frame. The Euler angles and translation matrix T were passed to the SAUV controller for navigation.

3.2.2. Pose Estimation in the Case of Incomplete Observations

Pose estimation can be solved by normal PnP algorithms using 2D-3D-point correspondences in the case of complete observations. However, PnP algorithms cannot give correct solutions if some landmarks are not observed without knowing which landmarks are missing. In this subsection, we propose a method to estimate pose in the case of incomplete observations by minimizing the reprojection error.
In order to estimate pose in the case of incomplete observations, the method first identifies which landmarks are missing. Then, the normal PnP algorithm can be used to compute the pose after identifying missing landmarks. Each landmark was numbered as its identification as shown in Figure 6a. The identification of all detected landmarks is denoted by a vector i m = { 1 , 2 , 3 , , m } , where m is the number of predefined landmarks. For example, vector i 8 = { 1 , 2 , 3 , , 7 } means that there are eight landmarks in total mounted on the docking station, but the eighth landmark cannot be observed due to some reasons. Computing identification of landmarks i m is not hard in complete observations. Take the configuration of landmarks in Figure 6b as an example. The configuration consists of eight landmarks. The upper-right connected component in the binary images obtained by the detection phase corresponds to the first landmark. The upper-left connected component corresponds to the eighth landmark. The rest can be done in the same manner. The 2D-3D-point correspondences are gained in this way in the case of complete observations. However, this method does not work for the identification of landmarks in the case of incomplete observations. We propose to identify landmarks in the case of incomplete observations by finding an identification vector i m that minimizes the reprojection error, that is,
a r g m i n i m j = 1 k ( Ω i m w j x j i m ) 2 ,
where k ( 4 k m ) denotes the number of landmarks detected in the detection phase, w j are the 3D coordinates of the jth landmark defined in Equation (6), x j i m are the 2D coordinates of the jth landmarks determined by i m , and Ω i m is the projection matrix in Equation (6), which is computed by using w j and x j i m . To be specific, k is computed by analyzing connected components in the binary image obtained in the detection phase. Then, vector i m that minimizes the reprojection error is searched in the space of k landmarks. To search for the optimal i m , x j i m was first computed for each i m . Then, Ω i m was computed using the corresponding x j i m and w j . Next, w j was projected to the image plane by Ω i m , and the corresponding reprojection error was computed. Finally, the i m with the minimal reprojection error was the optimal i m . After obtaining the optimal i m , the relative pose can be computed using normal PnP algorithms.

4. Experiment Results

In this section, we give the experiment results of our proposed vision-based underwater recovery framework for AUVs. We first show that our proposed detection method outperformed the state-of-art method in landmark remote detection. Next, the LCB method is demonstrated on an underwater-docking dataset. Finally, we jointly validate our proposed method for detection and pose estimation in field experiments.

4.1. Experiment Detection Results

4.1.1. Comparison of Detection: Algorithm Performance

A histogram-based adaptive thresholding (HATS) method was proposed for landmark detection in [20]. We compare our proposed LCB method with HATS in this subsection. The comparison results are shown in Figure 7. Comparison images were consecutive frames taken in an AUV recovery task. HATS argues that landmark images contain a direct-path component and a glow component resulting from the scattering and absorption effects of the water body. The direct-path component carries much larger power than the glow component. Hence, the direct-path component corresponds to a peak in the tail of the histogram, as shown in Rows 5 and 6 of Figure 7. However, we found that HATS performed well in landmark detection at short distances, but failed at remote distances, because pixels of direct landmark components dominate the image at short distances, as shown in Rows 5 and 6 of Figure 7. The peak corresponding to the direct component was obvious in the tail of the histogram, but this was not the case at remote distances. The direct component of a landmark may only correspond to several pixels in the image at remote distances, resulting in a flat tail in the histogram, as illustrated in Rows 1–4 of Figure 7. In this case, landmarks cannot be detected using HATS, whereas remote detection is imperative for recovery since the remote detection of landmarks can provide the SAUV with more time and space to adjust its pose for docking, which vastly increases the docking success rate. Figure 7 shows that the LCB algorithm was effective for landmark detection at both short and remote distances, outperforming the state-of-the-art HATS method.
We further compared the LCB and HATS methods on a dataset, namely D r e c o v e r y (Publicly available at https://drive.google.com/open?id=1m3OIQmtYUEwRmA-rN_Xg_eBYaUM1rarz), which contains 2306 images of active underwater landmarks. D r e c o v e r y was collected in shallow-water field experiments with the landmarks and docking station deployed in [16]. Eight landmarks ( m = 8 ) were mounted on the docking station. The coordinates of each landmark in the 2D image were manually labeled as the ground truth in D r e c o v e r y . Detection was regarded as a failed detection if the number of detected landmarks was not equal to the number of predefined landmarks ( k m ). This can be regarded as an incomplete observation and solved by our proposed pose-estimation method if the number of detected landmarks is less than the number of predefined landmarks ( 4 k < m ). The Euclidean distance of each landmark between the location detected by the LCB algorithm and ground truth was computed for measuring location error if k = m . The location error was evaluated, as it significantly affected the accuracy of the pose-estimation algorithm given the known intrinsic matrix Λ and n 3D points w = ( x r , y r , z r ) in the reference frame, as illustrated by (6). The comparison results are shown in Table 1. For the methods in Rows 1 and 2 in Table 1, the threshold was computed using HATS if a peak existed in the tail of the histogram. Otherwise, a fixed threshold was used. The first line shows results in which HATS was applied to the whole image. Both the successful detection and location errors were very poor. This was primarily due to the ambient light illustrated in the third column of Figure 7. HATS was applied to the ROI detected in the coarse detection phase in the second row of Table 1. Both failed detection and location error were improved thanks to coarse detection. Location error was 1.91 , which is acceptable for successful detection, but the number of failed detections was very high. This was because HATS only works at short distances, as illustrated in the beginning of this subsection. The third row of Table 1 shows the results of our proposed LCB method. It performed much better than the two previous methods, in both the number of failed detection and location errors.

4.1.2. Detection-Performance Analysis

The LCB method consists of three phases: (1) coarse detection, (2) LoG filtering, and (3) blockwise segmentation, as described in Section 3.1. In this section, we discuss the contribution of each phase to the detection performance of active underwater landmarks. Analysis was performed on D r e c o v e r y . Analysis results are shown in Table 2. The first row of Table 2 shows the detection performance by globally applying Otsu’s method to the image. Otsu’s method failed in the detection of all images in D r e c o v e r y because it is severely affected by ambient light. Otsu’s method assumes that there are two classes of pixels that follow a bimodal histogram in images and finds a threshold to minimize the intraclass variance. Otsu’s method tends to segment the region close to the light source instead of individual landmarks, as shown in Figure 8b. The method globally applying LoG and then using Otsu’s method on the filtered image in the second row of Table 2 performed better than Otsu’s method in terms of resistance to ambient light, but poorly in nonuniform spreading, as seen in Figure 8c. Applying Otsu’s method to the image filtered by LoG block-wisely (Row 3 in Table 2) still separated the image in blocks that did not contain landmarks into two classes, introducing much blob noise, as shown in Figure 8d. Subsequently, we discuss the performance of methods with coarse detection (Rows 4–8). Comparing the method in the second and fourth rows, the number of failed detections was decreased by 75 images by coarse detection. Comparing the method in the fourth and fifth rows, the number of failed detections dramatically dropped, from 322 to 9 images, by adopting blockwise segmentation. We compare the performance of methods with various σ in Rows 5–8 of Table 2. Results showed that the method with σ = 3 had the worst performance, and other methods ( σ = 5 , 7 , 9 ) performed almost the same. The poor performance of the method with σ = 3 was primarily due to a small standard deviation in the LoG filter being sensitive to noise.

4.2. Pose-Estimation Algorithm Analysis

In this subsection, we analyze the accuracy of the pose-estimation algorithm and its sensitivity to the location error introduced by the detection in 2D images. Since it is very hard to obtain the underwater ground truth of the relative pose between the SAUV and the docking station, accuracy was evaluated by ground experiments. In the ground experiments, a docking station of one meter in diameter was kept stationary on the ground, and a camera captured landmark images mounted on the docking station with various positions and orientations. To estimate the relative pose, the 2D location of each landmark in the 2D images was first annotated. Next, the relative pose using the pose-estimation algorithm was computed to assess accuracy in the absence of location error. Then, two levels of Gaussian noise, with standard deviations σ = 3 and σ = 5 , were added to the annotated 2D location of each landmark to evaluate the accuracy of the pose-estimation algorithm in the presence of location errors in different degrees. Note that the intrinsic parameters of the camera used in the ground experiments were different from those used in water due to a different optical medium. Figure 9 and Table 3 (Part of the data in the table was published in our previous work [16].) show images for estimating the pose and corresponding results, respectively. Table 3 gives the mean estimation results averaged over 1000 trials for each degree of noise. It is shown in Table 3 that the mean orientation error was 1.823 degrees, and the mean position error was 6.306 mm in the absence of location error, that is σ = 0 . The mean orientation and position error increased to 2.168 degrees and 7.039 mm, respectively, as the noise level went to σ = 3 . With σ = 5 , the mean orientation and position error became 2.770 degrees and 9.818 mm, respectively. The aforementioned results validate the accuracy of the pose-estimation algorithm and its robustness to noise.

4.3. Field Experiments

We demonstrate our proposed LCB method and pose-estimation method by the automated recovery of AUVs in field experiments. Field experiments were conducted on an experimental ship, as shown in Figure 10a, anchored in Qiandao Lake, China, using systems given in Section 2. The method in the fifth row of Table 2 was employed for landmark detection. In order to validate our proposed pose-estimation method for both complete and incomplete observations, field experiments were performed with two configurations of landmarks. One was complete observations, that is all landmarks could be observed ( k = m = 8 ), as shown in Figure 11a. The other was to simulate incomplete observations ( 4 k < m ). To this end, a landmark was removed in advance, as shown in Figure 11b, but without letting the pose-estimation algorithm know this prior information.
Both the mother vessel (see Figure 10b,d) and the SAUV (see Figure 10b,c) started running at a random position with a random heading. The distance between them was about 500 m. The mother vessel ran near the water surface. Both acoustic and optical sensors were combined for recovery in the field experiments. The acoustic sensor was utilized for rough navigation at a long distance. Once the SAUV received the navigation information provided by our proposed vision-based underwater recycling algorithm, the SAUV switched to using it for navigation. Two runs were performed for each landmark configuration. The SAUV was recovered successfully by the mother vessel in all of these runs. Figure 12 shows the trajectories of all four of these runs. Figure 13, Figure 14, Figure 15 and Figure 16 give the images of the recycling process. Runs 1 and 2 demonstrate the LCB algorithm and pose-estimation method in complete observations. We validated the LCB algorithm and pose-estimation method for the incomplete observation in Runs 3 and 4. The average running time of the proposed framework was 0.17 s per frame.

5. Conclusions

A vision-based framework was proposed in this work for recovering an AUV with another AUV in shallow water. The framework consisted of a detection phase and a pose-estimation phase. In order to overcome challenges resulting from ambient light and nonuniform spreading in shallow water, an LCB algorithm was developed for the detection of underwater landmarks. Experiments showed that the LCB algorithm outperformed the state-of-the-art method in terms of location error. For the pose-estimation phase, a novel pose-estimation method was proposed for the incomplete observation of all underwater landmarks, which is practical, but has been neglected by most previous works. Pose estimation for incomplete observations enables the recovery of AUVs, even if some landmarks are broken or occluded by biofouling, which is very likely in long-duration missions. In ground experiments, we observed that the mean position and orientation error were 1.823 degrees and 6.306 mm, respectively, in the absence of noise, and 2.770 degrees and 9.818 mm, respectively, in the presence of strong noise. Field experiments were performed to recover a sub-AUV by a mother vessel in a lake using our proposed framework. The field experiments validated that our developed framework was effective for recovering AUVs in shallow water.

Author Contributions

Methodology, S.L.; resources, H.X.; validation, L.G.; supervision, Y.L.

Funding

This research was funded by Defense Technology Innovation Funds (No. 18-H863-00-TS-002-034-01).

Acknowledgments

We thank Chuang Yu, Xiaojun Han, Bofeng Liu, and Bin Yang for their help in field experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Park, J.Y.; Huan Jun, B.; Mook Lee, P.; Oh, J. Experiments on vision guided docking of an autonomous underwater vehicle using one camera. Ocean Eng. 2009, 36, 48–61. [Google Scholar] [CrossRef]
  2. Maki, T.; Shiroku, R.; Sato, Y.; Matsuda, T.; Sakamaki, T.; Ura, T. Docking method for hovering type AUVs by acoustic and visual positioning. In Proceedings of the 2013 IEEE International Underwater Technology Symposium, Tokyo, Japan, 5–8 March 2013; pp. 1–6. [Google Scholar] [CrossRef]
  3. Li, Y.; Jiang, Y.; Cao, J.; Wang, B.; Li, Y. AUV docking experiments based on vision positioning using two cameras. Ocean Eng. 2015, 110, 163–173. [Google Scholar] [CrossRef]
  4. Palomeras, N.; Vallicrosa, G.; Mallios, A.; Bosch, J.; Vidal, E.; Hurtos, N.; Carreras, M.; Ridao, P. AUV homing and docking for remote operations. Ocean Eng. 2018, 154, 106–120. [Google Scholar] [CrossRef]
  5. Sarda, E.I.; Dhanak, M.R. A USV-Based Automated Launch and Recovery System for AUVs. IEEE J. Ocean. Eng. 2017, 42, 37–55. [Google Scholar] [CrossRef]
  6. Raspante, F. Underwater mobile docking of autonomous underwater vehicles. In Proceedings of the OCEANS 2012 Conference, Hampton Roads, VA, USA, 21–24 May 2012; pp. 1–15. [Google Scholar]
  7. Irani, R.; Kehoe, D.; Spencer, W.; Watt, G.; Gillis, C.; Carretero, J.; Dubay, R. Towards a UUV launch and recovery system on a slowly moving submarine. In Proceedings of the Warship 2014: Naval Submarines & UUV’s, Bath, UK, 18–19 June 2014; pp. 93–103. [Google Scholar]
  8. Renilson, M. A simplified concept for recovering a UUV to a submarine. Underw. Technol. 2014, 32, 193–197. [Google Scholar] [CrossRef] [Green Version]
  9. Watt, G.D.; Roy, A.R.; Currie, J.; Gillis, C.B.; Giesbrecht, J.; Heard, G.J.; Birsan, M.; Seto, M.L.; Carretero, J.A.; Dubay, R.; et al. A concept for docking a UUV with a slowly moving submarine under waves. IEEE J. Ocean. Eng. 2016, 41, 471–498. [Google Scholar]
  10. Brizzolara, S.; Chryssostomidis, C. Design of an Unconventional ASV for Underwater Vehicles Recovery: Simulation of the motions for operations in rough seas. In Proceedings of the ASNE International Conference on Launch & Recovery, Linthicum, MD, USA, 14–15 November 2012. [Google Scholar]
  11. Singh, H.; Bellingham, J.G.; Hover, F.; Lemer, S.; Moran, B.A.; Von der Heydt, K.; Yoerger, D. Docking for an autonomous ocean sampling network. IEEE J. Ocean. Eng. 2001, 26, 498–514. [Google Scholar] [CrossRef]
  12. Wirtz, M.; Hildebrandt, M.; Gaudig, C. Design and test of a robust docking system for hovering AUVs. In Proceedings of the 2012 Oceans, Arlington, VA, USA, 22–24 October 2012; pp. 1–6. [Google Scholar]
  13. Kimball, P.W.; Clark, E.B.; Scully, M.; Richmond, K.; Flesher, C.; Lindzey, L.E.; Harman, J.; Huffstutler, K.; Lawrence, J.; Lelievre, S.; et al. The ARTEMIS under-ice AUV docking system. J. Field Robot. 2018, 35, 299–308. [Google Scholar] [CrossRef]
  14. Pai, S.; Guerrini, P.; Potter, J.; Maguer, A.; Chitre, M.; Biagini, S. Autonomous initial capture system for AUV recovery. In Proceedings of the 3rd International Conference and Exhibition on Underwater Acoustic Measurements: Technologies and Results, Nafplion, Greece, 21–26 June 2009. [Google Scholar]
  15. Li, D.J.; Chen, Y.H.; Shi, J.G.; Yang, C.J. Autonomous underwater vehicle docking system for cabled ocean observatory network. Ocean Eng. 2015, 109, 127–134. [Google Scholar] [CrossRef]
  16. Liu, S.; Ozay, M.; Okatani, T.; Xu, H.; Sun, K.; Lin, Y. Detection and Pose Estimation for Short-Range Vision-Based Underwater Docking. IEEE Access 2019, 7, 2720–2749. [Google Scholar] [CrossRef]
  17. Feezor, M.D.; Yates Sorrell, F.; Blankinship, P.R.; Bellingham, J.G. Autonomous underwater vehicle homing/docking via electromagnetic guidance. IEEE J. Ocean. Eng. 2001, 26, 515–521. [Google Scholar] [CrossRef]
  18. Cowen, S.; Briest, S.; Dombrowski, J. Underwater docking of autonomous undersea vehicles using optical terminal guidance. In Proceedings of the Oceans ’97. MTS/IEEE Conference, Halifax, NS, Canada, 6–9 October 1997; Volume 2, pp. 1143–1147. [Google Scholar] [CrossRef]
  19. Sato, Y.; Maki, T.; Masuda, K.; Matsuda, T.; Sakamaki, T. Autonomous docking of hovering type AUV to seafloor charging station based on acoustic and visual sensing. In Proceedings of the 2017 IEEE Underwater Technology (UT), Busan, Korea, 21–24 February 2017; pp. 1–6. [Google Scholar]
  20. Ghosh, S.; Ray, R.; Vadali, S.R.K.; Shome, S.N.; Nandy, S. Reliable pose estimation of underwater dock using single camera: A scene invariant approach. Mach. Vis. Appl. 2016, 27, 221–236. [Google Scholar] [CrossRef]
  21. Hong, Y.H.; Kim, J.Y.; Oh, J.H.; Lee, P.M.; Jeon, B.H.; Oh, K.H. Development of the homing and docking algorithm for AUV. In Proceedings of the Thirteenth International Offshore and Polar Engineering Conference, Honolulu, HI, USA, 25–30 May 2003. [Google Scholar]
  22. Kong, H.; Akakin, H.C.; Sarma, S.E. A generalized Laplacian of Gaussian filter for blob detection and its applications. IEEE Trans. Cybern. 2013, 43, 1719–1733. [Google Scholar] [CrossRef] [PubMed]
  23. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Trans. Syst. Man Cybern. 1979, 9, 62–66. [Google Scholar] [CrossRef]
  24. ABDEL-AZIZ, Y. Direct linear transformation from comparator coordinates into object space in close-range photogrammetry. In Proceedings of the ASP Symposium on Close-Range Photogrammetry, Falls Church, VA, USA, 11 December 1971; pp. 1–18. [Google Scholar]
  25. Araújo, H.; Carceroni, R.L.; Brown, C.M. A Fully Projective Formulation to Improve the Accuracy of Lowe’s Pose-Estimation Algorithm. Comput. Vis. Image Underst. 1998, 70, 227–238. [Google Scholar] [CrossRef]
  26. Oberkampf, D.; DeMenthon, D.; Davis, L.S. Iterative Pose Estimation Using Coplanar Feature Points. Comput. Vis. Image Underst. 1996, 63, 495–511. [Google Scholar] [CrossRef]
  27. Lu, C.; Hager, G.D.; Mjolsness, E. Fast and globally convergent pose estimation from video images. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 610–622. [Google Scholar] [CrossRef]
  28. Li, S.; Xu, C.; Xie, M. A Robust O(n) Solution to the Perspective-n-Point Problem. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 1444–1450. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Autonomous underwater vehicle (AUV) recovery illustrated by (a) the simulation model and (b) field experiments, captured by an unmanned aerial vehicle. (c) Landmarks mounted on the docking-station rim. (d) Camera and acoustic sensor on the AUV head.
Figure 1. Autonomous underwater vehicle (AUV) recovery illustrated by (a) the simulation model and (b) field experiments, captured by an unmanned aerial vehicle. (c) Landmarks mounted on the docking-station rim. (d) Camera and acoustic sensor on the AUV head.
Sensors 19 01889 g001
Figure 2. (a,b) Ambient light in AUV recovery in shallow water; (c,d) nonuniform spreading of landmarks in AUV recovery.
Figure 2. (a,b) Ambient light in AUV recovery in shallow water; (c,d) nonuniform spreading of landmarks in AUV recovery.
Sensors 19 01889 g002
Figure 3. (First row) 1D pixel values along the black dotted line in the red bounding box. (Second row) Negatives of the Laplacian-of-Gaussian (LoG filters) with σ = 3 , σ = 5 , σ = 7 , and σ = 9 , from left to right. (Third row) Corresponding results obtained by convolving the original signal in the first row with the filter in the second row.
Figure 3. (First row) 1D pixel values along the black dotted line in the red bounding box. (Second row) Negatives of the Laplacian-of-Gaussian (LoG filters) with σ = 3 , σ = 5 , σ = 7 , and σ = 9 , from left to right. (Third row) Corresponding results obtained by convolving the original signal in the first row with the filter in the second row.
Sensors 19 01889 g003
Figure 4. Coordinate frames in recovery.
Figure 4. Coordinate frames in recovery.
Sensors 19 01889 g004
Figure 5. Constraint given by Equation (7).
Figure 5. Constraint given by Equation (7).
Sensors 19 01889 g005
Figure 6. (a) Original image patch obtained by the coarse detection phase. (b) Definition of identification for (a). (c) Binary image of (a), computed by our proposed Laplacian-of-Gaussian-based coarse-to-fine blockwise (LCB) method, introduced in Section 3.1.
Figure 6. (a) Original image patch obtained by the coarse detection phase. (b) Definition of identification for (a). (c) Binary image of (a), computed by our proposed Laplacian-of-Gaussian-based coarse-to-fine blockwise (LCB) method, introduced in Section 3.1.
Sensors 19 01889 g006
Figure 7. (First column) Distance-ordered landmark images captured in a run of underwater recycling. The distance between camera and docking station in the z c direction is shown in the lower right corner of each image. The green rectangle indicates the region of interest (ROI) predicted by the coarse detection phase. (Second column) Histograms of the whole image and ROI. The histogram in red corresponds to the whole image, indicated by the red rectangle. The histogram in green corresponds to the ROI, indicated by the green rectangle. (Third and fourth column) Detection results from applying HATS to the whole image (third column) and ROI (fourth column). HATS is only effective and used in the case where a peak exists in the tail (Rows 5 and 6). For cases in which a peak does not exist in the tail, a fixed threshold is used for segmentation. The fixed threshold was set to 180 in Rows 1–4. (Fifth column) Detection results of our proposed method ( σ = 5 ).
Figure 7. (First column) Distance-ordered landmark images captured in a run of underwater recycling. The distance between camera and docking station in the z c direction is shown in the lower right corner of each image. The green rectangle indicates the region of interest (ROI) predicted by the coarse detection phase. (Second column) Histograms of the whole image and ROI. The histogram in red corresponds to the whole image, indicated by the red rectangle. The histogram in green corresponds to the ROI, indicated by the green rectangle. (Third and fourth column) Detection results from applying HATS to the whole image (third column) and ROI (fourth column). HATS is only effective and used in the case where a peak exists in the tail (Rows 5 and 6). For cases in which a peak does not exist in the tail, a fixed threshold is used for segmentation. The fixed threshold was set to 180 in Rows 1–4. (Fifth column) Detection results of our proposed method ( σ = 5 ).
Sensors 19 01889 g007
Figure 8. (a) Original image. (bd) Results obtained by using the method in the first, second, and third row of Table 2, respectively.
Figure 8. (a) Original image. (bd) Results obtained by using the method in the first, second, and third row of Table 2, respectively.
Sensors 19 01889 g008
Figure 9. Landmarks for pose estimation in ground experiments.
Figure 9. Landmarks for pose estimation in ground experiments.
Sensors 19 01889 g009
Figure 10. (a) Experiment ship in field experiments. (bd) Mother vessel and sub-AUV (SAUV) in field experiments.
Figure 10. (a) Experiment ship in field experiments. (bd) Mother vessel and sub-AUV (SAUV) in field experiments.
Sensors 19 01889 g010
Figure 11. Docking stations with eight and seven landmarks.
Figure 11. Docking stations with eight and seven landmarks.
Sensors 19 01889 g011
Figure 12. Trajectories of four recycling runs.
Figure 12. Trajectories of four recycling runs.
Sensors 19 01889 g012
Figure 13. Images of the recycling process of Run 1.
Figure 13. Images of the recycling process of Run 1.
Sensors 19 01889 g013
Figure 14. Images of the recycling process of Run 2.
Figure 14. Images of the recycling process of Run 2.
Sensors 19 01889 g014
Figure 15. Images of the recycling process of Run 3.
Figure 15. Images of the recycling process of Run 3.
Sensors 19 01889 g015
Figure 16. Images of the recycling process of Run 4.
Figure 16. Images of the recycling process of Run 4.
Sensors 19 01889 g016
Table 1. Performance comparison of histogram-based adaptive thresholding (HATS) and LCB methods.
Table 1. Performance comparison of histogram-based adaptive thresholding (HATS) and LCB methods.
MethodNumber of Failed DetectionsLocation Errors (Pixel per Landmark)
1. HATS190718.13
2. HATS (ROI)13701.91
3. LCB ( σ = 5 )91.68
Table 2. Detection performance of various methods.
Table 2. Detection performance of various methods.
MethodNumber of Failed DetectionsLocation Errors (Pixel per Landmark)
1. Otsu2306N/A
2. LoG ( σ = 5 ) + Otsu3971.83
3. LoG ( σ = 5 ) + Block-wise + Otsu17157.39
4. Coarse detection + LoG ( σ = 5 ) + Otsu3221.64
5. Coarse detection + LoG ( σ = 5 ) + Blockwise + Otsu91.68
6. Coarse detection + LoG ( σ = 3 ) + Blockwise + Otsu261.67
7. Coarse detection + LoG ( σ = 7 ) + Blockwise + Otsu101.70
8. Coarse detection + LoG ( σ = 9 ) + Blockwise + Otsu111.70
Table 3. Ground pose-estimation performance. Results in the table correspond to the figures given in Figure 9.
Table 3. Ground pose-estimation performance. Results in the table correspond to the figures given in Figure 9.
No.Ground Truth σ = 0 σ = 3 σ = 5
Orien. (deg)Pos. (mm)Orien. (deg)Pos. (mm)Orien. (deg)Pos. (mm)Orien. (deg)Pos. (mm)
(a)Yaw: 0.0x: 99.21.23110.3660.96099.9160.29398.862
Pitch: 4.5y: −73.62.809−71.5171.932−72.2162.069−71.717
Roll: −0.8z: 3686.5−1.2733687.955−1.2583679.886−1.2443666.954
(b)Yaw: 6.8x: 98.12.11596.5591.19695.3181.12795.288
Pitch: −9.2y: 839.8−11.302837.126−11.911834.394−12.314832.257
Roll: −1.3z: 3598.1−2.1423590.579−2.1353598.303−2.1283572.824
(c)Yaw: 2.7x: 435.8−3.714433.207−4.509431.473−4.443430.364
Pitch: 12.2y: −1036.216.827−1029.84516.277−1028.30116.198−1024.642
Roll: −1.2z: 3553.8−1.9153538.244−1.8413531.964−1.8983518.536
(d)Yaw: 18.9x: −1212.822.741−1213.57622.503−1211.53722.265−1206.448
Pitch: −0.4y: −167.83.573−168.6603.067−168.7343.050−167.966
Roll: −2.3z: 3453.7−2.0423460.211−2.2373453.575−2.1833438.054
(e)Yaw: −37.0x: 515.5−35.471513.549−33.238518.459−29.508522.176
Pitch: 1.9y: −422.32.576−427.3952.563−430.4593.374−430.230
Roll: −2.9z: 4098.4−2.5714088.068−2.6244115.437−2.8174120.387
(f)Yaw: −28.3x: −393.6−27.256−394.989−26.977−395.335−25.288−396.326
Pitch: 2.5y: −374.23.573−374.4063.472−375.1743.615−377.715
Roll: −1.2z: 3307.9−1.3633304.232−1.3283309.443−1.4023329.823
(g)Yaw: −12.0x: −1296.5−11.813−1284.283−11.366−1285.865−10.421−1290.430
Pitch: 2.6y: −327.83.991−323.6704.033−324.0824.006−325.703
Roll: −1.9z: 3102.0−2.3173073.298−2.3263078.568−2.3313090.632
(h)Yaw: 10.4x: −59.84.375−62.9594.274−63.1643.078−63.927
Pitch: 22.8y: −966.925.979−972.44023.945−973.17922.672−971.425
Roll: −3.0z: 3728.8−3.4203745.472−3.5083741.764−3.6653730.413
(i)Yaw: 3.2x: 12.3−2.5013.982−2.12220.59310.69023.763
Pitch: 10.1y: −46.38.912−38.89813.662−38.7627.017−57.425
Roll: 5.4z: 3889.04.9653881.1856.7493898.2192.8793899.503
(j)Yaw: −1.4x: 252.5−5.338259.961−4.975245.1767.377239.694
Pitch: −17.6y: 823.0−16.601831.225−16.833814.238−19.823810.849
Roll: 1.8z: 3954.41.7333962.1130.9193945.5194.3243964.525
Mean ErrorN/AN/A1.8236.3062.1687.0392.7709.818

Share and Cite

MDPI and ACS Style

Liu, S.; Xu, H.; Lin, Y.; Gao, L. Visual Navigation for Recovering an AUV by Another AUV in Shallow Water. Sensors 2019, 19, 1889. https://doi.org/10.3390/s19081889

AMA Style

Liu S, Xu H, Lin Y, Gao L. Visual Navigation for Recovering an AUV by Another AUV in Shallow Water. Sensors. 2019; 19(8):1889. https://doi.org/10.3390/s19081889

Chicago/Turabian Style

Liu, Shuang, Hongli Xu, Yang Lin, and Lei Gao. 2019. "Visual Navigation for Recovering an AUV by Another AUV in Shallow Water" Sensors 19, no. 8: 1889. https://doi.org/10.3390/s19081889

APA Style

Liu, S., Xu, H., Lin, Y., & Gao, L. (2019). Visual Navigation for Recovering an AUV by Another AUV in Shallow Water. Sensors, 19(8), 1889. https://doi.org/10.3390/s19081889

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop