Skip to Content
JMSEJournal of Marine Science and Engineering
  • Article
  • Open Access

15 December 2021

An Underwater Visual Navigation Method Based on Multiple ArUco Markers

,
,
,
and
1
Maritime College, Guangdong Ocean University, Zhanjiang 524000, China
2
School of Engineering, Newcastle University, Newcastle upon Tyne NE1 7RU, UK
*
Author to whom correspondence should be addressed.

Abstract

Underwater navigation presents crucial issues because of the rapid attenuation of electronic magnetic waves. The conventional underwater navigation methods are achieved by acoustic equipment, such as the ultra-short-baseline localisation systems and Doppler velocity logs, etc. However, they suffer from low fresh rate, low bandwidth, environmental disturbance and high cost. In the paper, a novel underwater visual navigation is investigated based on the multiple ArUco markers. Unlike other underwater navigation approaches based on the artificial markers, the noise model of the pose estimation of a single marker and an optimal algorithm of the multiple markers are developed to increase the precision of the method. The experimental tests are conducted in the towing tank. The results show that the proposed method is able to localise the underwater vehicle accurately.

1. Introduction

The oceans cover 71% of the earth. Unmanned Underwater Vehicles (UUVs) are widely implemented to explore the oceans. The reliable underwater localisation systems are essential to increase the efficiency of the UUVs in underwater missions. In contrast to the localisation and navigation methods applied by land robotics, underwater navigation is a challenging problem, mainly because the electromagnetic wave attenuates rapidly in the water. It indicates that the methods of localisation or communications depending on the electromagnetic wave are unsuitable in underwater environments. Therefore, conventional methods of positioning of UUVs are achieved by acoustic beacons, such as Ultra-Short-Baseline Localisation (USBL) or Long Baseline Localisation (LBL). However, these devices require additional infrastructure, which means the cost is fairly high [1]. In addition to that, the acoustic equipment, including the Doppler Velocity Log (DVL), suffers from a low refresh rate, low bandwidth and noises in underwater environments.
There are many reports on solving underwater navigation issues using visual techniques [2,3,4,5]. Researchers fused the visual information of the seabed with other data collected by inertial sensors to predict the positions and attitudes of the UUVs [6,7]. However, the limitations such as insufficient illumination, scattering, and refraction, prevent underwater visual navigation methods from being implemented.
In this paper, a novel underwater localisation method based on the multiple ArUco markers has been investigated. The ArUco markers, reported by Jurado, can be detected automatically, and their positions relative to the camera are obtained as well [8]. While the ArUco markers were intended to serve for the augmented reality applications originally, it is widely utilised to locate the mobile robots [9,10]. In the proposed method, multiple markers with known positions are set out on the ground of a towing tank. After calibration of the camera in water, the positions and attitudes of the camera relative to the markers can be obtained by inverting the transformation matrices. Subsequently, an optimal algorithm will be used to update the position of the camera in the global coordinate system. Since the camera is fitted on the UUV, the positions of the UUVs are obtained by multiplying the geometric transformation matrices. The practical experiments have been conducted to prove that the proposed method is effective while being low-cost. The method is expected to be applied in underwater pipeline tracking or areas that are occupied by artificial structures. After placing the markers on the structures, the approach can estimate the positions and altitudes of UUVs precisely in the area.
The paper is structured as follows. The related work is outlined in Section 2. The introduction of ArUco markers and the testing platform are presented in Section 3 and Section 4. Subsequently, the brief theory and the process of the underwater camera calibration are shown in Section 5. In Section 6, the methodology is described in detail, and the layout of the markers is shown in Section 7. The experimental results and discussion are presented in Section 8. Section 9 concludes the approach and discusses the future work.

3. Introduction of ArUco Markers

There are three well-known artificial fiducial marker systems: ARToolKit [26], AprilTags [27], and ArUco [8]. Compared with others, the ArUco algorithm is able to detect and track the markers quickly and reliably. As shown in Figure 1, an adaptive threshold selection algorithm is implemented by computing the average value of the surrounding pixels of the specific pixel. In this way, all contours in the image are expected to be found, and non-square contours are filtered out. Followed by the projection operation using recognised contours, the code extraction, marker identification and error correction are processed. Eventually, the unique ID and the pose of the marker relative to the camera are estimated.
Figure 1. The ArUco Algorithm.

4. Testing Platform

The modified Videoray Pro 3 ROV was used to collect the data in the towing tank [28]. A stereo camera (Intel T265 Tracking camera) and an onboard computer (Latte Panda) are installed inside the watertight tube, which is mounted on the bottom of the Videoray Pro 3, as shown in Figure 2. During the collection process, the vehicle was operated remotely by a controller to follow specific paths. After tests, the collected data can be transferred to the terminal via WiFi.
Figure 2. VideoRay Pro 3 with tube.

5. Underwater Camera Calibration

The underwater camera calibration is essential for visual navigation systems. It not only can handle the distortion issue induced by the lenses but is also used to reduce the refraction effect. The camera calibration problem is solved by minimising the cost function [29]:
E = 1 2 i = 1 m [ H ( Θ , g i ) p i ] 2 ,
where a grid based on a chessboard consists of m points of g i coordinates with corresponding values p i . The term H is a function mapping grid points into the image plane. The Θ involves intrinsic, distortion, and extrinsic parameters.
The Levenberg–Marquardt approach can be applied to minimise the cost Function (1), solving for Θ . In order to minimise the cost function effectively, the initial estimates of camera parameters are processed by Direct Linear Translation (DLT) [30].
In practice, the camera calibration is processed by OpenCV tools, which are developed based on [31]. In Figure 3a, a classic black-white chessboard was placed on the bottom of the towing tank, and the vehicle was operated remotely to obtain images of the chessboard from different views. The process of the calibration through the chessboard pattern is shown in Figure 3b. The comparison is illustrated in Figure 3c,d. They display the original image and the calibrated image, respectively.
Figure 3. (a) The chessboard on the ground and the vehicle are operated to capture the images. (b) The calibration is processed based on the captured patterns. (c) The original image. (d) The same image after being calibrated.

6. Methodology

In the proposed method, the positions of the camera relative to different single markers are derived by using transformation matrices. Because the positions of the markers in the global coordinate system are known beforehand, the position of the camera in the global coordinate system is expected to be estimated. However, the positions measured by different markers do not indicate the real value due to the noise of underwater imaging. Hence, the noise model for the estimation by a single marker is constructed, and an optimal algorithm is designed to combine the predicted poses from the multiple markers to increase the accuracy of the localisation, as shown in Figure 4.
Figure 4. The flow chart of the method.

6.1. Position Estimated by the Single Marker

The ArUco libraries from OpenCV were implemented in this navigation method. By calling the estimatedPoseSingleMarkers function in the libraries, the rotation vector and translational vector of the marker relative to the camera were expected to be obtained. With the Rodrigues’ formula [32], the rotation vector was converted to the associated rotation matrix. Hence, the transformation matrix c a m e r a T m a r k e r was derived. However, m a r k e r T c a m e r a was needed in order to acquire the position and the attitude of the camera in a marker’s coordinate system.
In the free-noise phenomenon, the m a r k e r T c a m e r a can be derived by,
m a r k e r T c a m e r a = c a m e r a T m a r k e r 1 .
Because
c a m e r a T m a r k e r = R m c t m c 0 1
Then,
m a r k e r T c a m e r a = R m c T R m c T t m c 0 1 .
where u m = R m c T t m c is the translational vector of the camera relative to the marker. The term R m c T is the rotation matrix of the camera relative to the marker.
Before investigation about the white noise situation, the Lie algebra for the rotation matrix should be discussed. The R m c has nine individual entries, and eight degrees of freedom, because it is in the Special Orthogonal Group, i.e., R S O ( 3 ) and R T R = I . The Lie algebra ϕ , a 3-element vector, can be used to present the rotation matrix R m c [33]. Since the mapping from the rotation matrix R to Lie algebra ϕ complies with the Rodrigues’ formula [32], the Lie algebra ϕ actually equals the rotation vector.
The exponential mapping for R and ϕ is,
R = e x p ( ϕ ) .
The ϕ denotes the skew-symmetric matrix of the vector ϕ . The Lie algebra ϕ ¯ of the rotation matrix R T is,
ϕ ¯ = ( ln ( R T ) ) = ( ln e x p ( ϕ ) 1 ) = ϕ
Hence, the c a m e r a T m a r k e r can be presented by { ϕ m c ,   t m c } , the m a r k e r T c a m e r a can be presented by { ϕ m c ,   e x p ( ϕ m c ) t m c } .
As mentioned before, the noise caused by underwater imaging has been introduced into the translational vector t m c , and rotation matrix ϕ m c . According to [30,34],
{ ϕ ^ m c , t ^ m c } N ( { ϕ m c , t m c } , Σ mc )
and,
Σ mc = Q 1 0 0 Q 2
That means the { ϕ m c , t m c } obeys the Gaussian distribution. The Σ m c is a diagonal positive semidefinite matrix. The Q 1 and Q 2 are 3 × 3 matrices. The diagonal entries are constructed by the squared exponential covariance function [35], which is,
k i ( r i ) = e x p ( r i 2 2 l i 2 )
where the k i ( r i ) is the kernel function associated with the covariance function. The r i is the distance or radius on the xyz axis estimated by a single ArUco marker. The Σ = d i a g ( k i 1 1 ) . In this case, the | r i | 0 , the Σ i 0 . The l i is a hyperparameter defining the characteristic length-scale. In the paper, it is selected as 1 for estimated coordinates and 0.5 for the estimated rotation vectors. That means the smaller estimated value is more reliable.
Hence, according to [36], the probabilistic model of { ϕ c m , t c m } can be expressed as,
{ ϕ ^ c m , t ^ c m } N ( { ϕ m c , e x p ( ϕ m c ) t m c } , Σ c m )
where the ϕ ^ c m and t ^ are random variables and,
Σ cm = Q 1 0 0 M
However, the derivation of M is a little trivial. The e x p ( ϕ ^ m c ) is a nonlinear transform with respect to random variable ϕ ^ m c , and e x p ( ϕ ^ m c ) t ^ m c is a production of dual Gaussian distributions. According to [36], M can be,
M = [ e x p ( ϕ c m ) t c m ) e x p ( ϕ c m ) ] Q 2 [ e x p ( ϕ c m ) t c m ) e x p ( ϕ c m ) ] T
Now the derivation of position estimation by a single ArUco marker in the white noise situation is complete. In the next part, an optimal algorithm is described, which is used to increase the accuracy of the method based on the multiple markers.

6.2. The Optimal Algorithm Based on Multiple Markers

The ArUco algorithm can detect multiple markers and estimate the relative poses simultaneously. Assuming N markers observed, there are N numbers of translational and rotation vectors, i.e., { ϕ i , t i } and 0 < i N . An optimal algorithm was designed based on multiple markers to upgrade the position estimated from a single marker. The transform function T f 1 maps the coordinate from the reference of a single marker into the global reference. Because of the layout of the markers, the rotation vectors estimated from different markers are in the same reference. The detailed interpretation is presented in Section 7.
In the optimal algorithm, the cost function was built by employing the Mahalanobis distance. Subsequently, the ϕ ˜ and t ˜ need to be derived to minimise the Mahalanobis distance. There are other methods to achieve that, for example: the well-known RANSAC (Random Sample Consensus), and the weighted average method. However, they may both cause large errors in estimated poses. The marker closer to the camera can provide more accurate pose estimation in its coordinate system. In some cases, the RANSAC and weighted average methods may fail in finding the optimal pose, for instance, when the most similar observed values are estimated by markers at large distances. In the proposed method, once constructing the covariance function properly, the algorithm of minimisation of the Mahalanobis distance is expected to obtain the optimal pose successfully.
The Mahalanobis distance can be expressed as,
l = i = 1 N ( { ϕ i , T f 1 ( I d ) + t i ) } { ϕ ˜ , t ˜ g } ) T Σ i 1 ( { ϕ i , T f 1 ( I d ) + t i ) } { ϕ ˜ , t ˜ g } )
where the ϕ ˜ is the optimal rotation vector, and t ˜ g is the optimal translational vector in the global reference. The Σ i is identical to Σ c m ,
To minimise the Mahalanobis distance, let the differential of l be equal 0 with respect to { ϕ ˜ , t ˜ g } ,
l { ϕ ˜ , t ˜ g } = 0 .
The { ϕ ˜ , t ˜ g } can be derived as,
{ ϕ ˜ , t ˜ g } = ( i = 1 N Σ i ) 1 ( i = 1 N Σ i 1 { ϕ i , t i } ) .
Using the optimal algorithm, the optimal poses of the camera based on the multiple markers are derived. In the following section, the layout of the multiple markers will be introduced.

7. Layout of the Multiple Markers

In the towing tank, the multiple markers are set out on the ground. The layout of the markers is shown in Figure 5, and the reference of the markers is plotted in Figure 6.
Figure 5. Markers in the towing tank.
Figure 6. Reference of the markers.
According to Figure 6, the origin of the global coordinate system is located at the Id.6 marker. After obtaining the relative pose to a specific marker, the transfer function ( T f ) is derived as,
I d = T f ( r , c ) = 6 r + ( 1 ) r ( 6 c )
where I d is the unique ID of the marker. Symbols r and c are the rth row and cth column, respectively.
The inverse function of T f can be used to obtain r and c, I d ( r , c ) with T f 1 . The inverse function T f 1 can be derived by the following two equations,
r = [ I d 6 ] ;
c = ( 1 ) r + 1 ( I d 6 r ) + 6 .
where [ ] indicates the integer division operator. Using these equations, the ( c , r ) is calculated. Because the maximum distance of a single marker is about half a metre, the distance between markers on the grid is 0.5 m, the position of the marker in the global coordinate system is known by multiplying the constant scalar 0.5 by c and r as follows,
x m = 0.5 c ;
y m = 0.5 r .
The camera pose relative to the marker ( u m ), with the position of the marker in the global reference ( x m , y m , 0 ) , can be transferred to u g in the global coordinate system. Because all markers are parallel, the rotation matrix remains the same. Hence, the derivation of the camera pose in the global reference frame is completed.

8. Experimental Results and Discussion

As mentioned before, the experiments were conducted in the towing tank. The vehicle introduced in Section 4 was controlled by an operator remotely. The proposed underwater visual navigation was tested five times: two times with lawnmower patterns, three times with random closed shapes, and one time with a random shape. The navigation system provides the 3D trajectory in absolute scale, as shown in Figure 7.
Figure 7. The 3D trajectory estimated by the method.
Since the vehicle was mainly controlled to move in the horizontal plane, the 3D trajectories were converted into the 2D plots, as shown in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13.
Figure 8. Lawnmower pattern—1st.
Figure 9. Lawnmower pattern—2nd.
Figure 10. Closed trajectory—1st.
Figure 11. Closed trajectory—2nd.
Figure 12. Closed trajectory—3rd.
Figure 13. Random trajectory.
In Figure 8 and Figure 9, the vehicle was operated to travel along an approximated lawnmower path. The predicted trajectories in the figures mostly follow lawnmower patterns. The length of the vehicle is about 0.2 m. In Figure 8, the vehicle travelled along the lawnmower pattern and returned on the same side. In Figure 9, after completing the lawnmower-pattern journey, the vehicle docked on the opposite side. The results show the proposed method presents reasonable estimated trajectories.
In Figure 10, Figure 11 and Figure 12, the vehicle was controlled to complete a closed-loop path. These closed trajectories are generated arbitrarily by the operator. In Figure 12, the estimated trajectory is closed exactly at the end. In Figure 10 and Figure 11, there are small deviations between the start point and the end point in the estimated trajectories. As mentioned before, the length of the vehicle is over 0.2 m. The offsets may be caused by the operation error controlling the ROV or by hydrodynamic disturbances (waves, current). After that, the vehicle was controlled to complete a random trajectory, and the result is shown in Figure 13.
Unlike odometry methods, which suffer from unbounded cumulative error, the navigation method based on the ArUco markers is not dependent on the previous states. Hence, the error of the algorithm is bounded. However, in Figure 13, there are discontinuities in the estimated trajectory. These are caused by mistakes in the marker detection algorithm. The detection algorithm may recognise the marker incorrectly when the marker is far from the camera or is only partially visible. Because of this, the trajectories generated by the marker navigation system are not consistent.
For the quantitative evaluation, the positions of the markers in images can indicate the real position of the camera at that time. Three examples are selected randomly, shown in Table 1, where the markers’ positions and estimated positions are listed. The markers’ positions are known beforehand, and the camera positions are estimated by the proposed method. Through the comparison in Table 1, it is clear that the estimated position is close to the position of the corresponding marker. The deviations between the marker positions and estimated positions are less than 0.5 m, and most of them are about 0.2 m. Since the camera is not exactly on the centre top of the markers, the deviations are acceptable. According to Table 1, the localisation error of the proposed approach is bounded, and the accuracy of the method is guaranteed.
Table 1. Underwater ArUco method quantitative test (unit: metre).

9. Conclusions

In the paper, the detection algorithm of the ArUco markers is introduced first. Then, the testing platform and underwater camera calibration are presented as well. Based on the above, a novel underwater visual navigation method using the multiple ArUco markers is investigated. In the method, the noise model of the estimation by a single marker and an optimal algorithm for multiple markers are derived. The experimental results showed that the proposed method can estimate the positions and altitudes of UUVs precisely. The main contributions of the paper are listed as follows.
  • The multiple ArUco markers are utilised to achieve the accurate localisation of UUVs in certain underwater areas;
  • The probabilistic model of the position estimated by the ArUco is constructed by introducing the Lie algebra.
The advantage of the marker navigation over odometry is that the position error is bounded. In the underwater environments, where the artificial structures are laid out, the markers can be placed in the structures. In this case, the accurate positioning can be implemented in the area by using the proposed method. However, there is some chattering in the estimated trajectories. Some parts of estimated trajectories are not consistent. In future work, a smoothing algorithm is expected to be utilised to solve the issue.

Author Contributions

Conceptualization, Z.X. and A.J.M.; methodology, Z.X.; software, Z.X.; validation, Z.X.; formal analysis, Z.X.; investigation, Z.X.; resources, Z.X., J.N., A.J.M., M.H. and R.N.; data curation, J.N. and R.N.; writing—original draft preparation, Z.X.; writing—review and editing, Z.X.; visualization, Z.X.; supervision, R.N.; project administration, R.N.; funding acquisition, R.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the generous support of John and Vivien Prime in funding aspects of this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Batista, P.; Silvestre, C.; Oliveira, P. A sensor-based controller for homing of underactuated AUVs. IEEE Trans. Robot. 2009, 25, 701–716. [Google Scholar] [CrossRef]
  2. Eustice, R.M. Large-Area Visually Augmented Navigation for Autonomous Underwater Vehicles. Ph.D. Thesis, Massachusetts Institute of Technology and Woods Hole Oceanographic Institution, Falmouth, MA, USA, 2005. [Google Scholar]
  3. Caccia, M. Vision-based ROV horizontal motion control: Near-seafloor experimental results. Control Eng. Pract. 2007, 15, 703–714. [Google Scholar] [CrossRef]
  4. Corke, P.; Lobo, J.; Dias, J. An introduction to inertial and visual sensing. Int. J. Robot. Res. 2007, 26, 519–535. [Google Scholar] [CrossRef]
  5. Xu, Z.; Haroutunian, M.; Murphy, A.J.; Neasham, J.; Norman, R. An Integrated Visual Odometry System for Underwater Vehicles. IEEE J. Ocean. Eng. 2020, 46, 844–863. [Google Scholar] [CrossRef]
  6. Kim, A.; Eustice, R.M. Real-time visual SLAM for autonomous underwater hull inspection using visual saliency. IEEE Trans. Robot. 2013, 29, 719–733. [Google Scholar] [CrossRef] [Green Version]
  7. Leutenegger, S.; Lynen, S.; Bosse, M.; Siegwart, R.; Furgale, P. Keyframe-based visual–inertial odometry using nonlinear optimization. Int. J. Robot. Res. 2015, 34, 314–334. [Google Scholar] [CrossRef] [Green Version]
  8. Garrido-Jurado, S.; Muñoz-Salinas, R.; Madrid-Cuevas, F.J.; Marín-Jiménez, M.J. Automatic generation and detection of highly reliable fiducial markers under occlusion. Pattern Recognit. 2014, 47, 2280–2292. [Google Scholar] [CrossRef]
  9. Babinec, A.; Jurišica, L.; Hubinskỳ, P.; Duchoň, F. Visual localization of mobile robot using artificial markers. Procedia Eng. 2014, 96, 1–9. [Google Scholar] [CrossRef] [Green Version]
  10. Zheng, J.; Bi, S.; Cao, B.; Yang, D. Visual localization of inspection robot using extended kalman filter and aruco markers. In Proceedings of the 2018 IEEE International Conference on Robotics and Biomimetics (ROBIO), Kuala Lumpur, Malaysia, 12–15 December 2018; pp. 742–747. [Google Scholar]
  11. Xing, B.; Zhu, Q.; Pan, F.; Feng, X. Marker-based multi-sensor fusion indoor localization system for micro air vehicles. Sensors 2018, 18, 1706. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Meng, Y.; Lin, K.J.; Peng, B.; Tsai, B.; Shih, C.S. Arpico: Using pictures to build localization service for indoor iot applications. In Proceedings of the 2018 IEEE 11th Conference on Service-Oriented Computing and Applications (SOCA), Paris, France, 20–22 November 2018; pp. 105–112. [Google Scholar]
  13. Ferrera, M.; Moras, J.; Trouvé-Peloux, P.; Creuze, V. Real-Time Monocular Visual Odometry for Turbid and Dynamic Underwater Environments. Sensors 2019, 19, 687. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Rahman, S.; Li, A.Q.; Rekleitis, I. An Underwater SLAM System using Sonar, Visual, Inertial, and Depth Sensor. arXiv 2018, arXiv:1810.03200. [Google Scholar]
  15. Morales, R.; Keitler, P.; Maier, P.; Klinker, G. An underwater augmented reality system for commercial diving operations. In Proceedings of the OCEANS, Bremen, Germany, 11–14 May 2009; pp. 1–8. [Google Scholar]
  16. Jasiobedzki, P.; Se, S.; Bondy, M.; Jakola, R. Underwater 3D mapping and pose estimation for ROV operations. In Proceedings of the OCEANS, Quebec City, QC, Canada, 15–18 September 2008; pp. 1–6. [Google Scholar]
  17. Dos, S.C.; Diego, B.; Gaudig, C.; Fritsche, M.; Dos, R.; Marco, A.; Kirchner, F. An evaluation of artificial fiducial markers in underwater environments. In Proceedings of the OCEANS, Genova, Italy, 18–21 May 2015; pp. 1–6. [Google Scholar]
  18. Čejka, J.; Bruno, F.; Skarlatos, D.; Liarokapis, F. Detecting square markers in underwater environments. Remote Sens. 2019, 11, 459. [Google Scholar] [CrossRef] [Green Version]
  19. Žuži, M.; Čejka, J.; Bruno, F.; Skarlatos, D.; Liarokapis, F. Impact of dehazing on underwater marker detection for augmented reality. Front. Robot. AI 2018, 5, 92. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. Agarwal, A.; Maturana, D.; Scherer, S. Visual Odometry in Smoke Occluded Environments; Technical Report CMU-RI-TR-15-07; Robotics Institute, Carnegie Mellon University: Pittsburgh, PA, USA, 2014. [Google Scholar]
  21. João, M.V.; Da, C.G. Relative Location of Autonomous Underwater Vehicles Using Visual Markers. Master’s Thesis, FEUP, Porto, Portugal, 2020. [Google Scholar]
  22. Ren, R.; Zhang, L.; Yuan, Y.; Wang, Z.; Liu, L. Underwater Visual Tracking Method Based on KCF Algorithm of ArUco Marker. In Proceedings of the Global Oceans 2020: Singapore—US Gulf Coast, Singapore, 5–14 October 2020; pp. 1–6. [Google Scholar]
  23. Martínez-Barberá, H.; Bernal-Polo, P.; Herrero-Pérez, D. Sensor Modeling for Underwater Localization Using a Particle Filter. Sensors 2021, 21, 1549. [Google Scholar] [CrossRef] [PubMed]
  24. Chavez, A.G.; Mueller, C.A.; Doernbach, T.; Birk, A. Underwater navigation using visual markers in the context of intervention missions. Int. J. Adv. Robot. Syst. 2019, 16, 1729881419838967. [Google Scholar] [CrossRef]
  25. Zhang, Z.; Wang, C.; Zhang, Q.; Li, Y.; Feng, X.; Wang, Y. Research on Autonomous Grasping Control of Underwater Manipulator Based on Visual Servo. In Proceedings of the 2019 Chinese Automation Congress (CAC), Hangzhou, China, 22–24 November 2019; pp. 2904–2910. [Google Scholar]
  26. Kato, H.; Billinghurst, M. Marker tracking and hmd calibration for a video-based augmented reality conferencing system. In Proceedings of the 2nd IEEE and ACM International Workshop on Augmented Reality (IWAR’99), San Francisco, CA, USA, 20–21 October 1999; pp. 85–94. [Google Scholar]
  27. Olson, E. AprilTag: A robust and flexible visual fiducial system. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3400–3407. [Google Scholar]
  28. Bowes, P. Hydrodynamics Laboratory. 2010. Available online: https://www.ncl.ac.uk/engineering/about/facilities/marineoffshoresubseatechnology/hydrodynamics/#towingtank (accessed on 10 September 2019).
  29. Mei, C.; Rives, P. Single view point omnidirectional camera calibration from planar grids. In Proceedings of the 2007 IEEE International Conference on Robotics and Automation, Roma, Italy, 10–14 April 2007. [Google Scholar]
  30. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  31. Wang, J.S.; Lee, C.G. Self-adaptive recurrent neuro-fuzzy control of an autonomous underwater vehicle. IEEE Trans. Robot. Autom. 2003, 19, 283–295. [Google Scholar] [CrossRef]
  32. Mebius, J.E. Derivation of the Euler-Rodrigues formula for three-dimensional rotations from the general formula for four-dimensional rotations. arXiv 2007, arXiv:math/0701759. [Google Scholar]
  33. Blanco, J.L. A Tutorial on se (3) Transformation Parameterizations and On-Manifold Optimization; Technical Report; University of Malaga: Malaga, Spain, 2010. [Google Scholar]
  34. Gao, X.; Zhang, T.; Liu, Y.; Yan, Q.Y. 14 Lectures on Visual SLAM: From Theory to Practice; Publishing House of Electronics Industry: Beijing, China, 2017. [Google Scholar]
  35. Williams, C.K.; Rasmussen, C.E. Gaussian Processes for Machine Learning; MIT Press: Cambridge, MA, USA, 2006; Chapter 3. [Google Scholar]
  36. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.