Next Article in Journal
Research on a Cross-Domain Few-Shot Adaptive Classification Algorithm Based on Knowledge Distillation Technology
Next Article in Special Issue
Under-Actuated Motion Control of Haidou-1 ARV Using Data-Driven, Model-Free Adaptive Sliding Mode Control Method
Previous Article in Journal
A Deep Learning Approach for Surface Crack Classification and Segmentation in Unmanned Aerial Vehicle Assisted Infrastructure Inspections
Previous Article in Special Issue
Underwater Rescue Target Detection Based on Acoustic Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RU-SLAM: A Robust Deep-Learning Visual Simultaneous Localization and Mapping (SLAM) System for Weakly Textured Underwater Environments

1
Science and Technology on Underwater Vehicle Laboratory, Harbin Engineering University, Harbin 150001, China
2
Qingdao Innovation and Development Center, Harbin Engineering University, Qingdao 266000, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1937; https://doi.org/10.3390/s24061937
Submission received: 23 February 2024 / Revised: 15 March 2024 / Accepted: 15 March 2024 / Published: 18 March 2024
(This article belongs to the Special Issue Sensors, Modeling and Control for Intelligent Marine Robots)

Abstract

:
Accurate and robust simultaneous localization and mapping (SLAM) systems are crucial for autonomous underwater vehicles (AUVs) to perform missions in unknown environments. However, directly applying deep learning-based SLAM methods to underwater environments poses challenges due to weak textures, image degradation, and the inability to accurately annotate keypoints. In this paper, a robust deep-learning visual SLAM system is proposed. First, a feature generator named UWNet is designed to address weak texture and image degradation problems and extract more accurate keypoint features and their descriptors. Further, the idea of knowledge distillation is introduced based on an improved underwater imaging physical model to train the network in a self-supervised manner. Finally, UWNet is integrated into the ORB-SLAM3 to replace the traditional feature extractor. The extracted local and global features are respectively utilized in the feature tracking and closed-loop detection modules. Experimental results on public datasets and self-collected pool datasets verify that the proposed system maintains high accuracy and robustness in complex scenarios.

1. Introduction

In recent years, significant strides have been achieved in the research on Autonomous Underwater Vehicles (AUVs) with the increasing demand for marine resources. AUVs have emerged as a pivotal tool in the exploration and development of oceanic resources [1]. To undertake diverse missions effectively and safely, precise self-positioning is paramount for AUVs. The current dominant methods for AUV positioning encompass inertial navigation systems (INSs), acoustic positioning systems, geophysical navigation, and Simultaneous Localization and Mapping (SLAM) [2,3,4]. However, the accuracy of INSs gradually diminishes over time. Acoustic positioning systems and geophysical navigation suffer from pre-arranged transponders and pre-existing maps for localization, respectively. In contrast, SLAM enables AUVs to locate objects in unknown environments without prior information, which can reduce cumulative errors through back-end optimization and loop closure detection.
Despite these advantages, achieving robust underwater SLAM remains a formidable task due to the distinctive characteristics of underwater environments, including turbidity, dim lighting, and weak textures. These factors pose challenges for SLAM methods based on hand-crafted features such as SIFT [5], ORB [6], and Shi-Tomasi [7] in terms of ensuring consistent feature extraction and tracking. To address these issues, methods of image enhancement on image frames before feature extraction were proposed by scholars [8,9]. For example, an adversarial contrast learning method in [8] was designed for addressing underwater image degradation in visual SLAM. Considering the issue of water turbidity, ref. [9] proposed a model and model-free hybrid image enhancement approach. While the enhancement feature extraction works to some extent, challenges persist in obtaining accurate keypoint detection and corresponding descriptors due to the weak texture characteristics of underwater scenes.
On the other hand, scholars have tried to incorporate deep features into SLAM [10,11,12]. Methods based on deep learning possess the capability to process full-size images and simultaneously compute pixel-level keypoints and their descriptors. In GCN-SLAM [10], GCNv2 is integrated into ORB-SLAM2 [13] to predict the location keypoints and descriptions in images. HF-NET [14] is incorporated into ORB-SLAM2, resulting in a robust and efficient DXSLAM system [11]. To enhance the system stability, ref. [12] introduced a novel deep learning SLAM system that utilizes self-supervised learning to optimize the network. While deep learning-based SLAM methods often outperform traditional approaches in complex environments, they frequently require substantial labeled datasets for training. However, issues such as degradation of underwater images, such as blue-green bias and blurring, can affect the precise definition and labeling of keypoints.
Therefore, this paper presents a visual SLAM method named RU-SLAM that is subjected to dim and weakly textured underwater environments. It specifically addresses issues of weak texture, image degradation, and the difficulty in accurately labeling keypoints. In this system, an underwater image feature extraction network (UWNet) is integrated to extract features from input images. UWNet incorporates a channel attention [15] and a position attention module [16] into the local keypoint extraction branch. Additionally, it integrates a deformable convolutional module [17] into the local descriptor generation branch to enhance the accuracy of local keypoints and descriptors. To address the challenge of inaccurate keypoint labeling, a pseudo underwater image generator is designed based on a physical model, and UWNet is trained using a self-supervised approach. In summary, the contributions of this work are as follows:
  • A visual SLAM system, based on a deep feature extractor, demonstrates robust functionality in dimly lit and weakly textured underwater environments. This system efficiently extracts both local and global features from underwater images, facilitating tracking and loop closure detection threads, respectively.
  • A pseudo-underwater image generation method, based on a physical model, accurately simulates real underwater scenes. Utilizing this method as an intermediary, UWNet undergoes training via self-supervised learning. This process involves employing a teacher model to extract information from aerial images and guide UWNet’s learning process, enabling it to acquire the necessary knowledge effectively.
  • The experiments conducted on the public EuRoC land dataset [18], the AQUALOC underwater public dataset [19], and our Pool dataset showcase the exceptional localization performance and robustness of RU-SLAM, especially in challenging conditions characterized by weak texture and low lighting in underwater environments.
The structure of this paper is as follows: Section 2 provides a concise overview of the relevant literature. Section 3 presents detailed enhancements made to RU-SLAM. Experimental results and analysis of the proposed method are presented in Section 4. Finally, Section 5 concludes the paper.

2. Related Works

2.1. Works Based on Deep Learning for Feature Extraction

The algorithm for extracting features based on deep learning can be broadly categorized as weakly supervised or self-supervised. Weakly supervised methods primarily focus on optimizing the network using actual image pose information. For instance, CAPS [20] leverages the camera pose to establish antipodal constraints between image frames, optimizing network weights and demonstrating enhanced performance on various geometric tasks. Based on CAPS, a decoupled training strategy is proposed that employs a line-to-window search approach utilizing image pose values to effectively narrow down the search space for enhanced descriptor learning [21]. However, applying such training methods in underwater scenarios poses challenges, particularly when the pose accuracy obtained by sensors falls short of training requirements.
In recent years, self-supervised learning approaches have gained more attention. SuperPoint [22] utilized self-supervised learning to simultaneously train the feature extraction and descriptor generation networks. LF-Net [23] employed attention and asymmetric gradient backpropagation mechanisms for separate self-supervised learning. R2D2 [24] proposed a novel loss function based on micro-averageable accuracy for improving the reliability and repeatability of features, while ASLFeat [25] introduced deformable convolution. For computational efficiency, HF-NET [14] incorporated knowledge distillation, using [22,26] as teacher networks and designing a lightweight MobileNetV2-based [27] student network to extract image features. These advancements highlight that self-supervised learning methods can extract relatively accurate keypoints and descriptors without manual labeling. Therefore, a novel approach is proposed to tackle the issue of inaccurate image keypoint feature labels by employing self-supervised learning for both local and global feature extraction.

2.2. Visual SLAM

Traditional approaches of visual SLAM can be divided into direct and feature-based methods based on distinct data processing techniques. These methods, like DTAM [28], SVO [29], and LSD-SLAM [30], directly employ image brightness information to estimate the current location of the body. The other approach estimates the pose by extracting and matching keypoint features, where Shi-Tomasi [7] and ORB [6] are the more popular feature extraction algorithms. For instance, MonoSLAM [31], ORB-SLAM3 [32], and VINS-Mono [33] have extensively utilized them. The direct method, although efficient, is sensitive to changes in ambient lighting and is more challenging to initialize compared to the feature-based method, which exhibits poor performance on less textured or repetitive images.
Considering the challenges, visual SLAM based on deep learning has gradually emerged as a prominent research focus within the broader SLAM domain. These approaches typically employ deep learning networks to implement certain functionalities within traditional SLAM algorithms. For instance, GCNv2-SLAM [10] uses a graph neural network to generate image keypoint features. DXSLAM [11] leveraged a deep CNN, enhancing the stability of the SLAM system in intricate environments. RWT-SLAM [34] replaced the conventional SLAM module with a CNN-based LoFTR [35] network in ORB-SLAM2 to generate the description sub-module. The proposed study presents a cross-modal knowledge distillation framework, seamlessly integrated into the ORB-SLAM3 architecture [36]. In summary, deep learning-based methods outperform traditional visual SLAM in complex environments. However, these approaches have not effectively addressed challenges in underwater scenarios, such as weak texture, image degradation, and inaccurate keypoint labeling. Consequently, this paper introduces an underwater visual SLAM method based on deep learning, RU-SLAM.

3. RU-SLAM Method

3.1. Overview

The overall framework of the RU-SLAM is illustrated in Figure 1. This system is constructed upon ORB-SLAM3 with the underwater image frame (Real Frame) serving as input. It can be broadly categorized into three main parts based on thread tasks: tracking, local mapping, and closed-loop detection.
In the tracking thread, UWNet acts as an extractor to obtain local and global features of Real Frame. The channel attention (CA) and spatial attention modules (SA) are added at the forefront of the local feature point extraction branch (FPE) to process the input features in parallel before fusing the features. Furthermore, we introduce two deformable convolution (DCN) modules at the descriptor generation branch (DG). In the training phase, both the in-air RGB image (air image) and the pseudo-underwater image are utilized as inputs for the teacher models and the student mode, respectively. The information extracted from these teacher models is utilized to train the student model.
Local keypoint features and descriptors play a crucial role in system initialization, camera pose estimation, and determining keyframes (KFs) and their insertion into the local mapping thread. Upon receiving new KFs, the local mapping thread updates KFs and their connection relationships within the local map. Simultaneously, Bundle Adjustment (BA) is performed to optimize the position and map points of KFs in the current local map. Redundant KFs and their corresponding map points are then eliminated. The possibility of loop closure is assessed using a similarity score, which is computed based on the global descriptors of both the current KF and the KFs stored in the database. Geometric verification and temporal geometry verification are subsequently employed to determine whether correction for loop closure is necessary in the present region. Specific details of the improvement module are elaborated below.

3.2. Local and Global Feature Extraction Module

The approach for local and global feature extraction, based on self-supervised learning, is depicted in Figure 2. Initially, the proposed pseudo-underwater image generator is employed to acquire simulated underwater images (pseudo-underwater images). Subsequently, both local teacher models, SuperPoint, and global ones, NetVLAD, are utilized for feature extraction from the air image to guide UWNet in extracting corresponding features from the pseudo-underwater image. This process effectively mitigates the challenges associated with underwater image degradation. In the subsequent sections, the pseudo-underwater image generator and the UWNet components are described in more detail.

3.2.1. Pseudo Underwater Image Generator

1
Improved underwater imaging model
A widely employed classical physical model for underwater imaging is the Jaffe–McGlamery model [37]. Reputable scholars have observed that the limited visible range in underwater optical imaging often results in the general exclusion of the foreground scattering component.
The expression is presented in Equation (1).
M λ ( x ) = N λ ( x ) e β ( λ ) d ( x ) + A λ ( 1 e β ( λ ) d ( x ) ) , λ { r , g , b }
t λ ( x ) = e β ( λ ) d ( x )
where x represents the pixel coordinate value, λ stands for the color channel, M λ ( x ) denotes the synthesized underwater image, N λ ( x ) represents the RGB image in air, A λ ( x ) signifies the global background light, t λ ( x ) denotes optical medium transmittance, β ( λ ) is the attenuation coefficient of the water body, and d ( x ) is the distance from the target scene to the imaging device scene depth. The components of N λ ( x ) e β ( λ ) d ( x ) and A λ ( 1 e β ( λ ) d ( x ) ) are respectively responsible for direct attenuation and background light scattering. Additionally, attenuated atmospheric light and the auxiliary light sources are considered in underwater scenes. Equation (2) can be refined to Equation (3) following the approach suggested by [38].
M λ ( x ) = A λ R λ x t λ ( x ) + A λ ( 1 t λ ( x ) ) = L ( x ) η λ ( x ) R λ ( x ) t λ ( x ) + L ( x ) η λ ( x ) ( 1 t λ ( x ) )
In this Equation (3), N λ = A λ R λ ( x ) , A λ = L λ ( x ) η λ ( x ) , R λ ( x ) represents the reflected light, L ( x ) is the illumination intensity, and η λ ( x ) is the ambient light color.
The proposed model considers various factors to create a more realistic underwater environment. To simulate particulate impurities in the underwater images, Gaussian noise W is added to the model. Furthermore, image blurring may occur due to the movement of the AUV. Through the convolution operation, both the point spread function (PSF) and the direct transmission attenuation component are used to replicate image blurring. Overall, the modified model as Equation (4).
M λ ( x ) = { L ( x ) η λ ( x ) R λ ( x ) e β ( λ ) d ( x ) } P S F + L ( x ) η λ ( x ) ( 1 e β ( λ ) d ( x ) ) + W
2
Estimation of A λ and t λ
In general, the ambient illumination varies gently throughout the space except for the occluded region. Therefore, this paper assumes that A λ is constant in any localized region donate Ω i . The value V of Equation (3) in the Ω i can be obtained by Equation (5).
V = max x Ω i M λ ( x ) = max j Ω i ( L Ω i ( x ) η Ω i λ ( x ) R λ ( x ) t Ω i λ ( x ) + L Ω i ( x ) η Ω i λ ( x ) ( 1 t Ω i λ ( x ) ) )
The transmittance t Ω i λ ( x ) is locally invariant from [39]. According to the priori theory of maximum reflection proposed in [38], it can be obtained that max x Ω i R λ ( x ) 1 . Thus Equation (6) can be deduced.
η Ω i λ ( x ) = V L Ω i ( x )
based on the above assumption of in-variance in the local region, we can obtain Equation (7).
L Ω i ( x ) = max x { R , G , B } ( max x Ω i V )
The value of t λ is related to β ( λ ) and d ( x ) according to Equation (2). Since the dataset used for synthesis includes depth data, there is no need for the estimation of d ( x ) . We employed ten water types proposed by Jerlov [40] to estimate β ( λ ) .

3.2.2. UWNet Method

1
Network architecture
The degradation of underwater images causes extracting local and global features to be a complex task. We devised the UWNet network based on the inspiration from HF-Net, as shown in Figure 2. It utilizes MobileNetV2 as a backbone network to extract high-level features from the input image. The first seven layers of this backbone network serve as shared features, producing a feature map F1 at 1/8 resolution in the seventh layer and a feature map F2 at 1/32 resolution in the eighteenth layer. Subsequently, F1 and F2 are respectively used as inputs for both the local and global feature generators.
In the FPE branch, the feature map F1 is first inputted into both the SA and the CA to capture dependencies across spatial and channel dimensions. Their outputs from SA and CA are then fused to enhance the feature representation. Subsequently, the fused features pass through two consecutive convolutional blocks. The first block involves a convolution operation, followed by batch normalization (BatchNorm) and ReLU activation. This is followed by another convolutional operation with a 1 × 1 kernel size, resulting in 65 channels. The resulting features are reshaped using a SoftMax operation to generate the keypoint score map.
Concurrently, the feature map F1 is processed in the DG branch, where two deformable convolutional layers are utilized to capture crucial feature information and improve learning of scale kernel and rotation transformations. Following this, two convolutional operations are performed. The first convolutional block includes a 3 × 3 convolutional operation, BatchNorm, and ReLU activation. In contrast, the second convolutional block consists solely of a 1 × 1 convolutional operation with 256 channels. Finally, the output features undergo bilinear interpolation and L2-Norm operation to generate the local feature description subgraph.
The global feature generator is utilized to acquire the image’s global descriptor. F2 is fed into the NetVLAD layer and then undergoes a dimensionality reduction operation to obtain a global descriptor of size (4096). NetVLAD is primarily used for re-recognition or location tasks.
2
Train details
We employ the concept of knowledge distillation to optimize UWNet network parameters in a self-supervised multitasking manner, utilizing the RMSProp optimizer with an initial learning rate set to 0.003, employing a batch size of 16, and conducting training for 150 epochs. Traditional multitask learning typically involves jointly optimizing all losses and manually adjusting their weights. However, this approach often struggles to produce models that excel simultaneously across multiple tasks. Following the training strategy outlined in [14], the automatic selection of weights for different losses in multi-task learning and self-adjustment during training have been proven effective in enhancing the accuracy of individual tasks. Hence, we utilized Equation (8) as the loss function.
L = e α L 1 + e β L 2 + 2 e γ L 3 + α + β + γ
where L 1 = D s g D t 1 g 2 2 , L 2 = D s l D t 2 l 2 2 , and L 3 = C E ( P s , P t 2 ) . The values of α , β and γ are obtained by learning through network training. Furthermore, t1, t2, and s are respectively the NetVLAD, SuperPoint, and UWNet network. The global and local descriptors are D g and D l , respectively.

3.3. Feature Tracking

The tracking thread utilizes local keypoint features and descriptors extracted by UWNet in each input image frame to estimate camera motion. To be specific, the system was initially initialized. Subsequently, the keypoint features from the previous frame were projected onto the current frame at a constant speed, and matching was performed within a local window. In case of failure, an attempt was made to match between the current frame and reference KF. The standard k-nearest neighbor search and subsequent ratio test were implemented in our system for these two matching strategies. Afterwards, tracking of a reconstructed local map took place along with pose optimization. Finally, it was determined whether the current frame satisfies the criteria for being designated as a KF.

3.4. Loop Closing Module

Loop closure is crucial for eliminating cumulative errors in SLAM and establishing a globally consistent map. The current methodologies predominantly rely on the bag of words approach. However, this method neglects spatial relationships between features, potentially leading to false closures. Therefore, a closure detection method based on global descriptors has been employed in our system, which calculates the similarity between images based on global feature descriptors to determine whether the AUV revisits a location. By retrieving KFs with higher scores from the global descriptor database as candidate loops, the system mitigates the risk of false closures. Similarity judgment is performed using Euclidean distance calculation, as presented in Equation (9).
s c o r e ( x , y ) = 1 i N v i ( x ) v i ( y ) 2
where v ( x ) and v ( y ) are the global feature vectors of the query image and the image in the database, respectively, i represents the elements of the vector, and the size of N is 4096.

4. Experimental Analysis

The authenticity of the simulated underwater images is first validated, followed by an assessment of UWNet’s accuracy as a local feature generator compared to other algorithms. Finally, diverse datasets are used to evaluate the complete RU-SLAM system. The experiments are conducted on a computer equipped with an Intel Core i9-9900K CPU (3.60 GHz × 16 cores), 16 GB of RAM, and an NVIDIA RTX 2080Ti GPU.

4.1. Experiments with Pseudo-Underwater Image Generator

A dataset comprising N images with varying degrees of degradation and low texture was selected from the EUVP dataset [41] to estimate the parameter value A λ , as illustrated in Figure 3. Additionally, 200 images were randomly chosen as a test dataset from the NYU depth V2 dataset [42], which encompasses a total of 1449 RGB-D images. The specific details are as follows.
P regions of size 8 × 8 are randomly selected for each real underwater image. Equations (6) and (7) are then employed to calculate A λ within each region, resulting in A λ of N P denoted as A . In this paper, P and N are set to 8 and 10, respectively. For each image to be synthesized, one region is randomly selected from A as A λ , and t λ is obtained using Equation (2). The blur angle and length of the PSF function is 20 degrees and 10 pixels, while W is random Gaussian noise. The partially synthesized image data depicted in Figure 4 shows that the pseudo-underwater image exhibits varying background illumination, along with randomly added noise and different degrees of blurring. The clarity and contrast of the image progressively deteriorate under different attenuation coefficients, resembling the visual effect of real underwater images.
We employ a reference-free metric to evaluate the authenticity of pseudo-underwater images; it comprises the Underwater Color Image Quality Evaluation (UCIQE), Underwater Image Quality Measurement (UIQM), and its two attributes: Underwater Image Colorimetric Measurement (UICM) and Underwater Image Sharpness Measurement (UISM). The experimental results, as presented in Table 1, indicate that our algorithm can synthesize a more realistic underwater image compared to algorithms such as UWCNN [43], WaterGAN [44], and UGAN-P [45] for underwater image synthesis.

4.2. Evaluation of Local Keypoint and Descriptor in UWNet

The performance of the proposed local keypoint and descriptor in UWNet is assessed based on keypoint repeatability (%Rep.), matching score (%M.S.), and average matching accuracy (%MMA). Two datasets are utilized for evaluation: the Hpatches dataset [46] and a self-collected Underwater dataset (Upatches). The Upatches dataset comprises 30 original images obtained from the UIEB underwater public dataset [47], along with their corresponding homography-transformed images. These images exhibit characteristics of low-light conditions, turbidity, and low-texture underwater scenarios. It is noteworthy that the rotation angle does not exceed 180°.
The experiments were conducted using a dataset resolution of 480 × 640 and a limited number of feature points (2 k). According to the results presented in Table 2 on the Hpatches, our algorithm demonstrates competitive performance by outperforming other algorithms in terms of %M.S. and %MMA. For dataset Upatches, the proposed algorithm consistently outperforms other algorithms across all metrics. The experimental results based on these datasets reveals that our algorithm evinces commendable performance in both terrestrial and underwater datasets. Notably, the comparison of our algorithm’s performance with and without incorporating attention modules and deformable convolutions (Our w/o AM DCN) demonstrates a significant improvement in matching accuracy, thereby highlighting the efficacy of the proposed approach.
The visual results depicted in Figure 5 showcase the keypoint matching outcomes in selected Upatches data, where the yellow line represents correct correspondences. Even after RANSAC processing, our method consistently yields denser and more accurate matches compared with SIFT, LF-NET, and HF-NET, and this is especially evident in underwater images characterized by weak textures, low contrast, and blurring.

4.3. SLAM System Evaluation

The performance of RU-SLAM in both underwater and land environments is assessed using publicly available datasets, namely EuRoC and AQUALOC, as well as our own collected dataset called Pool. The Root Mean Square Error (RMSE) of the Absolute Pose Error (APE) and Relative Pose Error (RPE) between the estimated trajectory and the provided ground truth trajectory is computed using the evo tool as performance evaluation metrics, providing assessments of both global and local accuracy of the system. It should be noted that during the evaluation analysis, alignment is performed between the experimentally estimated trajectory and the provided ground truth. Specifically, in stereo vision scenarios, SE (3) transformation is employed for alignment, while in pure monocular vision, Sim (3) transformation is used to correct for scale due to a lack of depth information. Additionally, a relative interval of 0.1 m is set for RPE to enable more precise detection of positional changes and a thorough analysis of system performance. If the estimated trajectory is too short or diverges, the results will be marked as a failure (×). Furthermore, the average (Avg) in the table represents the mean of the sequences wherein all algorithms employed in the table achieve success.

4.3.1. EuRoC Dataset

The EuRoC dataset comprises two sections, namely Machine Hall and Vicon Room, which include stereo image and IMU data. The first dataset consists of five sequences: MH01, MH02, MH03, MH04, and MH05. These sequences are categorized into simple, medium, and difficult levels based on lighting conditions and scene textures. The Machine Hall dataset is employed for conducting experiments on stereo visual SLAM. The robustness of RU-SLAM is evaluated through a comparative analysis with ORB-SLAM3 and DXSLAM, based on results as presented in Table 3.
Where ORB-SLAM3 is the latest visual SLAM based on the ORB, and DXSLAM integrates the feature extractor based on the CNN deep convolutional network into the ORB-SLAM2 framework, the RU-SLAM exhibits superior performance compared to other methods in four sequences of the dataset, as evidenced by the observations in Table 3. Compared to ORB-SLAM3, our algorithm achieves a reduction in error of 44.45% and 49.15% on the MH01 and MH04 sequences, respectively. This success is primarily attributed to the simplicity of MH01, featuring a well-lit and textured scene conducive to obtaining more uniform feature points through the deep learning-based method. Conversely, MH04, being a challenging sequence with a large viewing angle and low light, exposes the limitations of traditional SLAM based on visual feature points in such conditions. The trajectory errors in sequences MH02, MH03, and MU04 are further reduced by our method compared to the ORB-SLAM3. However, the extent of error reduction in MH02 is not as significant as that observed in DXSLAM. The above analysis reveals that our algorithm displays admirable performance in land scenarios characterized by wide viewing angles and low lighting conditions.

4.3.2. AQUALOC Dataset

The AQUALOC dataset provides monocular, IMU, and depth data for underwater environments. It is categorized into two sections: harbor and archaeological, based on distinct acquisition environments. Additionally, it includes a ground truth trajectory that exhibits relatively high accuracy. To evaluate the performance of RU-SLAM in underwater scenarios, separate experiments were conducted using these two datasets of monocular data.
1
Archaeological
The archaeological dataset consists of 10 sequences collected on the seafloor at depths of several hundred meters. However, the propellers of the remotely operated vehicle (ROV) caused sediment disturbance on the seafloor, thereby introducing disruptions to visual feature recognition.
Detailed comparison results for the RMSE of ORB-SLAM3, DXSLAM, and RU-SLAM under APE and RPE are provided in Table 4. Additionally, Figure 6 presents trajectory comparisons across different algorithms for selected sequences. The superiority of the proposed method over others is evident from Table 4. Owing to interference from seafloor sediment and a lack of sufficient seafloor texture features, ORB-SLAM3 exhibits poor performance in this dataset. The average APE and RPE of the RU-SLAM algorithm decreased by 69.95% and 55.50%, respectively, in contrast to DXSLAM, thereby demonstrating the enhanced accuracy and robustness of our method.
2
Harbor
The harbor dataset comprises a total of seven sequences capturing scenarios of coastal ports at depths of 3–4 m. In this experiment, the study primarily focuses on the first six sequences, and present comparative test results between the ORB-SLAM3 and HFNET-SLAM algorithms and the proposed RU-SLAM algorithm. HFNET-SLAM refers to an algorithm where the unimproved HF-NET is used instead of the feature extraction module in ORB-SLAM3.
According to Table 5, our algorithm exhibits superior performance in underwater environments compared to the other two methods. However, the APE differences among the three algorithms in sequence 03 are minimal. This can be attributed to the clear water quality and abundant texture of this sequence, which facilitate both deep learning and traditional methods, enabling the attainment of satisfactory experimental results. In Sequence 04, both ORB-SLAM3 and HFNET-SLAM encounter localization failures due to the presence of visually poor features, resulting in incomplete positioning trajectories. In contrast, RU-SLAM successfully extracts features and accurately estimates camera poses in this challenging visual environment. Additionally, experimental results demonstrate that RU-SLAM exhibits enhanced robustness and stability compared to the original HFNET-SLAM.

4.3.3. Pool Dataset

In addition to utilizing the publicly available data set, we conducted our own data collection to assess the performance of RU-SLAM. Three sequences are captured by an ROV, as shown in Figure 7a, equipped with a monocular camera in a 30 × 50 × 10 m3 pool characterized by low illumination and weak textures. It is worth noting that the velocity of the ROV ranges between 0.3 and 0.5 m/s. There are artificial rock formations of various sizes strategically positioned at the bottom to function as obstacles. The ROV was remotely controlled via a console to navigate around obstacles and its movements were recorded during the experiment. The actual setup is illustrated in Figure 7b. The sequences consisted of the following: 01 circumnavigating the four sides of the pool with a maximum turn angle of 90 degrees; 02 maneuvering around obstacles within the pool in a complete circle; and 03 circumnavigating the large obstacle at the center of the tank bottom and forming a closed loop.
The estimated trajectory results, obtained by comparing our system with ORB-SLAM3, are depicted in Figure 8. It is noteworthy that this component serves as a valuable complement due to the absence of ground truth measurements. As depicted in Figure 8a, when the texture at the bottom and walls of the pool is weak and repetitive, ORB-SLAM3 provides an incorrect rotation prediction, resulting in the trajectory deviating from the actual walking path. In Sequence 02, both ORB-SLAM3 and our algorithm exhibit good localization performance. This is mainly because Sequence b was recorded in water five to six meters from the pool bottom and moved around artificial hills under optimal lighting conditions. As shown in Figure 8c, RU-SLAM demonstrates robust performance in handling images with weak texture.

5. Conclusions

A robust visual SLAM system based on depth features is proposed for accurate positioning in challenging underwater environments with low illumination and weak texture. We introduce the UWNet network to generate dense local and global features. Thanks to the diverse receptive field facilitated by the attention mechanism and deformable convolution, high quality descriptors for scenes with limited texture and image degradation can be obtained. A pseudo-underwater image generator is presented, and self-supervised training is employed for UWNet to make the extract features more adaptable to underwater scenes. Moreover, we seamlessly integrate UWNet into ORB-SLAM3, and our experimental results on diverse public datasets as well as our data demonstrate the exceptional accuracy and robustness of our system.

Author Contributions

Conceptualization, Z.W.; methodology, Q.C.; software, X.M.; validation, Z.W., Q.C. and X.M.; formal analysis, Z.W.; resources, Q.C.; writing—original draft preparation, Q.C.; writing—review and editing, Z.W.; supervision, Z.W.; funding acquisition, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by The National Key Research and Development Program of China, grant number 2023YFB4707003 and The National Natural Science Foundation of China, grant numbers 52131101 and 52025111.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the data that support the findings in this study are given by contacting the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, S.; Zhao, S.; An, D.; Liu, J.; Wang, H.; Feng, Y.; Li, D.; Zhao, R. Visual SLAM for Underwater Vehicles: A Survey. Comput. Sci. Rev. 2022, 46, 100510. [Google Scholar] [CrossRef]
  2. Sahoo, A.; Dwivedy, S.K.; Robi, P.S. Advancements in the Field of Autonomous Underwater Vehicle. Ocean Eng. 2019, 181, 145–160. [Google Scholar] [CrossRef]
  3. Zhang, B.; Ji, D.; Liu, S.; Zhu, X.; Xu, W. Autonomous Underwater Vehicle Navigation: A Review. Ocean Eng. 2023, 273, 113861. [Google Scholar] [CrossRef]
  4. Wu, Y.; Ta, X.; Xiao, R.; Wei, Y.; An, D.; Li, D. Survey of Underwater Robot Positioning Navigation. Appl. Ocean Res. 2019, 90, 101845. [Google Scholar] [CrossRef]
  5. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  6. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An Efficient Alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2564–2571. [Google Scholar]
  7. Shi, J.; Tomasi, C. Good Features to Track. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition CVPR-94, Seattle, WA, USA, 21–23 June 1994; IEEE Computer Society Press: Piscataway, NJ, USA, 1994; pp. 593–600. [Google Scholar]
  8. Zheng, Z.; Xin, Z.; Yu, Z.; Yeung, S.-K. Real-Time GAN-Based Image Enhancement for Robust Underwater Monocular SLAM. Front. Mar. Sci. 2023, 10, 1161399. [Google Scholar] [CrossRef]
  9. Cho, Y.; Kim, A. Channel Invariant Online Visibility Enhancement for Visual SLAM in a Turbid Environment. J. Field Robot. 2018, 35, 1080–1100. [Google Scholar] [CrossRef]
  10. Tang, J.; Ericson, L.; Folkesson, J.; Jensfelt, P. GCNv2: Efficient Correspondence Prediction for Real-Time SLAM 2019. IEEE Robot. Autom. Lett. 2019, 4, 3505–3512. [Google Scholar]
  11. Li, D.; Shi, X.; Long, Q.; Liu, S.; Yang, W.; Wang, F.; Wei, Q.; Qiao, F. DXSLAM: A Robust and Efficient Visual SLAM System with Deep Features. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 4958–4965. [Google Scholar]
  12. Li, Z.; Cao, J.; Hao, Q.; Zhao, X.; Ning, Y.; Li, D. DAN-SuperPoint: Self-Supervised Feature Point Detection Algorithm with Dual Attention Network. Sensors 2022, 22, 1940. [Google Scholar] [CrossRef] [PubMed]
  13. Mur-Artal, R.; Tardos, J.D. ORB-SLAM2: An Open-Source SLAM System for Monocular, Stereo and RGB-D Cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  14. Sarlin, P.-E.; Cadena, C.; Siegwart, R.; Dymczyk, M. From Coarse to Fine: Robust Hierarchical Localization at Large Scale. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12716–12725. [Google Scholar]
  15. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  17. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  18. Burri, M.; Nikolic, J.; Gohl, P.; Schneider, T.; Rehder, J.; Omari, S.; Achtelik, M.W.; Siegwart, R. The EuRoC Micro Aerial Vehicle Datasets. Int. J. Robot. Res. 2016, 35, 1157–1163. [Google Scholar] [CrossRef]
  19. Ferrera, M.; Creuze, V.; Moras, J.; Trouvé-Peloux, P. AQUALOC: An Underwater Dataset for Visual-Inertial-Pressure Localization. Int. J. Robot. Res. 2019, 38, 1549–1559. [Google Scholar] [CrossRef]
  20. Wang, Q.; Zhou, X.; Hariharan, B.; Snavely, N. Learning Feature Descriptors Using Camera Pose Supervision. In Computer Vision—ECCV 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 12346, pp. 757–774. ISBN 978-3-030-58451-1. [Google Scholar]
  21. Li, K.; Wang, L.; Liu, L.; Ran, Q.; Xu, K.; Guo, Y. Decoupling Makes Weakly Supervised Local Feature Better. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2022), New Orleans, LA, USA, 18–24 June 2022; IEEE Computer Society: Los Alamitos, CA, USA, 2022; pp. 15817–15827. [Google Scholar]
  22. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-Supervised Interest Point Detection and Description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar]
  23. Ono, Y.; Trulls, E.; Fua, P.; Yi, K.M. LF-Net: Learning Local Features from Images. In Proceedings of the Advances in Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  24. Revaud, J.; Weinzaepfel, P.; De Souza, C.; Pion, N.; Csurka, G.; Cabon, Y.; Humenberger, M. R2D2: Repeatable and Reliable Detector and Descriptor. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS), Vancouver, BC, USA, 8–14 December 2019. [Google Scholar]
  25. Luo, Z.; Zhou, L.; Bai, X.; Chen, H.; Zhang, J.; Yao, Y.; Li, S.; Fang, T.; Quan, L. ASLFeat: Learning Local Features of Accurate Shape and Localization. In Proceedings of the 7th International Conference on Signal Processing, Noida, India, 27–28 February 2020; pp. 1389–1392. [Google Scholar]
  26. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN Architecture for Weakly Supervised Place Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 5297–5307. [Google Scholar]
  27. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  28. Newcombe, R.A.; Lovegrove, S.J.; Davison, A.J. DTAM: Dense Tracking and Mapping in Real-Time. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 2320–2327. [Google Scholar]
  29. Forster, C.; Pizzoli, M.; Scaramuzza, D. SVO: Fast Semi-Direct Monocular Visual Odometry. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 15–22. [Google Scholar]
  30. Engel, J.; Schöps, T.; Cremers, D. LSD-SLAM Large-Scale Direct Monocular SLAM. In Computer Vision—ECCV 2014; Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2014; Volume 8690, pp. 834–849. ISBN 978-3-319-10604-5. [Google Scholar]
  31. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-Time Single Camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef]
  32. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.M.; Tardós, J.D. ORB-SLAM3: An Accurate Open-Source Library for Visual, Visual-Inertial and Multi-Map SLAM. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  33. Qin, T.; Li, P.; Shen, S. VINS-Mono: A Robust and Versatile Monocular Visual-Inertial State Estimator. IEEE Trans. Robot. 2018, 34, 1004–1020. [Google Scholar] [CrossRef]
  34. Peng, Q.; Xiang, Z.; Fan, Y.; Zhao, T.; Zhao, X. RWT-SLAM: Robust Visual SLAM for Highly Weak-Textured Environments. arXiv 2022, arXiv:2207.03539. [Google Scholar]
  35. Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-Free Local Feature Matching with Transformers. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2021, Nashville, TN, USA, 20–25 June 2021; IEEE Computer Society: Los Alamitos, CA, USA, 2021; pp. 8918–8927. [Google Scholar]
  36. Yang, J.; Gong, M.; Nair, G.; Lee, J.H.; Monty, J.; Pu, Y. Knowledge Distillation for Feature Extraction in Underwater VSLAM. In Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5163–5169. [Google Scholar]
  37. Narasimhan, S.G.; Nayar, S.K. Vision and the Atmosphere. In Proceedings of the ACM SIGGRAPH ASIA 2008 Courses on—SIGGRAPH Asia ’08, New York, NY, USA, 11–15 August 2008; ACM Press: Singapore, 2008; pp. 1–22. [Google Scholar]
  38. Zhang, J.; Cao, Y.; Fang, S.; Kang, Y.; Chen, C.W. Fast Haze Removal for Nighttime Image Using Maximum Reflectance Prior. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 7016–7024. [Google Scholar]
  39. Liang, Z.; Ding, X.; Wang, Y.; Yan, X.; Fu, X. GUDCP: Generalization of Underwater Dark Channel Prior for Underwater Image Restoration. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 4879–4884. [Google Scholar] [CrossRef]
  40. Liu, X.; Chen, B.M. A Systematic Approach to Synthesize Underwater Images Benchmark Dataset and Beyond. In Proceedings of the 2019 IEEE 15th International Conference on Control and Automation (ICCA), Edinburgh, UK, 16–19 July 2019; pp. 1517–1522. [Google Scholar]
  41. Islam, M.J.; Xia, Y.; Sattar, J. Fast Underwater Image Enhancement for Improved Visual Perception. IEEE Robot. Autom. Lett. 2020, 5, 3227–3234. [Google Scholar] [CrossRef]
  42. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor Segmentation and Support Inference from RGBD Images. In Computer Vision—ECCV 2012; Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7576, pp. 746–760. ISBN 978-3-642-33714-7. [Google Scholar]
  43. Li, C.; Anwar, S.; Porikli, F. Underwater Scene Prior Inspired Deep Underwater Image and Video Enhancement. Pattern Recognit. 2020, 98, 107038. [Google Scholar] [CrossRef]
  44. Li, J.; Skinner, K.A.; Eustice, R.M.; Johnson-Roberson, M. WaterGAN: Unsupervised Generative Network to Enable Real-Time Color Correction of Monocular Underwater Images. IEEE Robot. Autom. Lett. 2017, 3, 387–394. [Google Scholar] [CrossRef]
  45. Fabbri, C.; Islam, M.J.; Sattar, J. Enhancing Underwater Imagery Using Generative Adversarial Networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 7159–7165. [Google Scholar]
  46. Balntas, V.; Lenc, K.; Vedaldi, A.; Mikolajczyk, K. HPatches: A Benchmark and Evaluation of Handcrafted and Learned Local Descriptors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5173–5182. [Google Scholar]
  47. Li, C.; Guo, C.; Ren, W.; Cong, R.; Hou, J.; Kwong, S.; Tao, D. An Underwater Image Enhancement Benchmark Dataset and Beyond. IEEE Trans. Image Process. 2020, 29, 4376–4389. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The overall framework of the RU-SLAM system.
Figure 1. The overall framework of the RU-SLAM system.
Sensors 24 01937 g001
Figure 2. Framework of local and global feature extraction methods for images.
Figure 2. Framework of local and global feature extraction methods for images.
Sensors 24 01937 g002
Figure 3. Underwater real image. (a) blue-green (b) turbidity (c) bluish (d) greenish.
Figure 3. Underwater real image. (a) blue-green (b) turbidity (c) bluish (d) greenish.
Sensors 24 01937 g003
Figure 4. Pseudo-underwater images generated with different attenuation coefficients.
Figure 4. Pseudo-underwater images generated with different attenuation coefficients.
Sensors 24 01937 g004
Figure 5. Results of image keypoint matching using different methods on the Upatches dataset: (a) SIFT; (b) LF-NET; (c) HF-NET; (d) Ours.
Figure 5. Results of image keypoint matching using different methods on the Upatches dataset: (a) SIFT; (b) LF-NET; (c) HF-NET; (d) Ours.
Sensors 24 01937 g005
Figure 6. Comparison of partial sequence trajectory results from different algorithms in Archaeological datasets. (a) Sequences 01; (b) Sequences 03; (c) Sequences 05; (d) Sequences 07.
Figure 6. Comparison of partial sequence trajectory results from different algorithms in Archaeological datasets. (a) Sequences 01; (b) Sequences 03; (c) Sequences 05; (d) Sequences 07.
Sensors 24 01937 g006
Figure 7. The ROV and Pool environment, including (a) Physical drawing of ROV; (b) Pool Rockery Placement.
Figure 7. The ROV and Pool environment, including (a) Physical drawing of ROV; (b) Pool Rockery Placement.
Sensors 24 01937 g007
Figure 8. Estimated trajectories resulting from RU-SLAM and ORB-SLAM3 in our own Pool dataset. (a) Sequence 01; (b) Sequence 02; (c) Sequence 03.
Figure 8. Estimated trajectories resulting from RU-SLAM and ORB-SLAM3 in our own Pool dataset. (a) Sequence 01; (b) Sequence 02; (c) Sequence 03.
Sensors 24 01937 g008
Table 1. Evaluation of reference-free underwater image quality metrics on the same test set using different image synthesis methods. The best result is shown in the bold numbers.
Table 1. Evaluation of reference-free underwater image quality metrics on the same test set using different image synthesis methods. The best result is shown in the bold numbers.
MethodUCIQEUIQMUICMUISM
WaterGAN0.4300.51410.3511.153
UGAN-P0.3810.4138.4901.004
UWCNN0.3280.2348.1830.453
Ours0.2710.1904.8090.652
Table 2. Quantitative evaluation results on Hpatches and Upatches datasets for image matching. The best result is shown in the bold numbers.
Table 2. Quantitative evaluation results on Hpatches and Upatches datasets for image matching. The best result is shown in the bold numbers.
MethodHpathesUpatches
%Rep.%M.S.%MMA%Rep.%M.S.%MMA
SIFT35.8028.1065.4035.9929.9070.86
LF-NET42.4432.2762.6028.6817.6930.00
SuperPoint52.3545.8872.1142.0046.9370.00
ASLFeat62.2445.1072.2852.6534.7658.64
HF-NET54.8045.3272.5055.9146.4963.33
Our w/o AM DCN48.2038.6065.9052.7141.5353.33
Ours56.0047.0074.4057.0047.2873.33
Table 3. The RMS of APE (m) obtained from various SLAM methods applied in the Machine Hall. The best result is shown in the bold numbers.
Table 3. The RMS of APE (m) obtained from various SLAM methods applied in the Machine Hall. The best result is shown in the bold numbers.
SequenceORB-SLAM3DXSLAMOurs
MH010.0270.0220.015
MH020.0370.0150.016
MH030.0280.0250.022
MH040.1180.1210.060
MH050.0720.0640.043
Avg0.0560.0440.031
Table 4. The results are obtained from various SLAM methods applied in the archaeological dataset. The best result is shown in the bold numbers.
Table 4. The results are obtained from various SLAM methods applied in the archaeological dataset. The best result is shown in the bold numbers.
SequenceORB-SLAM3DXSLAMOurs
APERPEAPERPEAPERPE
01× 1×0.1080.0170.0650.009
02××××××
03××0.1540.0950.1240.055
041.4500.5072.9310.4920.7630.315
050.2630.1300.1350.0390.1060.047
06××××0.1830.020
071.4680.2190.8910.1970.1320.086
080.1010.0490.1530.6600.1140.048
09××0.0360.2270.0270.282
100.3430.0310.5500.0180.1610.012
Avg0.7250.1870.9320.2810.2550.101
1 Represents trace failure.
Table 5. The results obtained from various SLAM methods applied in the harbor. The best result is shown in the bold numbers.
Table 5. The results obtained from various SLAM methods applied in the harbor. The best result is shown in the bold numbers.
SequenceORB-SLAM3HFNET-SLAMOurs
APERPEAPERPEAPERPE
010.1720.0330.1080.0220.0800.018
020.9320.4300.4720.4180.3490.374
030.0300.0220.0260.0090.0280.008
04××××1.4400.873
050.1410.0540.0930.0430.0740.041
060.0400.0200.0260.0140.0200.013
Avg0.2630.1120.1450.1010.1100.090
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Cheng, Q.; Mu, X. RU-SLAM: A Robust Deep-Learning Visual Simultaneous Localization and Mapping (SLAM) System for Weakly Textured Underwater Environments. Sensors 2024, 24, 1937. https://doi.org/10.3390/s24061937

AMA Style

Wang Z, Cheng Q, Mu X. RU-SLAM: A Robust Deep-Learning Visual Simultaneous Localization and Mapping (SLAM) System for Weakly Textured Underwater Environments. Sensors. 2024; 24(6):1937. https://doi.org/10.3390/s24061937

Chicago/Turabian Style

Wang, Zhuo, Qin Cheng, and Xiaokai Mu. 2024. "RU-SLAM: A Robust Deep-Learning Visual Simultaneous Localization and Mapping (SLAM) System for Weakly Textured Underwater Environments" Sensors 24, no. 6: 1937. https://doi.org/10.3390/s24061937

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop