Next Article in Journal
Flight Trajectory Prediction Based on Automatic Dependent Surveillance-Broadcast Data Fusion with Interacting Multiple Model and Informer Framework
Previous Article in Journal
Study of In-Vehicle Ethernet Message Scheduling Based on the Adaptive Frame Segmentation Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DA-IRRK: Data-Adaptive Iteratively Reweighted Robust Kernel-Based Approach for Back-End Optimization in Visual SLAM

Electrical and Power Engineering, Yingxi Campus, Taiyuan University of Technology, No. 79 Yingze West Street, Wanbailin District, Taiyuan 030024, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(8), 2529; https://doi.org/10.3390/s25082529
Submission received: 5 March 2025 / Revised: 12 April 2025 / Accepted: 14 April 2025 / Published: 17 April 2025
(This article belongs to the Section Navigation and Positioning)

Abstract

:
Back-end optimization is a key process to eliminate the cumulative error in Visual Simultaneous Localization and Mapping (VSLAM). Existing VSLAM frameworks often use kernel function-based back-end optimization methods. However, these methods typically rely on fixed kernel parameters based on the chi-square test, assuming Gaussian-distributed reprojection errors. In practice, though, reprojection errors are not always Gaussian, which can reduce robustness and accuracy. Therefore, we propose a data-adaptive iteratively reweighted robust kernel (DA-IRRK) approach, which combines median absolute deviation (MAD) with iteratively reweighted strategies. The robustness parameters are adaptively adjusted according to the MAD of reprojection errors, and the Huber kernel function is used to demonstrate the implementation of the back-end optimization process. The method is compared with other robust function-based approaches via the EuRoC dataset and the KITTI dataset, showing adaptability across different VSLAM frameworks and demonstrating significant improvements in trajectory accuracy on the vast majority of dataset sequences. The statistical analysis of the results from the perspective of reprojection error indicates DA-IRRK can tackle non-Gaussian noises better than the compared methods.

1. Introduction

The evolution of Visual Simultaneous Localization and Mapping (VSLAM) technology has been driven primarily by its widespread adoption in mobile robotics, autonomous navigation systems, and immersive reality applications [1]. However, real-world environmental complexities continue to challenge existing VSLAM approaches [2]. Key challenges include maintaining accuracy and robustness when processing noisy, incomplete data containing outliers.
Fundamentally, VSLAM solves state estimation by determining camera poses and reconstructing 3D environments from visual data. The technique’s dependence on visual features makes it vulnerable to motion artifacts and texture ambiguities [3]. Mismatched feature correspondences often generate outliers that compromise system accuracy [4].
Back-end optimization plays a pivotal role in refining VSLAM’s mapping precision. Pioneer VSLAM implementations predominantly utilized filter-based back-end optimizations. Such filters often fail to achieve global consistency while exhibiting outlier sensitivity [5]. Modern VSLAM systems like ORB-SLAM2/3 [3,6] predominantly leverage nonlinear optimization.These approaches formulate pose estimation and mapping as optimization tasks, minimizing cost functions through specialized algorithms.
Contemporary frameworks like ORB-SLAM3 and CCM-SLAM [7] enhance robustness through kernel-function-based optimization. Conventional implementations typically employ fixed-parameter kernel functions. Diverse kernel functions (Huber, Cauchy, Tukey, Geman-McClure) [8,9,10,11] have emerged for different scenarios. Kernel selection remains challenging, given environmental variability and limited function choices. Parameter estimation is further complicated by interdependent shape and scaling parameters. Moreover, few enhanced methods have been validated in actual VSLAM implementations.
To address the above issues, various recent studies have considered data adaptation methods. Ref. [12] proposed a generalized robust loss function for machine learning only. Ref. [13] proposed a new data-driven criterion to solve the robustness parameter of the Huber estimator, to achieve an adaptive Huber estimator and [14] applied the iteratively reweighted strategy to the adaptive robust parameter estimation to optimize the linear regression model problem.
Inspired by [13,14], we propose an iteratively reweighted robust kernel function method based on data adaptation to solve the VSLAM back-end optimization problem. Different from previous methods, the proposed method only needs to estimate the robustness parameter of the kernel function according to the reprojection error data in VSLAM, which can avoid the coupling problem of multi-parameter estimation. Furthermore, the robustness parameter can be adaptively adjusted according to the environmental changes to achieve better optimization results for different scenes. Our main contributions are as follows:
(1)
A data-adaptive iteratively reweighted robust kernel-based (DA-IRRK) method is proposed for back-end optimization in VSLAM. In this method, a kernel function is adopted as the objective function for the back-end optimization problem. The robustness parameter in the kernel function is adaptively updated according to the reprojection error, which is reflected through the median absolute deviation (MAD). The proposed method brings out robustness for different scenarios. In addition, the formulated back-end optimization problem is solved iteratively through a reweighted updating process.
(2)
The proposed method is implemented in different VSLAM frameworks, including ORB-SLAM3, JORB-SLAM, and CCM-SLAM, to demonstrate its effectiveness in visual-only SLAM, multi-sensor fusion SLAM, and collaborative VSLAM. The proposed method is tested on both indoor and outdoor datasets and compared with other robust kernel methods as well as the state-of-the-art MCMCC method.
(3)
The performance difference between the proposed method and other methods is analyzed from the perspective of reprojection error statistics, which provides insights into the VSLAM back-end problem in the context of adaptivity.
The remainder of the paper is organized as follows. Section 2 reviews related work on the VSLAM framework and back-end optimization. Section 3 describes the back-end optimization problem in VSLAM, presents the proposed DA-IRRK method, and takes the Huber kernel as an example to demonstrate the implementation of the DA-IRRK method. Section 4 shows experimental validation for three different modes of visual-only SLAM, multi-sensor fusion VSLAM, and collaborative VSLAM in different VSLAM frameworks, including ORB-SLAM3, JORB-SLAM, and CCM-SLAM. Section 5 draws conclusions and outlines future work.

2. Related Work

In this section, we focus on back-end optimization methods for VSLAM. Since different VSLAM frameworks use different back-end optimization methods, we first analyze different VSLAM frameworks and then investigate the existing back-end optimization methods. Finally, we discuss data-driven adaptive-based optimization methods.

2.1. VSLAM Frameworks

VSLAM frameworks can be classified into three categories: single-robot VSLAM, multi-robot VSLAM, and CNN-based VSLAM [15]. Single-robot VSLAM systems, such as ORB-SLAM3 [6], are extensively studied for their fast speed and promising accuracy in various scenarios. Multi-robot VSLAM frameworks, such as CCM-SLAM [7] and JORB-SLAM [16], use multiple robots to map complex environments and can address the problems of high individual cost and centralized computation faced by single-robot SLAM systems [17]. Advanced CNN-based VSLAM frameworks [18,19] have also been explored due to the performance improvement by Convolutional Neural Networks (CNN). However, CNN-based VSLAMs, often time-consuming and scenario-specific, need further performance enhancements for real-time applications. In this study, we adopt the single-robot SLAM framework ORB-SLAM3 and the multi-robot VSLAM frameworks CCM-SLAM and JORB-SLAM to demonstrate and verify the proposed method.

2.2. Back-End Optimization

Back-end optimization refines the preliminary estimates from the front end in VSLAM. Filter-based [5] methods and nonlinear optimization methods [20] are currently the most popular for back-end optimization. In early VSLAM implementations, filter-based methods, such as the Extended Kalman Filter (EKF) method, were prevalent. However, the EKF lacks an outlier elimination mechanism [21,22], and even a few outliers may significantly degrade the optimization performance. Furthermore, in large-scale VSLAM, the growing number of frames and map points slows the optimization process as it requires dynamic updates of both the mean and the variance, along with the state size. On the other hand, nonlinear optimization methods, such as bundle adjustment, which formulate the state estimation of the camera pose and map points as an optimization problem, dominate the mainstream SLAM frameworks (such as PTAM [23], ORB-SLAM3, and DSO [24]). Notably, most modern SLAM systems, including ORB-SLAM3, rely on general-purpose optimization back-ends like g2o [25] rather than implementing their own optimization modules. These frameworks usually adopt Gauss–Newton [26] or Levenberg–Marquardt (L–M) algorithms [27] to solve the optimization problem.

2.3. Robust Kernel Functions

Most nonlinear optimization methods aim to minimize the sum of squared error to counter mismatches. If the data samples are outliers, this approach tends to fail to identify incorrect data and cause estimation errors. Consequently, using robust kernel functions as the objective function has become a common practice to prevent errors on one edge from overshadowing others [28]. Refs. [4,29] applied these kernel functions to various estimation problems in computer vision and robotics. Ref. [30] analyzed popular robust kernel functions for alignment problems, offering suggestions for kernel function selection based on the application scenarios.
However, the abovementioned methods manually select different robust kernel functions depending on the situation, lacking flexibility and robustness for new scenes. Ref. [31] introduced a generalized adaptive robust kernel function, which makes the above robust kernel functions a special case of the generalized robust kernel function. By setting the shape parameter differently, the generalized adaptive robust kernel function can automatically select a suitable robust kernel function according to different situations. Ref. [32] improved the generalized robust kernel family based on [31] by adapting robust kernel shapes through the probability distribution of the generalized loss function and applying the improved method to nonlinear optimization problems. Ref. [33] associated the M-estimator with an elliptic probability distribution. They estimate hyper-parameters for each kernel type based on the residual distribution and perform model comparisons to determine the best kernel for the situation at hand. All of the above adaptive methods focus on estimating shape parameters and then finding a suitable robust kernel function.

2.4. Adaptive Methods

With the popularity of data-driven approaches, various recent works have reconsidered data-adaptive methods, giving rise to two fundamental parameter optimization paradigms: adaptive methods and self-learning mechanisms [34]. Adaptive methods (e.g., the M-estimator parameter adjustment proposed by [35]) primarily tune model parameters by analyzing real-time data distribution characteristics (such as reprojection errors), with [35] demonstrating that such adaptive methods significantly enhance robustness compared with static M-estimators across different data distributions. In contrast, self-learning mechanisms (as seen in recent path planning research) achieve continuous system self-optimization through built-in feedback loops.
Early ideas for adaptively choosing the tuning parameters of M-estimators are presented in [35] based on different data distributions. Recent studies by [13,36] have further validated the necessity of data-dependent adjustment for Huber estimator parameters to effectively suppress outliers, showing these adaptive parameters can be customized according to variations in sample size, dimensionality, and noise characteristics. Ref. [14] advanced this direction by combining iteratively reweighted least squares with adaptive algorithms for parameter estimation, successfully applying this methodology to linear regression model optimization problems.
In this paper, we propose a data-adaptive iteratively reweighted robust kernel-based (DA-IRRK) approach by drawing on the data-driven adaptive robust Huber kernel function method and iteratively reweighted strategy in [14,35], in which the robustness parameter of the robust kernel function is adaptively changed based on the distribution of reprojection error in the real environment, rather than a fixed robustness parameter in all cases. The proposed method can be implemented in the back end of any VSLAM framework.

3. Back-End Optimization Based on DA-IRRK

Figure 1 illustrates the general VSLAM framework that the proposed back-end optimization method applies to. We focus on the VSLAM back-end module. We elaborated the proposed method and its implementation following the preliminary work related to back-end optimization.

3.1. Back-End Optimization

State estimation in VSLAM involves determining the unknown parameters of the motion model based on noisy observations z i j , which include camera position χ c and map point χ m , collectively termed χ . The state estimation problem can be viewed as a nonlinear least squares (LS) optimization problem in the back end. While Equation (1) presents the general formulation, practical implementations in frameworks like ORB-SLAM3 using g2o solve a modified version that accounts for additional engineering considerations [25]. Back-end optimization consists of both local and global optimization: local optimization refines the pose and map points of a few keyframes locally, while global optimization optimizes all keyframes after loop closure detection. The optimization problem is typically solved by minimizing the objective function of reprojection error, expressed as follows:
X * = argmin L ( χ ) = argmin 1 2 i = 1 N j = 1 M ω i j e i j ( χ ) 2
where i varies from 1 to N, and N is the number of frames; j varies from 1 to M, and M is the number of map points that can be seen from the position of the i-th frame. The reprojection error, denoted as e i j ( χ ) = z i j f i j ( χ ) refers to the discrepancy between the camera observation z i j and the reprojected 3D map point χ m j onto the i-th 2D frame using the projection function f i j ( · ) . The weight ω i j represents the information matrix (inverse covariance Σ i j 1 ), encoding the uncertainty of observation z i j under the Gaussian assumption z i j N ( f i j ( χ ) , Σ i j ) . The observation z i j represents χ m j in the i-th frame, typically in pixel coordinates. The importance of the i-th reprojection error is represented by its weight ω i j . Figure 2 illustrates the reprojection process in VSLAM. While X * is statistically optimal when the error of the observation z i j in the i-th frame follows a Gaussian distribution, it may deviate from this optimal solution in cases involving non-Gaussian noises [11].
When e i j ( χ ) follows a Gaussian distribution, the weighted LS problem shown in Equation (1) can be directly solved by the Gauss–Newton algorithm or Levenberg–Marquardt (L–M) algorithm. However, due to the non-convex characteristics of the image and the influence of environmental noises, e i j ( χ ) typically follows a sub-Gaussian distribution or non-Gaussian distribution. The estimation results obtained using only the Gauss–Newton algorithm or L–M algorithm often deviate from the true value, leading to a decrease in localization and mapping accuracy in VSLAM.

3.2. DA-IRRK

In the field of VSLAM, heavy-tailed distribution data affected by sub-Gaussian or non-Gaussian noises are often observed in both datasets and real scene experiments, leading to results with large errors [37]. Robust kernel functions with specific robustness parameters are widely used in the field of VSLAM to solve the back-end optimization problem and improve mapping accuracy, as the robustness parameter can balance accuracy and robustness. According to [38], the robustness parameter should be adapted to the sample size, dimensionality, noise variance, and confidence level of the experimental data. While Equation (2) presents a general form of robust kernel formulation, practical implementations like ORB-SLAM3 properly incorporate the information matrix of measurements when applying the Huber kernel in their g2o-based optimization back-end. In VSLAM, Equation (2) shows the objective function based on the robust kernel function, where l ( · ) represents the robust loss or kernel.
X * = argmin L ( χ ) = argmin i = 1 N j = 1 M l e i j ( χ )
In previous studies, the robustness parameter is typically set to a fixed value to adapt to specific scene noises. However, this approach requires re-determining the robustness parameter whenever the scene changes, which results in the designed back-end optimization algorithm having poor robustness and low mapping accuracy in practical applications. To address this issue, an adaptive back-end optimization algorithm is proposed that can adaptively adjust the robustness parameters according to different scenarios, thereby improving the robustness of the back-end and enhancing the mapping accuracy.
The Huber kernel function is a popular robust kernel function in VSLAM, which is widely used in mature VSLAM back-end optimization modules such as ORB-SLAM2 and ORB-SLAM3. Therefore, in the following algorithm presentation, the proposed method is demonstrated using the Huber kernel function as an example. The Huber kernel function and its corresponding first-order derivatives are presented in Equations (3) and (4), where τ represents the robustness parameter and e is the error.
l Huber ( e ) = 1 2 e 2 , i f | e | τ τ | e | 1 2 τ 2 , i f | e | > τ
φ Huber ( e ) = e , i f | e | τ s g n ( e ) τ , i f | e | > τ
Equations (3) and (4) demonstrate that τ plays a crucial role in determining the weight of the errors in the optimization process. When e is less than or equal to τ , the back-end optimization problem based on the Huber kernel function can be regarded as an LS problem, corresponding to the case where ω i j = 1 in Equation (1). On the other hand, when e exceeds τ , the influence of error samples on the optimization process is significantly suppressed. A comprehensive analysis provided by [20] concludes that the choice of τ determines the robustness and sensitivity of the back-end optimization algorithm. Therefore, we propose an adaptive tuning approach for τ in our back-end optimization algorithm based on the Huber kernel function. The aim is to solve the optimization problem shown in Equation (2) using an iteratively reweighted strategy.
To solve Equation (2), we utilize the adaptive method proposed by [39] and integrate it with existing back-end optimization algorithms, such as the L–M algorithm, which can be implemented in the following five steps:
(1)
Calculate the current reprojection error based on the camera pose and map points at the front end;
(2)
Determine the adaptive threshold σ according to Equation (6) after computing the MAD of reprojection errors using Equation (5), where m e d i a n x represents the median of a set of data and x i represents the i-th sample in the set. The MAD strategy is preferred over standard deviation due to its higher breakdown point (50% vs. 0%) for outlier resistance [40], which is crucial for handling outliers in SLAM. The coefficient 1.4826 in Equation (6) equals 1 / Φ 1 ( 0.75 ) , where Φ 1 is the inverse standard normal CDF. This scaling ensures σ M A D matches the standard deviation for normally distributed data, maintaining compatibility with Gaussian kernels [41]. This adaptive mechanism enables automatic adjustment of the robustness parameters based on real-time sensor data characteristics, representing a key advantage over traditional fixed-parameter methods like MCMCC. These computations are performed in the tangent space of the manifold, which provides a vector space approximation for the nonlinear optimization problem while maintaining the geometric properties of the original space. It is worth noting that the MAD strategy may not be sufficiently robust for data distributions containing a larger number of large outliers;
MAD ( x ) = m e d i a n x i m e d i a n ( x )
σ M A D e i j ( χ ) = MAD e i j ( χ ) φ 1 ( 0.75 )
(3)
Calculate the robustness parameter in the Huber kernel function using Equation (7), where c represents the scaling factor. The scale factor c = 5.99 corresponds to 95% χ 2 confidence. T is the vector form of τ ;
T = c σ M A D e i j ( χ )
(4)
Solve the objective function iteratively through a reweighted updating process;
(5)
Utilize the L–M algorithm to update the camera pose and map points.
After completing the previous three steps, it is necessary to find the derivative of L ( χ ) before performing the subsequent steps, which are represented as Equation (8), and the iteratively reweighted strategy is used for further simplification, as in Equation (9). The iteratively reweighted strategy has also been used in [42] to achieve reweighting in the optimization equations, and experimental results show its effectiveness.
L ( χ ) χ = i = 1 N j = 1 M d l e i j ( χ ) d e i j ( χ ) e i j ( χ ) χ
L ( χ ) χ = i = 1 N j = 1 M d l e i j ( χ ) d e i j ( χ ) e i j ( χ ) χ = i = 1 N j = 1 M ω i j e i j ( χ ) e i j ( χ ) χ = i = 1 N j = 1 M ω i j ( χ ) e i j ( χ ) f i j ( χ ) χ = 0
where the first-order derivative of l e i j ( χ ) is defined as l e i j ( χ ) = d l e i j ( χ ) d e i j ( χ ) ) , l e i j ( χ ) represents the influence function and ω i j ( χ ) = l e i j ( χ ) e i j ( χ ) is the weight function. After simplifying Equation (2) using the iteratively reweighted strategy described above, the optimization of Equation (2) can be further achieved by the L–M algorithm.
In the L–M algorithm, the projection equation f i j ( χ ) is firstly Taylor expanded (Equation (10)), then Equations (11) and (12) are solved sequentially, and finally, the matrix form of the solution to Equation (12) is expressed as Equation (13), which is the update vector Δ χ of the L–M algorithm after adding the damping coefficient λ (Equation (14)).
f i j ( χ ) f i j χ k + f i j χ k χ χ χ k = f i j χ k + J · Δ χ
e i j ( χ ) = z i j f i j ( χ ) = z i j f i j χ k + f i j χ k f i j ( χ ) Δ e i j J · Δ χ
L ( Δ χ ) Δ χ = i = 1 N j = 1 M ω i j ( Δ χ ) · Δ e i j J · Δ χ · f i j ( Δ χ ) Δ χ = i = 1 N j = 1 M ω i j ( Δ χ ) · Δ e i j J · Δ χ · J = 0
Δ χ = J T W J 1 J T W Δ E
Δ χ = J T W J + λ · diag J T W J 1 J T W Δ E
where W is treated as a diagonal matrix.
During the iteration process, when λ approaches zero, it signifies that the L–M algorithm behaves similarly to the Gauss–Newton method. Conversely, a high damping coefficient aligns the update vector with the direction of the gradient, making the L–M algorithm a compromise between the Gauss–Newton algorithm and gradient descent. The implementation of the proposed method DA-IRRK is summarized in Algorithm 1.
It is worth noting that ω i j is updated in each iteration, which is also why we name our method an iterative reweighted method. As the iteration carries on, the weight ω i j reflects the importance of each e i j with better and better accuracy, leading to the approaching of the global solution.
Algorithm 1 DA-IRRK-based back-end optimization.
Require: Camera pose χ c and map point χ m , denoted as χ in the algorithm;
Ensure: Updated model parameters χ c and χ m ;
  1:
Initialization: Set the initial camera pose and map point χ 0 . Define the robustness parameter T of the robust kernel and set the parameter σ of the DA-IRRK method to zero;
  2:
for  k = 1 to N do
  3:
    if  k = 1  then
  4:
        Calculate the e i j ( χ ) using e i j ( χ ) = z i j f i j ( χ ) , update e i j ( χ ) using Equation (2);
  5:
        Compute σ and T according to Equations (5)–(7);
  6:
        Define the updated σ as σ 0 , which is passed into the DA-IRRK algorithm;
  7:
    else
  8:
        With the incoming σ 0 and updated e i j ( χ ) , compute T of the DA-IRRK method using Equation (7);
  9:
        Substitute the obtained T into Equations (3) and (4);
10:
        The objective function of Equation (2) is derived using Equations (8)–(10) to obtain the update quantity Δ χ ;
11:
        Equation (13) is deformed into Equation (14) and χ c and χ m are updated iteratively using the L–M algorithm;
12:
    end if
13:
end for

4. Experiments

We evaluate the proposed DA-IRRK method along with other robust kernel function-based back-end optimization methods on both indoor and outdoor datasets in different VSLAM frameworks, including ORB-SLAM3, CCM-SLAM [7], and JORB-SLAM [16]. The EuRoC dataset is utilized for indoor testing, while the KITTI dataset is used for outdoor validation to assess the accuracy and robustness of the DA-IRRK method. Eigen3 [43], G2O [25], PCL [44], Pangolin [45], and OpenCV [46] are employed for implementing the proposed method. All experiments are conducted on a laptop with an Intel Core i5 12400F CPU and 16 GB of RAM running Ubuntu 18.04. Here, we present the detailed implementation of the DA-IRRK method for visual SLAM back-end optimization, specifically realized within the Edge and Solver modules of the G2O library.
  • Adaptive Edge Module: New EdgeSE3-DAIRRK class implements the following:
    • Dynamic kernel parameter computation (Equations (5) and (6)).
    • Real-time robust kernel selection (Equation (7)).
    • Information matrix weighting mechanism.
  • Solver Module: Enhanced LinearSolverEigen features the following:
    • Optimized sparse matrix storage pattern.
    • Improved marginalization strategy.

4.1. Experimental Datasets

The DA-IRRK method was evaluated across two distinct environments: indoor scenarios using the EuRoC dataset [47] and outdoor settings with KITTI [48]. EuRoC captures three environment types: ETH Zurich’s machine halls (MH01-MH05), calibration chambers (V101-V103), and standard rooms (V201-V203). Based on feature availability and UAV flight dynamics, EuRoC sequences are categorized by difficulty:
  • Simple: MH01 (80.6 m), MH02 (73.5 m), V101 (58.6 m), V201 (36.5 m).
  • Moderate: MH03 (130.9 m), V102 (75.9 m), V202 (83.2 m).
  • Challenging: V103 (79.0 m), V203 (86.1 m), MH04 (91.7 m), MH05 (97.6 m).
The KITTI dataset includes urban, residential, rural, and highway scenarios. The primary data acquisition equipment includes a 64-beam Velodyne laser scanner, four cameras, and a GPS/IMU unit. We select 11 sequences (00–10) to verify the proposed method in the outdoor environment. The sequence lengths are as follows: sequence 00: 3724.187 m; sequence 01: 2453.203 m; sequence 02: 5067.233 m; sequence 03: 560.888 m; sequence 04: 393.645 m; sequence 05: 2205.576 m; sequence 06: 1232.876 m; sequence 07: 694.697 m; sequence 08: 3222.795 m; sequence 09: 1705.051 m; sequence 10: 919.518 m.

4.2. Experimental Frameworks and Evaluation Benchmarks

4.2.1. Experimental Frameworks

We evaluate the robustness of the proposed method by implementing it in ORB-SLAM3 for single-robot applications and in JORB-SLAM and CCM-SLAM for multi-robot systems.
ORB-SLAM3 is a single-robot VSLAM framework that supports multiple cameras and offers visual, visual-inertial, and multi-map SLAM for various applications. It is regarded as the most advanced VSLAM framework to date and is widely used as a benchmark for both indoor and outdoor testing. Notably, the DA-IRRK method is evaluated in ORB-SLAM3 for both single-robot VSLAM and multi-sensor fusion-based VSLAM, as the framework supports both visual-only and visual-inertial modes.
JORB-SLAM is a feature-based multi-robot VSLAM framework that extends ORB-SLAM2 to multiple agents, focusing on rapid map fusion and improved coverage. The system includes multiple ORB-SLAM2 clients for local mapping and a central server for global map fusion. The DA-IRRK method is tested within this framework, as it enhances robustness by adding additional constraints to the AprilTag.
CCM-SLAM is a centralized collaborative multi-robot VSLAM framework based on ORB-SLAM2 that ensures agent autonomy while fostering cooperation through a central server for map fusion and optimization, maintaining efficiency despite potential data loss or delays. Unlike JORB-SLAM, where each agent runs a complete sub-map, CCM-SLAM is widely used as a comparative collaborative VSLAM framework for multi-agent applications, such as CORB2I-SLAM [49] and COVINS [50]. We implement the DA-IRRK method in each CCM-SLAM agent and verify the performance improvement after modification.

4.2.2. Evaluation Metrics

The Absolute Trajectory Error (ATE) [51] is a widely used metric in VSLAM evaluation, measuring the global consistency of the estimated trajectory by comparing it to the ground truth. The accuracy is expressed through the Root Mean Square Error (RMSE) of the ATE. In this paper, the EVO [52] tool is used to compute ATE, providing an assessment of the trajectory accuracy of the VSLAM system against the ground truth. EVO also offers data analysis and visualization for odometry and SLAM evaluation.

4.3. Case Study

4.3.1. Visual-Only SLAM

This section evaluates the proposed method in visual-only SLAM mode using ORB-SLAM3 with a single agent and a visual sensor. We compare it with three robust kernel-based methods (Huber, Tukey, and Cauchy, all using ORB-SLAM3’s default kernel parameters) and the state-of-the-art (SOTA) method, MCMCC [42], to assess the RMSE of ATE across different scenarios using the EuRoC and KITTI datasets. Monocular scale blur causes errors in ORB-SLAM3 in open environments, complicating comparisons. Therefore, we evaluate these methods solely in stereo mode, with results presented in Table 1 and Table 2, where the top performer is bolded and the second-best is underlined.
Table 1 demonstrates that the proposed DA-IRRK method achieves higher accuracy in most indoor sequences compared with other methods, while the MCMCC method outperforms ours on MH03, V103, V201, and V203. Table 2 shows that the proposed DA-IRRK method shows lower accuracy than the comparing methods in sequences 03, 06, 09, and 10 on the KITTI dataset. Meanwhile, the MCMCC method outperforms ours in sequences 03 and 10. The analysis method in Table 2 is the same as that in Table 1.
(a)
Adaptive Robustness in DA-IRRK vs. Fixed-Parameter Huber Kernel
To further investigate these findings, we analyze sequences V101 and V203 for the EuRoC dataset and sequences 01 and 09 for the KITTI dataset as examples. As noted in Section 3.2, the DA-IRRK method adaptively adjusts the robustness parameter based on reprojection errors to minimize the impact of outliers on mapping accuracy, employing the MAD strategy for robust estimation of data distributions containing outliers (refer to Equations (5)–(7) in Section 3.2).
Figure 3 shows the correlation between reprojection error and robustness parameter for sequences V101 and V203 using the DA-IRRK method, while Figure 4 presents the same for the robust Huber kernel method. The Huber kernel’s robustness parameter is fixed at 5.99 for monoculars and 7.81 for stereo cameras, calculated using the chi-square test [41]. Unlike the Huber kernel, DA-IRRK’s parameter τ varies with reprojection errors, reflecting its data-dependent adaptation. However, the MAD-based estimation in DA-IRRK exhibits sensitivity to extreme outliers, leading to suboptimal performance in sequences like V203 with pronounced non-Gaussian noise (e.g., due to motion blur). Similarly, Figure 5 and Figure 6 show how robustness parameters and reprojection errors vary when applying DA-IRRK and the robust Huber kernel methods to ORB-SLAM3 on sequences 01 and 09. While DA-IRRK’s parameter varies with error, the Huber kernel’s remains constant. Sequence 09 exhibits larger reprojection errors, corresponding to higher robustness parameters. However, as τ is derived from the median error, its magnitude does not strictly correlate with error size—a limitation that occasionally compromises accuracy.
(b)
Noise Distribution Analysis and MAD Strategy Efficacy
We assess the impact of the MAD strategy by visualizing the reprojection error distributions for both global and local keyframes in individual sequences from the EuRoC dataset, as shown in Figure 7 and Figure 8. The reprojection error distributions of global keyframes in each sequence show a non-Gaussian pattern. In contrast, local keyframes (Figure 8)—sampled at critical phases (initialization, motion bursts, and stable tracking)—reveal sub-Gaussian distributions for sequences MH01, MH02, MH04, MH05, V101, V102, and V202, while others (e.g., MH03, V201, V203) retain non-Gaussian traits. The MAD-based DA-IRRK method effectively suppresses sub-Gaussian heavy-tailed outliers, achieving superior accuracy in 7 of 11 sequences (Table 1). However, its performance lags behind MCMCC for non-Gaussian outliers (e.g., V203), highlighting a limitation in handling extreme noise. This instability warrants future improvement, as noted in Section 5. Specifically, Figure 7 shows that reprojection errors in sequence V101 are mainly concentrated in the range of 0 to 20, with fewer errors between 20 and 60. In contrast, for sequence V203, errors are concentrated in the 0 to 20 range, but more errors fall between 20 and 60, impacting mapping accuracy. It is important to note that the results presented are snapshots selected from local optimal optimizations at random instances. Reprojection error distributions vary across different instances.
To investigate the root reason for varying reprojection error distributions, we analyze the scenes of sequences V101 and V203. In Figure 9, sequence V101 features brighter scenes and sequence V203 displays blurred scenes caused by fast-moving UAVs. V101 is captured when the UAV flies slowly, resulting in clear scene graphs. On the other hand, sequence V203 is recorded when the UAV is fast-moving, leading to a blurred texture in the scene graph [47]. As a consequence, the front-end feature extraction and matching process of VSLAM produces more mismatches due to the blurred texture, thereby increasing both the number and magnitude of reprojection error outliers.
The reprojection error distributions of individual sequences from the KITTI dataset in global and local keyframes are shown in Figure 10 and Figure 11. The results indicate that reprojection errors of global keyframes tend to be non-Gaussian distributed, while those of local keyframes are prone to be sub-Gaussian distributed on sequences 00, 01, 02, 04, 05, 07, and 08, whereas on the other sequences, the reprojection errors tend to be non-Gaussian distributed, both in the global keyframes and the local keyframes. Combined with Table 2, we can see that the MCMCC method proves effective at suppressing outliers of non-Gaussian type with heavy-tailed distribution, leading to higher mapping accuracy for sequences 03 and 10.
Specifically, Figure 11 shows that most reprojection errors of sequence 01 are concentrated within the range of 0 to 20, with only a few error samples in the range of 20 to 60. In contrast, reprojection errors of sequence 09 reach a maximum of 125, and a large number of reprojection errors lie in the range of 25 to 75. The average error is larger than that on the other seven sequences, causing the degradation of the mapping accuracy.
Figure 12 shows that sequence 01 exhibits clearer and more static features with minimal lighting variations, while sequence 09 contains dynamic elements (e.g., a moving car) and displays obvious lighting variations. For sequence 09, the effects of lighting and dynamic features result in more and larger reprojection error outliers.
Based on the above analysis, the results on the KITTI dataset coincide with those on the EuRoC dataset, both demonstrating the effectiveness of the proposed DA-IRRK method for single-robot VSLAM applications with the visual-only case.

4.3.2. Multi-Sensor Fusion VSLAM

To test the proposed DA-IRRK method in a multi-sensor fusion VSLAM mode, we also consider using ORB-SLAM3 in stereo visual-inertial mode. We switch from the visual-only mode to the visual-inertial mode in ORB-SLAM3. In this experiment, we specifically integrate the DA-IRRK method into the g2o optimization library while keeping the IMU-related components unchanged in the backend processing pipeline. Evaluation of the EuRoC dataset is assessed using the RMSE of ATE. Since [42] did not test the proposed MCMCC method in visual-inertial mode, we only provide the results of the proposed DA-IRRK method and the robust Huber kernel method to avoid redundancy since Huber is the most widely used compared with other kernels, and it is only inferior to the proposed method and MCMCC according to Table 1 and Table 2. Detailed results are presented in Table 3, where the top performer is bolded. The third-row Impr_huber in Table 3 quantifies the improvement of the proposed method compared with the Huber function-based method.
Table 3 shows that mapping accuracy has been improved on 9 out of 11 sequences, with 6 sequences indicating an improvement of more than 10%. The highest accuracy improvement is seen at 26.29%. Only the accuracy of mapping on sequences V201 and V203 decreases but by less than 9.5%. This serves as evidence of the effectiveness of the DA-IRRK method in visual-inertial mode.
To analyze the reasons for the degradation of mapping accuracy on sequences V201 and V203, we conducted a comparative analysis using sequences V201 and V202 as examples. In accordance with Section 4.3.1, our analysis begins with the proposed DA-IRRK method. Figure 13 and Figure 14 show the relationship between reprojection error and robustness parameters for both methods on sequences V201 and V202. The results consolidate those in the visual-only SLAM case.

4.3.3. Collaborative VSLAM

In this subsection, the effectiveness of the DA-IRRK approach is tested in collaborative VSLAM frameworks such as JORB-SLAM and CCM-SLAM, with only two agents adopted. Specifically, as in [7,16], we also test the proposed method in JORB-SLAM on the KITTI dataset and CCM-SLAM on the EuRoC dataset. JORB-SLAM divides a trajectory segment into overlapping parts in its experiments to achieve mapping through multi-robot collaboration, while CCM-SLAM achieves mapping by multiple robots collaborating to run respective maps.
In the experiments with CCM-SLAM, sequences from EuRoC are paired together (e.g., MH01-MH02, where agent A runs MH01 and agent B runs MH02) with map fusion performed at a central station. Evaluation is based on the RMSE of ATE compared only with the most widely used robust Huber kernel to avoid redundancy. Results are detailed in Table 4 and Table 5. The third row, Impr_huber, in Table 4 and Table 5, quantifies the improvement of the proposed method compared with the Huber function-based method.
The rule is developed to evaluate accuracy within a collaborative framework: the accuracy of the final trajectory is considered improved if both agent A and agent B enhance their trajectory accuracy after running on the dataset; additionally, the accuracy of the final trajectory is also considered enhanced if only one agent (A or B) improves its trajectory accuracy after running on the dataset but with the improvement higher than the accuracy degradation of the other; conversely, in any other case, we consider the accuracy of the final trajectory decreased. Following the above stipulation, we can see from Table 4 that the mapping accuracy on sequences 04, 07, and 08 decreases, while the mapping accuracy on the other 7 sequences improves greatly, with sequences 01, 02, 03, 06, and 10 improving by more than 10.23%.
We take sequences 07 and 06 as an example to demonstrate the performance difference. From Figure 15 and Figure 16, the robustness parameter of the DA-IRRK method fluctuates with the reprojection error, whereas that of the robust Huber kernel method remains constant. We can also see that the robustness parameter of sequence 06 is smaller than that of sequence 07, and similarly, the reprojection error of sequence 06 is smaller than that of sequence 07, which demonstrates that sequence 06 has a better mapping accuracy, which is consistent with the results in Table 4. In the reprojection error analysis of the sub-sequences of sequences 06 and 07, as shown in Figure 17. It is worth noting that sequence 06 tends to be a sub-Gaussian distribution, while sequence 07 tends to be a non-Gaussian distribution, and according to [14], we can know that the proposed DA-IRRK method is more effective in suppressing the sub-Gaussian distribution of error data, and consequently, the mapping accuracy of sequence 06 is better.
The reprojection error is related to the actual environment, and different environments have different reprojection error distributions. As can be seen from Figure 18, the trajectory of sequence 06 is relatively simple, characterized by numerous trees and moderate lighting, which provides clearer feature points for optimal tracking. On the other hand, the scene of sequence 07 has strong lighting, resulting in unclear feature points and, consequently, more mismatches during tracking, thereby reducing mapping accuracy.
The experimental results in the CCM-SLAM framework are presented in Table 5. We can see a significant accuracy difference between the two agents on MH03-MH04 and MH04-MH05, with one agent showing a much larger improvement than the other. Table 5 reveals that the DA-IRRK method outperforms the robust Huber kernel method in terms of mapping accuracy within the CCM-SLAM framework. This unexpected result opens up new possibilities for future advancements in VSLAM collaboration by improving accuracy levels.
Summarizing the above analysis, we can conclude the proposed DA-IRRK method not only works properly in different visual collaborative SLAM frameworks but also shows better mapping accuracy, which proves the effectiveness as well as the robustness of the proposed method in collaborative VSLAM frameworks.
The experiments in Section 4.3 demonstrate that while the DA-IRRK method improves mapping accuracy over the robust Huber kernel, this comes with an increased computational cost. To evaluate real-time performance, we conducted timing analysis on the EuRoC V103 sequence using ORB-SLAM3 (Table 6, where the top performer is bolded.). Results show the DA-IRRK method maintains real-time capability (≤33 ms/frame at 30 Hz) in both visual-only and visual-inertial modes, with only marginal time increases (≤0.1 s) versus the Huber kernel. The visual-inertial mode’s reduced keyframe optimization further minimizes computational overhead. Our frame-rate compliance tests confirm the method’s practical suitability for real-time VSLAM applications.

5. Conclusions

In VSLAM, the robust kernel method is commonly used for back-end optimization, but its fixed robustness parameter limits performance in varying scenarios. To address this, we developed a data-driven, adaptive, iteratively reweighted robust kernel method to improve accuracy and robustness. Experiments in both indoor and outdoor environments show that our method outperforms robust Huber, Cauchy, and Tukey kernels, as well as state-of-the-art MCMCC methods, particularly when sub-Gaussian distributions occur during reprojection. Additionally, multi-sensor experiments reveal that our method surpasses the robust Huber kernel, with lower performance only in sequences with large reprojection errors. Furthermore, in the JORB-SLAM and CCM-SLAM frameworks, our method not only operates efficiently but also achieves higher accuracy. However, it shares a limitation with the robust Huber kernel in non-Gaussian environments, where it underperforms compared with the latest MCMCC method. Future work will focus on (1) developing data-adaptive methods for non-Gaussian features in VSLAM; (2) extending the approach to multi-modal sensor fusion applications; and (3) optimizing real-time performance for edge computing devices.

Author Contributions

Conceptualization, Z.H. and J.W.; methodology, L.C.; software, Z.H.; validation, Z.H. and J.W.; formal analysis, Z.Z.; investigation, Z.Z. and G.Y.; resources, X.X.; data curation, Z.H.; writing—original draft preparation, Z.H.; writing—review and editing, L.C.; funding acquisition, L.C. All authors have read and agreed to the published version of the manuscript.

Funding

The work was surpported by the National Natural Science Foundation of China (62073232), Natural Science Foundation of Shanxi Province (202403021221054), International Science and Technology Cooperation Program of Shanxi Province (202104041101030) Natural Science Foundation of Shanxi Province (201901D211079).

Data Availability Statement

The dataset used in the paper is the public KITTI raw data, which can be downloaded at https://www.cvlibs.net/datasets/kitti/raw_data.php, (accessed on 1 December 2024). The other dataset used in the paper is the public EuRoC Dataset, which can be downloaded at https://projects.asl.ethz.ch/datasets/doku.php?id=kmavvisualinertialdatasets, (accessed on 1 December 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Jia, Y.; Yan, X.; Xu, Y. A Survey of simultaneous localization and mapping for robot. In Proceedings of the 2019 IEEE 4th Advanced Information Technology, Electronic and Automation Control Conference (IAEAC), Chengdu, China, 20–22 December 2019; pp. 857–861. [Google Scholar]
  2. Placed, J.A.; Strader, J.; Carrillo, H.; Atanasov, N.; Indelman, V.; Carlone, L.; Castellanos, J.A. A survey on active simultaneous localization and mapping: State of the art and new frontiers. IEEE Trans. Robot. 2023, 39, 1686–1705. [Google Scholar] [CrossRef]
  3. Mur-Artal, R.; Tardós, J.D. Orb-slam2: An open-source slam system for monocular, stereo, and rgb-d cameras. IEEE Trans. Robot. 2017, 33, 1255–1262. [Google Scholar] [CrossRef]
  4. Zhang, Z. Parameter estimation techniques: A tutorial with application to conic fitting. Image Vis. Comput. 1997, 15, 59–76. [Google Scholar] [CrossRef]
  5. Chen, S.Y. Kalman filter for robot vision: A survey. IEEE Trans. Ind. Electron. 2011, 59, 4409–4420. [Google Scholar] [CrossRef]
  6. Campos, C.; Elvira, R.; Rodríguez, J.J.G.; Montiel, J.M.; Tardós, J.D. Orb-slam3: An accurate open-source library for visual, visual–inertial, and multimap slam. IEEE Trans. Robot. 2021, 37, 1874–1890. [Google Scholar] [CrossRef]
  7. Schmuck, P.; Chli, M. CCM-SLAM: Robust and efficient centralized collaborative monocular simultaneous localization and mapping for robotic teams. J. Field Robot. 2019, 36, 763–781. [Google Scholar] [CrossRef]
  8. Huber, P.J. Robust estimation of a location parameter. In Breakthroughs in Statistics: Methodology and Distribution; Springer: New York, NY, USA, 1992; pp. 492–518. [Google Scholar]
  9. Steven, K. De Schrijver and El-Houssaine Aghezzaf and Hendrik Vanmaele. Double precision rational approximation algorithm for the inverse standard normal first order loss function. Appl. Math. Comput. 2012, 219, 1375–1382. [Google Scholar]
  10. Black, M.J.; Anandan, P. The robust estimation of multiple motions: Parametric and piecewise-smooth flow fields. Comput. Vis. Image Underst. 1996, 63, 75–104. [Google Scholar] [CrossRef]
  11. Bas, E. Robust fuzzy regression functions approaches. Inf. Sci. 2022, 613, 419–434. [Google Scholar] [CrossRef]
  12. Fu, S.; Wang, X.; Tang, J.; Lan, S.; Tian, Y. Generalized robust loss functions for machine learning. Neural Netw. 2024, 171, 200–214. [Google Scholar] [CrossRef]
  13. Wang, L.; Zheng, C.; Zhou, W.; Zhou, W.X. A new principle for tuning-free Huber regression. Stat. Sin. 2021, 31, 2153–2177. [Google Scholar] [CrossRef]
  14. Kargoll, B.; Omidalizarandi, M.; Loth, I.; Paffenholz, J.A.; Alkhatib, H. An iteratively reweighted least-squares approach to adaptive robust adjustment of parameters in linear regression models with autoregressive and t-distributed deviations. J. Geod. 2018, 92, 271–297. [Google Scholar] [CrossRef]
  15. Huang, B.; Zhao, J.; Liu, J. A survey of simultaneous localization and mapping with an envision in 6G wireless networks. arXiv 2019, arXiv:1909.05214. [Google Scholar]
  16. Chakraborty, K.; Deegan, M.; Kulkarni, P.; Searle, C.; Zhong, Y. Jorb-slam: A Jointly Optimized Multi-Robot Visual Slam. 2022. Available online: https://um-mobrob-t12-w19.github.io/docs/report.pdf (accessed on 1 December 2024).
  17. Zhao, X.; Shao, S.; Wang, T.; Fang, C.; Zhang, J.; Zhao, H. A review of multi-robot collaborative simultaneous localization and mapping. In Proceedings of the 2023 IEEE International Conference on Unmanned Systems (ICUS), Hefei, China, 13–15 October 2023; pp. 900–905. [Google Scholar]
  18. Wei, X.; Zhang, Y.; Li, Z.; Fu, Y.; Xue, X. Deepsfm: Structure from motion via deep bundle adjustment. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28 2020; Proceedings, Part I 16. Springer International Publishing: Cham, Switzerland, 2020; pp. 230–247. [Google Scholar]
  19. Teed, Z.; Deng, J. Droid-slam: Deep visual slam for monocular, stereo, and rgb-d cameras. Adv. Neural Inf. Process. Syst. 2021, 34, 16558–16569. [Google Scholar]
  20. Saxena, A.; Chiu, C.Y.; Shrivastava, R.; Menke, J.; Sastry, S. Simultaneous localization and mapping: Through the lens of nonlinear optimization. IEEE Robot. Autom. Lett. 2022, 7, 7148–7155. [Google Scholar] [CrossRef]
  21. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef]
  22. Zheng, B.; Zhang, Z. An Improved EKF-SLAM for Mars Surface Exploration. Int. J. Aerosp. Eng. 2019, 2019, 7637469. [Google Scholar] [CrossRef]
  23. Klein, G.; Murray, D. Parallel tracking and mapping for small AR workspaces. In Proceedings of the 2007 6th IEEE and ACM International Symposium on Mixed and Augmented Reality, Nara, Japan, 13–16 November 2007; pp. 225–234. [Google Scholar]
  24. Engel, J.; Koltun, V.; Cremers, D. Direct sparse odometry. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 611–625. [Google Scholar] [CrossRef]
  25. Kümmerle, R.; Grisetti, G.; Strasdat, H.; Konolige, K.; Burgard, W. G2o: A general framework for graph optimization. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 3607–3613. [Google Scholar]
  26. Wang, Y. Gauss–newton method. Wiley Interdiscip. Rev. Comput. Stat. 2012, 4, 415–420. [Google Scholar] [CrossRef]
  27. Gavin, H.P. The Levenberg-Marquardt algorithm for nonlinear least squares curve-fitting problems. Dep. Civ. Environ. Eng. Duke Univ. August 2019, 3, 1–23. [Google Scholar]
  28. MacTavish, K.; Barfoot, T.D. At all costs: A comparison of robust cost functions for camera correspondence outliers. In Proceedings of the 2015 12th Conference on Computer and Robot Vision, Halifax, NS, Canada, 3–5 June 2015; pp. 62–69. [Google Scholar]
  29. Bosse, M.; Agamennoni, G.; Gilitschenski, I. Robust estimation and applications in robotics. Found. Trends® Robot. 2016, 4, 225–269. [Google Scholar] [CrossRef]
  30. Babin, P.; Giguere, P.; Pomerleau, F. Analysis of robust functions for registration algorithms. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 1451–1457. [Google Scholar]
  31. Barron, J.T. A general and adaptive robust loss function. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4331–4339. [Google Scholar]
  32. Chebrolu, N.; Läbe, T.; Vysotska, O.; Behley, J.; Stachniss, C. Adaptive robust kernels for non-linear least squares problems. IEEE Robot. Autom. Lett. 2021, 6, 2240–2247. [Google Scholar] [CrossRef]
  33. Agamennoni, G.; Furgale, P.; Siegwart, R. Self-tuning M-estimators. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 4628–4635. [Google Scholar]
  34. Ye, F.; Duan, P.; Meng, L.; Sang, H.; Gao, K. An enhanced artificial bee colony algorithm with self-learning optimization mechanism for multi-objective path planning problem. Eng. Appl. Artif. Intell. 2025, 149, 110444. [Google Scholar] [CrossRef]
  35. Kelly, G. Adaptive choice of tuning constant for robust regression estimators. J. R. Stat. Soc. Ser. D Stat. 1996, 45, 35–40. [Google Scholar] [CrossRef]
  36. Sun, Q.; Zhou, W.X.; Fan, J. Adaptive huber regression. J. Am. Stat. Assoc. 2020, 115, 254–265. [Google Scholar] [CrossRef]
  37. Catoni, O. Challenging the empirical mean and empirical variance: A deviation study. Ann. l’IHP Probab. Stat. 2012, 48, 1148–1185. [Google Scholar] [CrossRef]
  38. Fan, J.; Li, Q.; Wang, Y. Estimation of high dimensional mean regression in the absence of symmetry and light tail assumptions. J. R. Stat. Soc. Ser. B Stat. Methodol. 2017, 79, 247–265. [Google Scholar] [CrossRef] [PubMed]
  39. Ruppert, D. Robust Statistics: The Approach Based on Influence Functions. Technometrics 1987, 29, 240–241. [Google Scholar] [CrossRef]
  40. Huber, P.J. Robust Statistics. In International Encyclopedia of Statistical Science; Springer: Berlin/Heidelberg, Germany, 2011; pp. 1248–1251. [Google Scholar]
  41. Olson, E.; Strom, J.; Morton, R.; Richardson, A.; Ranganathan, P.; Goeddel, R.; Bulic, M.; Crossman, J.; Marinier, B. Progress toward multi-robot reconnaissance and the MAGIC 2010 competition. J. Field Robot. 2012, 29, 762–792. [Google Scholar] [CrossRef]
  42. Cheng, L.; Wang, T.; Xu, X.; Yan, G.; Ren, M.; Zhang, Z. Nonlinear back-end optimization method for VSLAM with multi-convex combined maximum correntropy criterion. ISA Trans. 2023, 142, 731–746. [Google Scholar] [CrossRef]
  43. Guennebaud, G.; Jacob, B. Eigen v3. 2010. Available online: http://eigen.tuxfamily.org (accessed on 1 December 2024).
  44. Będkowski, J.; Pełka, M.; Majek, K.; Fitri, T.; Naruniec, J. Open source robotic 3D mapping framework with ROS—Robot Operating System, PC—Point Cloud Library and Cloud Compare. In Proceedings of the 2015 International Conference on Electrical Engineering and Informatics (ICEEI), Denpasar, Indonesia, 10–11 August 2015; pp. 644–649. [Google Scholar]
  45. Chen, X.; Dathathri, R.; Gill, G.; Pingali, K. Pangolin: An efficient and flexible graph mining system on CPU and GPU. Proc. VLDB Endow. 2020, 13, 1190–1205. [Google Scholar] [CrossRef]
  46. Pulli, K.; Baksheev, A.; Kornyakov, K.; Eruhimov, V. Anatoly and Kornyakov, Kirill and Eruhimov, Victor. Real-time computer vision with OpenCV. Commun. ACM 2012, 55, 61–69. [Google Scholar] [CrossRef]
  47. Burri, M.; Nikolic, J.; Gohl, P.; Schneider, T.; Rehder, J.; Omari, S.; Achtelik, M.W.; Siegwart, R. The EuRoC micro aerial vehicle datasets. Int. J. Robot. Res. 2016, 35, 1157–1163. [Google Scholar] [CrossRef]
  48. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? In the kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  49. Saha, A.; Dhara, B.C.; Umer, S.; AlZubi, A.A.; Alanazi, J.M.; Yurii, K. Corb2i-slam: An adaptive collaborative visual-inertial slam for multiple robots. Electronics 2022, 11, 2814. [Google Scholar] [CrossRef]
  50. Schmuck, P.; Ziegler, T.; Karrer, M.; Perraudin, J.; Chli, M. Covins: Visual-inertial slam for centralized collaboration. In Proceedings of the 2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Bari, Italy, 4–8 October 2021; pp. 171–176. [Google Scholar]
  51. Zhang, Z.; Scaramuzza, D. A tutorial on quantitative trajectory evaluation for visual (-inertial) odometry. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 7244–7251. [Google Scholar]
  52. Rebecq, H.; Horstschaefer, T.; Gallego, G.; Scaramuzza, D. EVO: A Geometric Approach to Event-Based 6-DOF Parallel Tracking and Mapping in Real Time. IEEE Robot. Autom. Lett. 2017, 2, 593–600. [Google Scholar] [CrossRef]
Figure 1. General framework. The framework illustrates three forms of VSLAM: single-robot VSLAM, multi-sensor fusion VSLAM, and multi-robot VSLAM. The data input module transmits data to the front-end module for feature extraction, IMU data integration, and RANSAC optimization. The back-end module performs iterative optimization. Optimized camera positions and map points serve as inputs for the mapping module, facilitating mapping.
Figure 1. General framework. The framework illustrates three forms of VSLAM: single-robot VSLAM, multi-sensor fusion VSLAM, and multi-robot VSLAM. The data input module transmits data to the front-end module for feature extraction, IMU data integration, and RANSAC optimization. The back-end module performs iterative optimization. Optimized camera positions and map points serve as inputs for the mapping module, facilitating mapping.
Sensors 25 02529 g001
Figure 2. Illustration of the reprojection process. χ m j represents the 3D map point in global reference system, and χ c i denotes the camera position. The black box signifies the pixel coordinate system, with the black point indicating the observation and the red point representing the 2D reprojection. The blue solid line depicts the projection process, while the red dashed line illustrates the reprojection process. Additionally, two gray dashed lines trace both the map point trajectory and camera trajectory, respectively.
Figure 2. Illustration of the reprojection process. χ m j represents the 3D map point in global reference system, and χ c i denotes the camera position. The black box signifies the pixel coordinate system, with the black point indicating the observation and the red point representing the 2D reprojection. The blue solid line depicts the projection process, while the red dashed line illustrates the reprojection process. Additionally, two gray dashed lines trace both the map point trajectory and camera trajectory, respectively.
Sensors 25 02529 g002
Figure 3. Relationship between reprojection errors and robustness parameter τ for the DA-IRRK method in sequences V101 and V203. (a) The left is for V101. (b) The right is for V203.
Figure 3. Relationship between reprojection errors and robustness parameter τ for the DA-IRRK method in sequences V101 and V203. (a) The left is for V101. (b) The right is for V203.
Sensors 25 02529 g003
Figure 4. Relationship between reprojection errors and robustness parameter τ for robust Huber kernel method in sequences V101 and V203. (a) The left is for V101. (b) The right is for V203.
Figure 4. Relationship between reprojection errors and robustness parameter τ for robust Huber kernel method in sequences V101 and V203. (a) The left is for V101. (b) The right is for V203.
Sensors 25 02529 g004
Figure 5. Relationship between reprojection errors and robustness parameter τ for the DA-IRRK method in sequences 01 and 09. (a) The left is for 01. (b) The right is for 09.
Figure 5. Relationship between reprojection errors and robustness parameter τ for the DA-IRRK method in sequences 01 and 09. (a) The left is for 01. (b) The right is for 09.
Sensors 25 02529 g005
Figure 6. Relationship between reprojection errors and robustness parameter τ for the robust Huber kernel method in sequences 01 and 09. (a) The left is for 01. (b) The right is for 09.
Figure 6. Relationship between reprojection errors and robustness parameter τ for the robust Huber kernel method in sequences 01 and 09. (a) The left is for 01. (b) The right is for 09.
Sensors 25 02529 g006
Figure 7. Reprojection error distribution of global keyframes across sequences MH01-V203 in the EuRoC dataset.
Figure 7. Reprojection error distribution of global keyframes across sequences MH01-V203 in the EuRoC dataset.
Sensors 25 02529 g007
Figure 8. Reprojection error distribution plots of localized keyframes at three different timesteps for sequences MH01-V203 in the EuRoC dataset. The three horizontal subplots per sequence represent one moment. (ak) correspond to sequences MH01-V203, respectively.
Figure 8. Reprojection error distribution plots of localized keyframes at three different timesteps for sequences MH01-V203 in the EuRoC dataset. The three horizontal subplots per sequence represent one moment. (ak) correspond to sequences MH01-V203, respectively.
Sensors 25 02529 g008aSensors 25 02529 g008b
Figure 9. Scenes from EuRoC dataset. (a) The left is from V101. (b) The right is from V203.
Figure 9. Scenes from EuRoC dataset. (a) The left is from V101. (b) The right is from V203.
Sensors 25 02529 g009
Figure 10. Reprojection error distribution of global keyframes across sequences 00–10 in the KITTI dataset.
Figure 10. Reprojection error distribution of global keyframes across sequences 00–10 in the KITTI dataset.
Sensors 25 02529 g010
Figure 11. Reprojection error distribution plots of localized keyframes at three different timesteps for sequences 00–10 in the KITTI dataset. The three horizontal subplots per sequence represent one moment. (ak) correspond to sequences 00–10, respectively.
Figure 11. Reprojection error distribution plots of localized keyframes at three different timesteps for sequences 00–10 in the KITTI dataset. The three horizontal subplots per sequence represent one moment. (ak) correspond to sequences 00–10, respectively.
Sensors 25 02529 g011aSensors 25 02529 g011b
Figure 12. Scenes from EuRoC dataset. (a) The left is from V101. (b) The right is from V203.
Figure 12. Scenes from EuRoC dataset. (a) The left is from V101. (b) The right is from V203.
Sensors 25 02529 g012
Figure 13. Relationship between reprojection errors and robustness parameter τ using the DA-IRRK method on sequences V201 and V202. (a) The left is for V201. (b) The right is for V202.
Figure 13. Relationship between reprojection errors and robustness parameter τ using the DA-IRRK method on sequences V201 and V202. (a) The left is for V201. (b) The right is for V202.
Sensors 25 02529 g013
Figure 14. Relationship between reprojection errors and robustness parameter τ using the robust Huber method on sequences V201 and V202. (a) The left is for V201. (b) The right is for V202.
Figure 14. Relationship between reprojection errors and robustness parameter τ using the robust Huber method on sequences V201 and V202. (a) The left is for V201. (b) The right is for V202.
Sensors 25 02529 g014
Figure 15. Relationship between the reprojection error and the robustness parameter τ using the DA-IRRK method on sequences 06 and 07. (a) The left is for 06. (b) The right is for 07.
Figure 15. Relationship between the reprojection error and the robustness parameter τ using the DA-IRRK method on sequences 06 and 07. (a) The left is for 06. (b) The right is for 07.
Sensors 25 02529 g015
Figure 16. Relationship between the reprojection error and the robustness parameter τ using the robust Huber method on sequences 06 and 07. (a) The left is for 06. (b) The right is for 07.
Figure 16. Relationship between the reprojection error and the robustness parameter τ using the robust Huber method on sequences 06 and 07. (a) The left is for 06. (b) The right is for 07.
Sensors 25 02529 g016
Figure 17. Reprojection error distribution of sequences 06 and 07 for local mapping in the JORB-SLAM framework.
Figure 17. Reprojection error distribution of sequences 06 and 07 for local mapping in the JORB-SLAM framework.
Sensors 25 02529 g017
Figure 18. Scene and trajectory on sequences 06 and 07 in the JORB-SLAM framework. (a,b) Sequence 06 trajectory, (c,d) Sequence 07 trajectory, (e) Sequence 06 scene, (f) Sequence 07 scene.
Figure 18. Scene and trajectory on sequences 06 and 07 in the JORB-SLAM framework. (a,b) Sequence 06 trajectory, (c,d) Sequence 07 trajectory, (e) Sequence 06 scene, (f) Sequence 07 scene.
Sensors 25 02529 g018
Table 1. RMSE ATE (m) on Stereo EuRoC in ORB-SLAM3.
Table 1. RMSE ATE (m) on Stereo EuRoC in ORB-SLAM3.
MH01MH02MH03MH04MH05V101V102V103V201V202V203
Ours0.03630.03080.04050.06010.04610.07880.06030.07380.07480.05570.5490
Huber0.03910.03330.04110.06390.05290.08260.06370.07490.07290.05820.4940
MCMCC0.03800.03320.03730.06600.05180.08260.06410.07290.05280.05650.4150
Cauchy0.03950.03320.04420.08140.05340.08090.06720.11600.07440.06830.8560
Tukey0.03910.03450.04330.08920.05180.08150.06470.11400.08320.08590.8410
Table 2. RMSE ATE (m) on Stereo KITTI in ORB-SLAM3.
Table 2. RMSE ATE (m) on Stereo KITTI in ORB-SLAM3.
0001020304050607080910
Ours0.94455.03165.20600.28060.17940.35910.64490.42072.77051.77370.9458
Huber0.94545.89846.17950.29620.19900.39200.63570.48332.85971.71171.0910
MCMCC0.95675.68015.57420.23090.19780.37080.63090.42242.97521.79460.9226
Cauchy1.04489.17115.80030.29450.18330.39580.75840.79233.41131.77041.0993
Tukey0.99327.91787.58240.39730.29410.45720.58050.50242.88511.86021.6109
Table 3. Multi-sensor combination (visual-inertial): RMSE ATE (m) for Stereo-mode testing using ORB-SLAM3 on EuRoC dataset (indoor).
Table 3. Multi-sensor combination (visual-inertial): RMSE ATE (m) for Stereo-mode testing using ORB-SLAM3 on EuRoC dataset (indoor).
MH01MH02MH03MH04MH05V101V102V103V201V202V203
Ours0.01900.01540.02280.04080.05350.03260.00750.01610.01470.01020.0164
Huber0.02310.02090.02580.04540.05600.03380.00870.01710.01350.01190.0153
Impr_huber18.08%26.29%11.92%10.02%4.64%3.64%13.53%6.13%−9.44%14.68%−6.60%
Table 4. RMSE ATE (m) on Stereo-KITTI in JORB-SLAM.
Table 4. RMSE ATE (m) on Stereo-KITTI in JORB-SLAM.
0001020304050607080910
OursA0.18450.11280.05900.06640.06480.05640.01840.07851.07850.04490.0397
B0.10230.38910.08490.04410.06620.05120.02880.03580.48290.06430.0397
HuberA0.18810.15910.08280.06570.06140.06250.02260.07541.07930.05190.0442
B0.10930.48290.09730.05240.06450.05370.03350.03220.48220.06770.0558
Impr_huberA1.91%29.10%28.73%−0.94%−5.49%9.83%18.45%−4.12%0.07%13.52%10.23%
B6.40%19.42%12.82%15.84%−2.64%4.65%13.87%−11.12%−0.15%5.02%28.88%
Table 5. RMSE ATE (m) on Mono-EuRoC in CCM-SLAM.
Table 5. RMSE ATE (m) on Mono-EuRoC in CCM-SLAM.
MH01-MH02MH02-MH03MH03-MH04MH04-MH05V101-V102V102-V103V201-V202V202-V203
OursMH01 0.1296MH02 0.1202MH03 0.1506MH04 0.1710V101 0.0093V102 0.1046V201 0.0290-
MH02 0.1066MH03 0.1077MH04 0.1222MH05 0.1626V102 0.0347V103 0.1025V202 0.03576-
HuberMH01 0.3069MH02 0.1862MH03 0.2049MH04 1.0253V101 0.1261V102 0.1525V201 0.0860-
MH02 0.2508MH03 0.1397MH04 0.1186MH05 0.1620V102 0.1464V103 0.2125V202 0.0988-
Impr_huberMH01 57.77%MH02 35.45%MH03 26.50%MH04 83.32%V101 92.62%V102 31.41%V201 66.28%-
MH02 57.50%MH03 22.91%MH04 −3.04%MH05 −0.37%V102 76.30%V103 51.76%V202 59.21%-
Table 6. Computational elapsed time of the main modules on V103 of EuRoC in ORB-SLAM3 framework (meantime and standard deviation in ms).
Table 6. Computational elapsed time of the main modules on V103 of EuRoC in ORB-SLAM3 framework (meantime and standard deviation in ms).
AlHuberDA-IRRK (Ours)HuberDA-IRRK (Ours)
SettingsSensorStereoStereoMono-InertialMono-Inertial
Cam.Resolution752 × 480752 × 480600 × 350600 × 350
Cam.fps20 Hz20 Hz20 Hz20 Hz
IMU--200 Hz200 Hz
ORB Feat.1200120010001000
RMSE ATE0.082690.069680.037290.01611
TrackingORB extract14.56615.2596.1375.941
Stereo match2.77922.7234--
IMU integr.--0.04780.04515
Pose pred1.6332.2430.0510.044
Local Mapping (LM) Track2.9183.7034.4914.399
New Key Frame (KF) dec0.07150.07620.050.0463
Total24.41626.583--
Local Mapping (LM)KF Insert3.8794.0646.3376.317
Map Point (MP) Culling0.2540.2910.07430.0773
MP Creation10.05110.26920.79720.591
Local BA (LBA)38.44460.00763.5162.972
KF Culling3.4893.37917.83416.637
Total55.65577.275107.93107.26
LBA complexityLBA Edges9087.792031628.31727.6
LBA KF optimized21.06720.5845.9235.833
LBA KF fixed37.36635.17411
Map Size KFs165160341346
Map Size MPs797983481157211308
Full BAGlobal BA (GBA)130.73230.62775.59686.97
Map Update16.7934.178657.34540.312
Total147.52234.8832.94727.28
BA Size KFs7373139126
BA Size MPs5020510863305406
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hu, Z.; Cheng, L.; Wei, J.; Xu, X.; Zhang, Z.; Yan, G. DA-IRRK: Data-Adaptive Iteratively Reweighted Robust Kernel-Based Approach for Back-End Optimization in Visual SLAM. Sensors 2025, 25, 2529. https://doi.org/10.3390/s25082529

AMA Style

Hu Z, Cheng L, Wei J, Xu X, Zhang Z, Yan G. DA-IRRK: Data-Adaptive Iteratively Reweighted Robust Kernel-Based Approach for Back-End Optimization in Visual SLAM. Sensors. 2025; 25(8):2529. https://doi.org/10.3390/s25082529

Chicago/Turabian Style

Hu, Zhimin, Lan Cheng, Jiangxia Wei, Xinying Xu, Zhe Zhang, and Gaowei Yan. 2025. "DA-IRRK: Data-Adaptive Iteratively Reweighted Robust Kernel-Based Approach for Back-End Optimization in Visual SLAM" Sensors 25, no. 8: 2529. https://doi.org/10.3390/s25082529

APA Style

Hu, Z., Cheng, L., Wei, J., Xu, X., Zhang, Z., & Yan, G. (2025). DA-IRRK: Data-Adaptive Iteratively Reweighted Robust Kernel-Based Approach for Back-End Optimization in Visual SLAM. Sensors, 25(8), 2529. https://doi.org/10.3390/s25082529

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop