Next Article in Journal
An Efficient Boundary-Type Meshless Computational Approach for the Axial Compression on the Part Boundary of the Circular Shaft (Brazilian Test)
Next Article in Special Issue
UGAVs-MDVR: A Cluster-Based Multicast Routing Protocol for Unmanned Ground and Aerial Vehicles Communication in VANET
Previous Article in Journal
Study of Social Presence While Interacting in Metaverse with an Augmented Avatar during Autonomous Driving
Previous Article in Special Issue
Research on Data Fusion of Positioning System with a Fault Detection Mechanism for Autonomous Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Application and Assessment of Cooperative Localization in Three-Dimensional Vehicle Networks

by
Juan Carlos Oliveros
and
Hashem Ashrafiuon
*,†
Center for Nonlinear Dynamics and Control, Department of Mechanical Engineering, Villanova University, 800 Lancaster Ave, Villanova, PA 19085, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Appl. Sci. 2022, 12(22), 11805; https://doi.org/10.3390/app122211805
Submission received: 27 October 2022 / Revised: 16 November 2022 / Accepted: 16 November 2022 / Published: 20 November 2022

Abstract

:
The trajectory planning and control of multi-agent systems requires accurate localization, which may not be possible when GPS signals and fixed features required for SLAM are not available. Cooperative Localization (CL) in multi-agent systems offers a short-term solution that may significantly improve vehicle pose estimation. CL algorithms have been mainly developed and assessed for planar mobile robot networks due to complexities and singularities in three-dimensional (3D) motion. In this paper, we develop the required singularity-free equations and apply and assess an EKF-based CL for 3D vehicle networks. We assess the performance of CL with respect to the number of simultaneous and redundant measurements. We further assess CL performance with only relative position measurements available. Finally, experiments are performed to validate the proposed algorithms. We further investigate the effect of absolute position measurements in CL. Conclusions: Cooperative localization is an effective method when applied to 3D vehicle networks. However, CL does not improve localization with only relative position measurements, and thus previously reported results for 2D vehicle models were only effective due to relative orientation measurements. Absolute measurement reduces the overall localization errors much more significantly when there has been CL with prior relative position measurements.

1. Introduction

One of the vital technologies required for autonomous navigation is a localization technique that is capable of recognizing the position and orientation of a vehicle or vehicle networks, especially in GPS-denied environments. This problem is further exacerbated by environments that are often unstructured and lack distinct features (such as deserts), contain features that may change over time due to environment, or as the vehicles move to new areas (such as urban settings). Methods such as feature-based Simultaneous Localization and Mapping (SLAM) [1] are not capable of providing accurate localization in these environments. Collaborative systems of unmanned ground and air vehicles show great promise for various applications [2,3]. Cooperative Localization (CL) can be utilized for these systems to significantly reduce the rate of growth of estimation errors until GPS becomes available again. In addition, CL can help reduce the sample rate requirement for various SLAM methods in structured environments, since these methods require extensive computation. Cooperative SLAM techniques [4,5] have shown great promise for precise localization, especially when combined with other techniques such as the Extended Kalman Filter (EKF) [6]. However, the success of these techniques relies on the knowledge of the environment.
CL can utilize several techniques, including Kalman filters [7,8,9] and particle filters [10,11]. Utilizing these techniques, CL has been suggested as a powerful method to improve pose estimation in vehicle networks capable of inter-agent communication for data sharing [12]. In an early application of CL, an EKF algorithm [13] was developed that distributes vehicle state information while keeping track of the cross-covariance terms. These terms are created by relative measurements of each 2D vehicle’s pose. However, this method requires a completely synchronous and interconnected communication network. A decoupled CL algorithm is proposed [14] that eliminates the need for complete and constant between vehicles. Vehicles are assumed to be equipped with the capability of obtaining their own pose (via equations of motion and onboard sensors), measuring the poses of nearby vehicles relative to their own reference frame, and processing this information and communicating with other vehicles. CL allows a group of vehicles to cooperatively approximate an individual vehicle’s pose by sharing the relative and absolute pose measurements that each vehicle has obtained.

2. Motivation

Research works on multi-agent systems have focused on 2D vehicle models, even treating vehicles as simple nodes [15]. CL methods have focused primarily on maneuvering in 2D space [16,17,18,19] and particularly using wheeled mobile robots [20,21,22] and vehicles [23] due to the complexities involved in the singularity-free representation of 3D orientation. It is, however, critically important to apply CL to three-dimensional (3D) vehicle models to improve the localization of networks with homogeneous and heterogeneous air, marine, and ground vehicles. Others only consider relative localization without any regard to the global reference frames [24,25]. There is also a CL method [26] that only uses range measurements of a vehicle with access to absolute measurements via GPS and therefore does not take advantage of cross-covariances terms. A common theme in these methods is suffering from singularities due to using Euler angles to represent orientation. EKF-based CL with relative pose measurements to improve the absolute pose of networks with 3D vehicles has remained relatively unexplored due to the intricacies involved in the kinematic representation and linearization of state and measurement equations. Alternative relative localization methods [27] have also been explored. Euler angles are normally used to measure orientation, but are associated with singularities in aggressive vehicle maneuvers with relative or absolute angles reaching 90 ° . Quaternions, on the other hand, do not have any singularities but are not measurable by any sensors. Hence, they are calculated from the rotation matrix, which may be challenging in the case of relative orientation representation.

3. Related Works

EKFs have commonly been used as the standard technique for state estimation in CL. There are EKF-based methods that achieve localization utilizing stereo-vision and sensor fusion [28] and can be further extended to target tracking [29]. However, these methods require maximizing the overlap of each agent’s field of view to enhance localization. Many localization techniques rely on a central system or single/multiple “leader” agents responsible for sensor fusion [30,31,32]. Distributed or decentralized methods have been investigated [33,34,35] to ameliorate the computational bottlenecks associated with centralized systems. While these methods can be efficient and effective, they come at the cost of accuracy. A combination of centralized–decentralized CL in ad hoc networks shows great promise in improving localization without prior knowledge of an environment [14,36]. A distributed CL utilizing an EKF with low communication path requirements can also provide centralized-equivalent performance [37]. It is also possible to model the CL problem as a graph model to further utilize additional algorithms [38].
EKF-based CL algorithms are popular due to their ease of execution and computational efficiency. However, linearizing an extremely nonlinear system can lead to inaccuracies and numerical instabilities. Unscented Kalman Filter (UKF)-based methods were introduced in [39] with the advantage of not requiring any linearization. The UKF is deemed to be more robust [40] since it does not rely on the computation of the Jacobian matrix and is an excellent option for self-localization and tracking [41,42,43,44]. On the other hand, the UKF has a much higher computation cost compared to the EKF, especially considering the Cholesky factorizations at every step. As an alternative, the Squared Root UKF [45] does help mitigate the computation cost. However, implementation of UKF-based CL algorithms and simultaneous measurements is very complex and has not been fully explored. Table 1 lists some of these related works along with their advantages and limitations.

4. Main Contributions and Novelties

In this paper, we extend a previous conference proceedings publication [46] to include additional subcases, physical experiments, and identifying additional limitations to CL. We extend the centralized EKF-based CL algorithm presented in [13,14] to include 3D vehicle models and assess its performance. Specifically, we utilize quaternions to represent vehicle orientation in order to avoid singularities. For measurements, we introduce quaternions corresponding to relative orientation represented by the rotation matrix and present linearized measurement equations, i.e., the output matrix. We then assess the performance of CL in 3D vehicle networks through a series of Monte Carlo-type simulations. We use a sequential update procedure [47] to assess the effect of simultaneous and redundant measurements on individual vehicles, as well as the overall network. Assuming that absolute orientation can be accurately measured by onboard sensors, we further assess the performance of CL when only relative position measurements are available and investigate the circumstances under which the method could be effective or ineffective. Finally, experiments are preformed to verify the algorithm.
The main contributions and novelties of this publication are outlined as follows:
  • The development and application of singularity-free EKF-based CL for networks of three-dimensional vehicles;
  • The first comprehensive assessment of EKF-based CL with only relative position measurements;
  • The identification of limitations of CL with respect to the number of simultaneous measurements, redundant measurements, and relative position-only measurements;
  • The experimental verification of the EKF CL algorithm’s performance and limitations.
The paper is organized as follows. The vehicle model is presented in Section 5. A multi-vehicle system with state and measurement equations and their linearization are presented in Section 6, followed by the CL algorithm in Section 7. Simulations and experiments are presented in Section 8 and Section 9, respectively, followed by the concluding remarks in Section 10.

5. Vehicle Model

5.1. Vehicle Pose

Each vehicle is assigned a body-fixed reference frame X v Y v Z v . Figure 1 depicts the pose of the vehicle, which is then represented by the position of the origin of X v Y v Z v and its orientation with respect to the inertial reference frame X 0 Y 0 Z 0 . Hence, the vector r 0 v = [ x 0 v y 0 v z 0 v ] represents the position of the vehicle, and the orthogonal rotation matrix R 0 v that transforms the coordinates of any vector in X v Y v Z v to X 0 Y 0 Z 0 determines its orientation. R 0 v can be calculated using quaternions, which are favored over Euler angles since they do not suffer from singularities and require less computational overhead. The quaternion vector of the vehicle reference frame q ¯ v can be presented as:
q ¯ v = q 0 v q v , q v = q 1 v q 2 v q 3 v
where there is redundancy due to the four scalar values representing three rotational degrees of freedom. The constraint relating the quaternions is q ¯ v q ¯ v 1 = 0 , indicating that q ¯ v is a unit vector.
The orthogonal rotation matrix is given in terms of the quaternions as [48]:
R 0 v ( q ¯ v ) = [ 2 ( q 0 v ) 2 1 ] I 3 + 2 [ q v q v + q 0 v S ( q v ) ]
where S ( q v ) defines a skew-symmetric matrix of a vector such that
S ( q v ) = 0 q 3 v q 2 v q 3 v 0 q 1 v q 2 v q 1 v 0
and I 3 is the 3 × 3 identity matrix. Given the rotation matrix in (2), we can calculate q ¯ v , as described in [46].

5.2. Vehicle Velocity

In the global coordinate frame, X 0 Y 0 Z 0 , the vehicle translational velocity vector, r ˙ 0 v , is related to its body-fixed velocities in X v Y v Z v , r ˙ v v , as [49]:
r ˙ 0 v = R 0 v ( q ¯ v ) v v
where v v r ˙ v v . The angular velocity vector of a vehicle about its body-fixed reference frame, ω v = [ ω x v ω y v ω z v ] , is used to calculate the time derivative of the quaternion vector, q ¯ ˙ v , as:
q ¯ ˙ v = 1 2 G v ω v , ω v = 2 G v q ¯ ˙ v
G v = q v , q 0 v I 3 S ( q v )
The unit-length constraint equation that relates the time derivatives of the quaternions is also given as q ¯ v q ¯ ˙ v = 0 .

5.3. Vehicle State Equations

Considering the single vehicle shown in Figure 1 and as described in the previous section, the pose of the vehicle can be represented by r 0 v = [ x 0 v y 0 v z 0 v ] and q ¯ v . For a kinematic model, we may choose the state vector as a collection of the seven pose variables:
x v = r 0 v q ¯ v
Using (4) and (5), the driftless state equations may be written as:
x ˙ v = g v ( x v ) u v
where
g v = R 0 v 0 3 × 4 0 3 × 3 1 2 G v T , u v = v v ω v
where 0 n × m represents the n × m zero matrix and the vehicle input vector consists of the body-fixed velocities v v and ω v .
The 7 × 7 Jacobian matrix of state equations, J v , necessary for EKF linearization is calculated as:
J v = g v ( x v ) u v x v = 0 3 × 3 r ˙ 0 v q ¯ v 0 4 × 3 q ¯ ˙ v q ¯ v
where the nonzero components are:
r ˙ 0 v q ¯ v = r ˙ 0 v q 0 v r ˙ 0 v q v
q ¯ ˙ v q ¯ v = 1 2 0 ω v ω v S ( ω v )
and
r ˙ 0 v q 0 v = [ ( 4 q 0 v ) I 3 + 2 S ( q v ) ] v v
r ˙ 0 v q v = 2 q v v v + 2 ( q v v v ) I 3 2 q 0 v S ( v v )
It is important to note that several approaches have been proposed to adhere to the unit quaternion constraint during the integration of the state equations [50,51,52].

6. Multi-Vehicle System Model

6.1. State Equations

In order to formulate the CL algorithm, let us first consider a three-vehicle system shown in Figure 2. In the illustrated scenario, a vehicle designated as b or “base vehicle” obtains a pose measurement of another vehicle designated as t or “target vehicle” relative to its own body-fixed reference frame. Referring to (6)–(8), the system may be represented by the 14 states representing absolute poses of the two vehicles. The state equations are written as:
x ˙ = g ( x ) u x = r 0 b q ¯ b r 0 t q ¯ t u = v b ω b v t ω t g = diag R 0 b 1 2 G b R 0 t 1 2 G t
The 14 × 14 Jacobian matrix, J, for the two-vehicle network is simply defined as:
J = diag J b J t
where J b and J t are defined for vehicle b and vehicle t according to (9)–(11).

6.2. Measurement Equations

Consider a base vehicle b acquiring a relative pose measurement of a target vehicle t. A problem with quaternions is that they are not kinematic quantities that can be directly measured by any sensor. Thus, we assume the rotation matrix representing the relative orientation of t with respect to b, perhaps derived from Euler angles, is the measured quantity. We then derive the quaternions corresponding to the relative rotation matrix. Similarly, the relative range, bearing angle, and elevation angle of t with respect to b are the quantities normally measured that may be used to determine the relative position. Here, we define the measurement output vector, z , in terms of the relative position vector, r, and the three components of the quaternion vector q, which are determined from the relative orientation matrix:
z = r q .
Note that the q 0 component is not needed since it can be derived from q using the unit-length constraint equation. Using the system states, the relative position r can be written as:
r = R 0 b r 0 b t = R 0 b ( r 0 r 0 b ) = R 0 b r 0
where r 0 = r 0 r 0 b . Quaternions representing the relative orientation are derived from the relative rotation matrix:
R ( q ¯ ) = R 0 b ( q ¯ b ) R 0 t ( q ¯ t )
Vector q is then derived from the elements of R ( q ¯ ) [46].
Next, we linearize the measurement output equations, z = H x where:
H = H b H t = R 0 b r q ¯ b R 0 b 0 3 × 4 0 3 × 3 q q ¯ b 0 3 × 3 q q ¯ t .
The term r q ¯ b can be written in terms of its components as r q ¯ b = r q 0 b r q b . These components can then be defined similar to (11) as:
r q 0 b = [ ( 4 q 0 b ) I 3 + 2 S ( q b ) ] r 0
r q b = 2 q b r 0 + 2 ( q b r 0 ) I 3 2 q 0 b S ( r 0 ) .
Assuming q 0 0 , the remaining derivative terms can be shown as:
q q ¯ b = 1 4 q 0 p q ¯ b , q q ¯ t = 1 4 q 0 p q ¯ t
where p is represented in terms of the elements of the matrix R ( q ¯ )
p = R 32 R 23 R 13 R 31 R 21 R 12 ,
and the components of (19) are derived as:
p q 0 b = 16 q 0 b q 0 t q t + 2 [ R 0 t t r ( R 0 t ) I 3 ] q b p q 0 t = 16 q 0 b q 0 t q b 2 [ R 0 b t r ( R 0 b ) I 3 ] q t p q b = 2 q 0 b [ R 0 t t r ( R 0 t ) I 3 ] 2 S ( q b ) R 0 t + 2 S ( R 0 t q b ) p q t = 2 q 0 t [ R 0 b t r ( R 0 b ) I 3 ] + 2 S ( q t ) R 0 b 2 S ( R 0 b q t )
If q 0 = 0 , components of q q b and q q t are derived as:
q i q j b = 1 4 q i R i i q j b , q i 0 q i q j t = 1 4 q i R i i q j t , q i 0 q i q j b = q i q j t = 0 , q i = 0 , i , j = 1 , 2 , 3
Finally, the terms R i i q b and R i i q t can be derived as:
R i i q j b = 2 R 0 j i t q i b , i j R i i q j b = 2 R 0 j i t q i b + 2 j = 1 3 R 0 j i t q j b , i = j R i i q j t = 2 R b j i q i t , i j R i i q j t = 2 R 0 j i b q i t + 2 j = 1 3 R 0 j i b q j t , i = j

7. Cooperative Localization

Consider a network of n vehicles numbered 1 to n. A vehicle i, at sample time ( k ) , utilizes absolute measurements to propagate its pose using the discrete form of (7):
x i ( k + 1 ) = g i ( x i ( k ) ) u i ( k ) + γ i ( x i ( k ) ) η i ( k )
where x i , u i , and η i refer to the state, input, and process noise vectors, respectively, and g i ( x i ) and γ i ( x i ) are the input and process noise coefficient functions, respectively. Next, we assume that vehicle i in the set { 1 , , n } measures the relative pose of vehicle, where j i . Subsequently, the measurement vector z i j can be written as:
z i j ( k + 1 ) = h i j ( x i ( k ) , x j ( k ) ) + ν i ( k )
where ν i ( k ) represents the measurement noise and h i j = [ r i j q i j ] ; see (14).
For the EKF procedure, we denote variables at the propagation step with “−” and the updated step with “+”. We also use over “ ˆ ” for estimated state values. Thus, x ^ i and x ^ i + indicate the predicted and updated state estimates for vehicle i, respectively. The key to success of EKF-based CL is the cross-covariance [14], which is the correlation term between vehicles i and j due to the relative measurements. Given initial conditions x i ( 0 ) , P i j ( 0 ) = 0 7 × 7 , and the initial guess of P i i ( 0 ) = I 7 , the prediction step state estimates, covariance matrix, and cross-covariance matrix of vehicles i = 1 , , n at step k + 1 are given in terms of step k terms as:
x ^ i ( k + 1 ) = g i ( x i + ( k ) ) u i P i i ( k + 1 ) = J i P i i + J i + Γ i Q i Γ i P i j ( k + 1 ) = J i P i i + J j
where j = 1 , , n , j i , J i = g i x i ( k ) is calculated using (9)–(10), Γ i = γ i x i , and Q i > 0 is the covariance of process noise.
At the update step, the state vector and (cross) covariance matrices remain unchanged if there are no relative measurements. Let us now assume that a “base” vehicle b takes relative measurements of a “target” vehicle t. Then, the innovation error between the measurement z b t and the estimated output through propagated states is:
e b ( k + 1 ) = z b t h b t ( x ^ b ( k + 1 ) , x ^ t ( k + 1 ) ) z b t H b H t x ^ b ( k + 1 ) x ^ t ( k + 1 )
where [ H b H t ] is defined in (17). The cross-covariance terms modifies the innovation covariance matrix as follows:
S b t ( k + 1 ) = H b P b b H b + H t P t t H t + H b P b t H t + H t P t b H b + R b
Utilizing (27) and (28), the Kalman gains along with the states and covariance terms for all vehicles i = 1 , , n are updated as follows:
K i ( k + 1 ) = [ P i t H t + P i b H b ] S b t 1 x ^ i + ( k + 1 ) = x ^ i + K i e b P i i + ( k + 1 ) = P i i K i S b t K i P i j + ( k + 1 ) = P i j K i S b t K j
where j = 1 , , n , j i . The Kalman gain, due to its semi-definite nature, ensures that a decrease in localization error is achieved by any relative measurement, as long as the measurements are not too noisy and the EKF linearization is accurate. The step-by-step procedure for the centralized CL algorithm can be found in [14].
The structure of the EKF-based CL algorithm is illustrated as a diagram in Figure 3 for better understanding of the procedure.

8. Simulation Results

Given a propagation time step size of Δ t , the seven state equations of vehicle i in (4) and (5) can be written in discrete form as:
r 0 i ( k + 1 ) = r 0 i ( k ) + R 0 i ( k ) v i ( k ) + η v i Δ t q ¯ i ( k + 1 ) = q ¯ i ( k ) + 1 2 G i ( k ) ω i ( k ) + η ω i Δ t
where η v i and η ω i are the corresponding components of η i in (24) representing the input velocity process noise. Similarly, R 0 i ( q ¯ i ( k ) ) and 1 2 G i ( q ¯ i ( k ) ) are the components of γ i in (24).
The relative measurements obtained by all vehicles are polluted by additive Gaussian measurement noise (AWGN) with a standard deviation of 0.05 m for the position states (based on distances between vehicles) and 0.02 for the quaternion states of each vehicle. Vehicles may also be capable of obtaining absolute pose measurements with a standard deviation of 0.1 m for position and 0.02 for the quaternion states. As far as process noise is concerned, the standard deviations are 0.1 v i for η v i and 1 ° / s for η ω i across all vehicles. Finally, since there is additive noise in both measurement and process steps, after several trial runs, we concluded that about 100 runs in each simulation scenario are required to remove any significant effect stemming from outliers.
All simulations are run for 200 s and use the same input velocities and process and measurement noise properties in order to remove any input bias from the assessments. Input velocities are applied to each vehicle with varying magnitudes, as listed in Table 2. Since vehicles normally have a “forward” velocity input, 3D movement is attained through orientation changes, as a result of input angular velocities. Therefore, linear velocity is only applied in the x-direction. We selected Δ t = 0.02 s, which is sufficiently small for accurate propagation without any noise, given the magnitude of input velocities.
To provide a comprehensive assessment, we initially present a test case with a changing measurement scenario. Next, we evaluate the effect of sequentially adding measurements that do not directly involve a given vehicle to show how CL can improve its localization and the localization of the entire network. We then assess the effect of redundant measurements when two or more vehicles simultaneously measure each other’s poses. Next, we assess the effectiveness of CL when only relative positions are measured, followed by an investigation of how a single vehicle’s absolute measurement can affect localization in this scenario.

8.1. A Test Case

We initially simulate a test case to demonstrate the effectiveness of CL using the first four vehicles listed in Table 2. In the test case, there are no relative measurements for the first 40 s, and therefore, each state is estimated using propagation only. Then, some vehicles fall within the line of sight of other vehicles such that vehicle 1 measures the relative pose of vehicle 2, vehicle 2 measures vehicle 3, and vehicle 3 measures vehicle 4 from t = 40 s until t = 160 s. These measurements all happen simultaneously. For the final 40 s, all vehicles lose sight of each other, such that there are no relative measurements. However, vehicles 1 and 4 take absolute measurements of their own pose until t = 200 s possibly due to the availability of GPS or SLAM for those vehicles.
The actual paths of the four vehicles along with those estimated with and without CL are shown in Figure 4 where significant improvement in the estimation of the path is observed after 40 s when CL is applied. We denote the norm of error in position vector r 0 v as Δ r and the norm of error in orientation vector q v as Δ q . Figure 5 and Figure 6 compares the time history of Δ r and Δ q for the four vehicles with and without CL, along with their 2 σ plots. It is clear that there is a significant improvement in the estimation of the pose of all vehicles during CL (from 40 s to 160 s), including vehicle 1, whose relative pose is never measured. In addition, the absolute measurements of vehicles 1 and 4 not only greatly reduce localization errors for those vehicles but also for vehicles 2 and 3 due to their cross-covariance matrices with vehicles 1 and 4. Another observation is that there are discontinuities in state estimations when measurement scenarios are changed, caused by abrupt changes in the covariance matrix, as observed in the 2 σ plots in Figure 5 and Figure 6. These discontinuities, however, are not of any concern as long as the algorithm remains numerically stable.

8.2. Simultaneous Relative Pose Measurement Assessment

For CL, the number of relative measurements plays a significant role in lowering uncertainty. Therefore, we investigate how the error for a given vehicle (and across all vehicles) is affected by sequentially adding simultaneous measurements by other vehicles. We also assess what happens when vehicles make redundant measurements, such as two vehicles simultaneously taking relative measurements and exchanging information with each other. Hence, we consider two networks, one with all six vehicles, n = 6 , for assessment of additional simultaneous measurements (Scenario 1), and another with only the first three vehicles, n = 3 , for the assessment of redundant measurements (Scenario 2). In each scenario, we start with one vehicle taking the relative measurements of another vehicle. In subsequent cases, we continue to add more relative measurements to assess their contribution.

8.2.1. Scenario 1—Effect of Other Vehicle Measurements on Single-Vehicle Localization

The idea for Scenario 1 is to see how other vehicles taking additional simultaneous relative measurements affects the localization of a single vehicle and, consequently, all vehicles in the network. Table 3 shows five cases that were simulated, concentrating on the localization of vehicle 1. In case 1, relative measurements of vehicle 2 are taken by vehicle 1, while in case 2, vehicle 2 simultaneously takes relative measurements of vehicle 3 and so on. In each case, errors are averaged across the 100 simulations.
The mean of the position error norm ( Δ r ) and the orientation error norm ( Δ q ) for vehicle 1 in all 100 simulations are plotted in Figure 7 and Figure 8, respectively, for propagation without measurements and the five CL cases. Both Δ r and Δ q show that the performance increases as more relative measurements are added, and all CL cases show significant improvement over the propagation. An interesting observation is the significant error reduction in both Δ r and Δ q of vehicle 1 from Case 1 to Case 2 when vehicle 2 measures the relative pose of vehicle 3. Hence, the overall error for vehicle 1 is reduced even though it is not involved in the additional measurement. When subsequent measurements are added in cases 3 through 5, the localization errors continue to decrease but are relatively marginal. This demonstrates that the beneficial effects of CL may not outweigh its additional computational cost as the number of simultaneous measurements is increased. It should also be noted that the temporary dip in quaternion propagation error in Figure 8 is due to the complete rotation of the vehicles, bringing some components of the quaternion closer to their actual values.
Table 4 lists the error norm statistics across the 100 simulations for vehicle 1, which includes statistical terms such as mean and standard deviation for position r and orientation q. To calculate these terms, we determine the norms of Δ r and Δ q at each time step of each simulation, and their root means squared are derived over time. Finally, the mean, standard deviation, min, and max are derived across all simulations. In CL, the cross-covariances contribute to lower overall error for vehicles when additional measurements are performed. The statistics again illustrate that the error reduction beyond Case 2 is relatively marginal. However, it also demonstrates that the variability is significantly increased in CL (see the standard deviation of CL cases compared to propagation).
We also compared the same statistics across all vehicles. In this case, Δ r and Δ q are again calculated at each time step of each simulation, and the root mean squared of the means is derived across all time steps and then averaged for all vehicles. Finally, the mean, standard deviation, min, and max are derived across all simulations. The results are listed in Table 5. The data show a similar trend to Table 4, where there is a reduction in the error estimates of both Δ r and Δ q as simultaneous measurements are added. However, as expected, the reduction in overall localization errors continues to be significant as the number of vehicles that rely solely on propagation decreases. The same increase in variability is also observed in the location of all vehicles.
To demonstrate the importance of cross-covariances, we repeated the simulations in case 5 but without maintaining the cross-covariance updates as presented in [14]. The error norm statistics in the 100 simulations for vehicle 1 and all vehicles are listed as the last rows of Table 4 and Table 5. It is clear that our method performs better for the overall network and the localization of individual vehicles. In addition, we plotted the time history of the error norms Δ r and Δ q for this method in Figure 7 and Figure 8, respectively. It is observed that while initially the method in [14] performs better, the errors grow at much faster rate.

8.2.2. Scenario 2—Effect of Redundant Pose Measurements

Scenario 2 is added in order to assess simultaneous redundant pose measurements and communications. In this scenario, we only consider the first three vehicles, n = 3 , while presenting the localization errors across all vehicles. We again assume vehicle 1 measures the relative pose of vehicle 2 as Case 1 and then add measurements until all vehicles simultaneously measure the relative pose of all other vehicles in the network, as listed in Table 6.
Table 7 summarizes the mean, standard deviation, minimum, and maximum error statistics for position and orientation estimates across all vehicle. As expected, the initial errors are reduced as measurements are added. However, there is no improvement beyond Case 3. In fact, there seems to be an increase in localization errors beyond Case 4 attributed mainly to computational inaccuracies. Therefore, redundant measurements (i.e., two vehicles simultaneous measuring each other’s relative poses and communicating their estimations and statistical properties) can negatively affect CL.

8.3. Relative Position Only Measurement Assessment

In this section, we assume each vehicle can reliably obtain its own orientation using an IMU and/or a gyroscope and does not measure the relative orientation of other vehicles. Therefore, the only means of CL is through relative position measurements. We use Scenario 2 for the simulations, but only relative position measurements are considered. Table 8 lists the error statistics for only Δ r across the 100 simulations. Our first observation is that the inclusion of absolute orientation measurements drastically reduces the overall error across vehicles, since IMUs can provide relatively accurate orientation information. Our second observation is that CL does not achieve statistically significant improvement in any of the cases. Therefore, the additional computational cost of CL does not justify its use in conjunction with the kinematic model presented in this and other works. There were no advantages in applying CL to this set because the equations of motion (4) and (5) and consequently the state equations (7) and (30) do not depend on position variables. We can further conclude that the improvements observed in Section 8.2 are only due to relative orientation measurements. These conclusions have not been previously reported and are valid even when applying CL to 2D vehicle models employed by other researchers [13,14].

8.4. Effect of Absolute Position Measurements in CL

It has previously been reported that measuring the absolute position of one vehicle can reduce the localization errors of other vehicles in CL [14]. Therefore, CL with only relative position measurements may still be useful if a vehicle is able to attain the absolute position measurement via GPS, SLAM, etc. The three-vehicle system of Scenario 2 was utilized to illustrate the effect of absolute measurements. We consider two cases. In one case, the relative positions of vehicles 2 and 3 are measured by vehicles 1 and 2, respectively, for the first 100 s (CL). Afterwards, vehicle 1 measures its own absolute position for a short period of 1 s in addition to the relative measurements. Afterwards, there are no subsequent measurements by any of the vehicles. In the other case, there are no measurements for the first 100 s (no CL), after which vehicle 1 measures its own absolute position. There are no other measurements involved.
The mean error norm Δ r across all vehicles and simulations was calculated to be 0.85 ± 0.04 for the CL case and 1.05 ± 0.01 for the no CL case. This clearly indicates that the absolute measurement reduces the overall localization errors much more significantly when there has been CL with prior relative position measurements.
Additional details can be found in Figure 9 where the comparisons between the time histories of the two cases for the three vehicles are plotted. It is clear that CL does not make a significant contribution during the first 100 s. After the first 100 s vehicle 1 measures its own position and its localization error drops to near zero in both cases. However, despite the lack of relative position measurements from 100 s to 200 s, the localization errors for the “CL” case are significantly lower than the “no CL” case for vehicles 2 and 3. This is due to the cross-covariance evolution in the CL algorithm during the first 100 s of the CL case. In conclusion, we propose that when dependable absolute orientation measurements are accessible, CL with relative position measurement does not improve localization errors using the kinematic model. Nevertheless, if any vehicle can measure its absolute position at some point during the experiment, CL can still greatly reduce the position errors of all vehicles.

9. Experimental Results

We conduct two experiments with three vehicles to verify and evaluate the CL algorithm. The first experiment utilizes relative pose measurements, while the second one only takes relative position measurements. The duration for both experiments is 80 s, and each vehicle follows a predefined path. The actual path data are processed through the CL EKF algorithm as measurements obtained by a vehicle. Each vehicle propagates its own states at a rate of Δ t = 0.02 s. However, relative measurements are obtained every 0.2 s.
Figure 10 shows an overview of one of the vehicles. All three vehicles have the same generic configuration and possess similar capabilities. The vehicles are differential-drive mobile robots with two motors controlled via a dual motor control and a PI controller implemented on an Arduino Duo board using motor encoder measurements. A Raspberry Pi computer performs most of the computation required to navigate the robot autonomously. Relative positions are calculated via LiDAR measurements of distance and bearing angle. However, the relative orientations are measured using a motion capture system that tracks markers, which are otherwise unused. The 0.2 s measurement rate is the time required for the LiDAR to rotate fully and the required data processing.

9.1. Relative Pose Measurement Experiment

In these experiments, we investigated the effect of CL when relative pose measurements are only available for a very short period. For the first case (no CL), each vehicle propagates its own states for 40 s. Starting at 40 s, vehicle 3 measures its own pose for 1 s, after which there are no other measurements. In the second case (CL), the same scenario is repeated except that between 40 s and 41 s, vehicle 3 simultaneously measures the relative pose of vehicle 2 and 1 in addition to its own pose.
Figure 11 compares the time history of the position error norms of the two cases in this experiment for all three vehicles. As expected the performances for all three vehicles are identical for the first 40 s and vehicle 3’s localization is similar throughout. However, despite only 1 s of relative measurements (i.e., five samples), vehicle 1’s position estimation is dramatically improved. Vehicle 2, on the other hand, does not initially benefit from CL. However, as time passes, the localization errors start decreasing for CL while increasing for the “no CL” case.

9.2. Relative Position Measurement Experiment

Next, we performed experiments with only relative position measurements, similar to the simulations in Section 8.4, where the absolute orientation measurements are available throughout the experiments for all vehicles. In the first experiment (no CL), the three vehicles propagate their position throughout while vehicle 1 measures its absolute position at 40 s for only 1 s. The second experiment (CL) is similar to the first except that vehicles 1 and 2 measure the relative positions of vehicles 2 and 3, respectively, for the first 40 s. Note that vehicle simply propagate their positions for the last 39 s in both cases.
Figure 12 shows the comparisons between the time history of the position error norms of the two experiments for the three vehicles. For the first 40 s, it is evident that CL has a mixed performance, where vehicles 2 and 3 seem to benefit from it while vehicle 1 localization deteriorates. However, after 40 s when vehicle 1 measures its own position for only 1 s, there is a significant drop in estimation errors of vehicles 2 and 3 when compared to the “no CL” case. This experiment confirms that even with intermittent absolute orientation measurements, CL can be very effective in improving localization errors.

10. Conclusions

In this work, we implemented a centralized EKF-based CL algorithm to several networks of 3D vehicles. We build upon a previous publication [46] by adding additional cases, experimental results, and explicit limitations of the proposed CL algorithm. We derived state and relative measurement equations in terms of quaternions and also produced the singularity-free linearized matrices required for EKF. We utilize quaternions to avoid singularities linked to Euler angles. Simulation results demonstrated that CL considerably reduces state estimation errors when relative orientation measurements are available, though the performance does not proportionally improve with additional measurements. We further showed that CL does not provide any notable advantages when redundant measurements are applied. When utilizing CL with only relative position measurements, no improvement is observed, which is attributed to the absence of position variables in the state equations. However, CL with only position measurements can still provide significant improvement in state estimation if absolute measurements by a vehicle in the network becomes available any time during simulation. Furthermore, these conclusions are verified with experimental results that demonstrate the effect that absolute position and pose measurements have on CL. These experimental findings provide additional justification and credibility towards the CL algorithm presented in this paper.
The main limitation of the current algorithm is the use of a simplistic kinematic model that is not realistic for real-world applications. This limitation also leads to the lack of effective localization when only relative positions are employed. Future work involves updating the algorithm to utilize a fully nonlinear dynamic model and applying closed-loop control. There will also be closed-loop control experiments to further evaluate the effect CL has on a vehicle’s state estimation.

Author Contributions

All authors contributed to the study conception and design. Material preparation and data collection and analysis were performed by J.C.O. The first draft of the manuscript was written by J.C.O. and H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets generated during and/or analyzed during the current study are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CL Cooperative Localization
2D Two-dimensional
3D Three-dimensional
EKF Extended Kalman Filter
UKF Unscented Kalman Filter
GPS Global Positioning System
SLAM Simultaneous Localization And Mapping

References

  1. Gamini Dissanayake, M.W.M.; Newman, P.; Clark, S.; Durrant-Whyte, H.F.; Csorba, M. A solution to the simultaneous localization and map building (SLAM) problem. IEEE Trans. Robot. Autom. 2001, 17, 229–241. [Google Scholar] [CrossRef] [Green Version]
  2. Hager, C.; Zarzhitsky, D.; Kwon, H.; Pack, D. Cooperative target localization using heterogeneous unmanned ground and aerial vehicles. In Proceedings of the 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, Taipei, Taiwan, 18–22 October 2010; pp. 2952–2957. [Google Scholar] [CrossRef]
  3. Minaeian, S.; Liu, J.; Son, Y.J. Vision-Based Target Detection and Localization via a Team of Cooperative UAV and UGVs. IEEE Trans. Syst. Man Cybern. Syst. 2016, 46, 1005–1016. [Google Scholar] [CrossRef]
  4. Fabresse, F.R.; Caballero, F.; Maza, I.; Ollero, A. Robust Range-Only SLAM for Unmanned Aerial Systems. J. Intell. Robot. Syst. Theory Appl. 2016, 84, 297–310. [Google Scholar] [CrossRef]
  5. Waniek, N.; Biedermann, J.; Conradt, J. Cooperative SLAM on small mobile robots. In Proceedings of the 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO), Zhuhai, China, 6–9 December 2015; pp. 1810–1815. [Google Scholar] [CrossRef]
  6. Morrison, J.G.; Gávez-López, D.; Sibley, G. Scalable multirobot localization and mapping with relative maps: Introducing MOARSLAM. IEEE Control Syst. Mag. 2016, 36, 75–85. [Google Scholar] [CrossRef]
  7. Madhavan, R.; Fregene, K.; Parker, L. Distributed cooperative outdoor multirobot localisation and mapping. Auton. Robot. 2004, 17, 23–41. [Google Scholar] [CrossRef]
  8. Ivanjko, E.; Kitanov, A.; Petrovic, I. Model based Kalman Filter Mobile Robot Self-Localization. In Robot Localization Map Building; IntechOpen: London, UK, 2010; pp. 59–90. [Google Scholar] [CrossRef] [Green Version]
  9. Romaniuk, S.; Ambroziak, L.; Gosiewski, Z.; Isto, P. Real time localization system with Extended Kalman Filter for indoor applications. In Proceedings of the 2016 21st International Conference on Methods and Models in Automation and Robotics (MMAR), Miedzyzdroje, Poland, 29 August–1 September 2016; pp. 42–47. [Google Scholar] [CrossRef]
  10. Fox, D.; Burgard, W.; Kruppa, H.; Thrun, S. Probabilistic approach to collaborative multi-robot localization. Auton. Robot. 2000, 8, 325–344. [Google Scholar] [CrossRef]
  11. Prorok, A.; Martinoli, A. A reciprocal sampling algorithm for lightweight distributed multi-robot localization. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 3241–3247. [Google Scholar] [CrossRef] [Green Version]
  12. Al-Shareeda, M.A.; Manickam, S.; Mohammed, B.A.; Al-Mekhlafi, Z.G.; Qtaish, A.; Alzahrani, A.J.; Alshammari, G.; Sallam, A.A.; Almekhlafi, K. Provably Secure with Efficient Data Sharing Scheme for Fifth-Generation (5G)-Enabled Vehicular Networks without Road-Side Unit (RSU). Sustainability 2022, 14, 9961. [Google Scholar] [CrossRef]
  13. Roumeliotis, S.I.; Bekey, G.A. Distributed multirobot localization. IEEE Trans. Robot. Autom. 2002, 18, 781–795. [Google Scholar] [CrossRef] [Green Version]
  14. Kia, S.S.; Rounds, S.; Martinez, S. Cooperative Localization for Mobile Agents: A Recursive Decentralized Algorithm Based on Kalman-Filter Decoupling. IEEE Control Syst. Mag. 2016, 36, 86–101. [Google Scholar] [CrossRef] [Green Version]
  15. Li, W.; Jelfs, B.; Kealy, A.; Wang, X.; Moran, B. Cooperative Localization Using Distance Measurements for Mobile Nodes. Sensors 2021, 21, 1507. [Google Scholar] [CrossRef]
  16. Ligorio, G.; Sabatini, A.M. Extended Kalman filter-based methods for pose estimation using visual, inertial and magnetic sensors: Comparative analysis and performance evaluation. Sensors 2013, 13, 1919–1941. [Google Scholar] [CrossRef] [PubMed]
  17. Lin, J.; Lou, Z.; Zweigel, R.; Abel, D. Cooperative Localization of a Networked Multi-Vehicle System. IFAC-PapersOnLine 2019, 52, 67–72. [Google Scholar] [CrossRef]
  18. Wu, Y.; Peng, B.; Wymeersch, H.; Seco-Granados, G.; Kakkavas, A.; Garcia, M.H.C.; Stirling-Gallacher, R.A.; Casta, M.H.; Stirling-Gallacher, R.A. Cooperative Localization with Angular Measurements and Posterior Linearization. In Proceedings of the 2020 IEEE International Conference on Communications Workshops (ICC Workshops), Dublin, Ireland, 7–11 June 2020. [Google Scholar] [CrossRef]
  19. Qu, M.; Lu, H. Cooperative Simultaneous Localization and Mapping with Local Measurements in 2D Space. In Proceedings of the 2021 6th International Conference on Automation, Control and Robotics Engineering (CACRE), Dalian, China, 15–17 July 2021; pp. 509–513. [Google Scholar] [CrossRef]
  20. Han, Y.; Wei, C.; Li, R.; Wang, J.; Yu, H. A novel cooperative localization method based on IMU and UWB. Sensors 2020, 20, 467. [Google Scholar] [CrossRef] [PubMed]
  21. Héry, E.; Xu, P.; Bonnifait, P.; Hery, E.; Xu, P.; Bonnifait, P. Consistent decentralized cooperative localization for autonomous vehicles using LiDAR, GNSS, and HD maps. J. Field Robot. 2021, 38, 552–571. [Google Scholar] [CrossRef]
  22. Pires, A.G.; Rezeck, P.A.F.; Chaves, R.A.; Macharet, D.G.; Chaimowicz, L. Cooperative Localization and Mapping with Robotic Swarms. J. Intell. Robot. Syst. Theory Appl. 2021, 102, 47. [Google Scholar] [CrossRef]
  23. Ben, Y.; Sun, Y.; Li, Q.; Cui, W.; Zhang, Q. A Cooperative Localization Algorithm Considering Unknown Ocean Currents for Multiple AUVs. In Proceedings of the 2021 International Conference on Autonomous Unmanned Systems (ICAUS 2021), Changsha, China, 24–26 September 2021; Lecture Notes in Electrical Engineering. Springer Science and Business Media Deutschland GmbH: Berlin/Heidelberg, Germany, 2022; Volume 861, pp. 1379–1387. [Google Scholar] [CrossRef]
  24. Qi, Y.; Zhong, Y.; Shi, Z. Cooperative 3-D relative localization for UAV swarm by fusing UWB with IMU and GPS. J. Phys. Conf. Ser. 2020, 1642, 012028. [Google Scholar] [CrossRef]
  25. Li, W.; Liu, Y.; Li, X.; Shen, Y. Three-dimensional cooperative localization via space-air-ground integrated networks. China Commun. 2022, 19, 253–263. [Google Scholar] [CrossRef]
  26. Mayle, M.N.; Sharma, R. Cooperative Localization in a GPS-Limited Environment Using Inter-Vehicle Range Measurements for a System of Multiple, Non-Homogeneous Vehicles. In Proceedings of the AIAA Scitech 2021 Forum, Virtual Event, 11–15 and 19–21 January 2021; AIAA American Institute of Aeronautics and Astronautics, Inc.: Reston, VA, USA, 2021; pp. 1–18. [Google Scholar] [CrossRef]
  27. Wanasinghe, T.R.; George, G.K.; Gosine, R.G. Relative Localization Approach for Combined Aerial and Ground Robotic System. J. Intell. Robot. Syst. Theory Appl. 2015, 77, 113–133. [Google Scholar] [CrossRef]
  28. Piasco, N.; Marzat, J.; Sanfourche, M. Collaborative localization and formation flying using distributed stereo-vision. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 1202–1207. [Google Scholar] [CrossRef]
  29. Modalavalasa, N.; Sasi, G.; Rao, B.; Satya Prasad, K.; Ganesh, L.; Kumar, M.N.V.S.S. A new method of target tracking by EKF using bearing and elevation measurements for underwater environment. Rob. Auton. Syst. 2015, 74, 221–228. [Google Scholar] [CrossRef]
  30. De Silva, O.; Mann, G.K.I.I.; Gosine, R.G. Development of a relative localization scheme for ground-aerial multi-robot systems. In Proceedings of the 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vilamoura-Algarve, Portugal, 7–12 October 2012; pp. 870–875. [Google Scholar] [CrossRef]
  31. Wanasinghe, T.R.; Mann, G.K.I.; Gosine, R.G. Distributed collaborative localization for a heterogeneous multi-robot system. In Proceedings of the 2014 IEEE 27th Canadian Conference on Electrical and Computer Engineering (CCECE), Toronto, ON, Canada, 4–7 May 2014; pp. 1–6. [Google Scholar] [CrossRef]
  32. Wanasinghe, T.R.; Mann, G.K.I.I.; Gosine, R.G. Distributed Leader-Assistive Localization Method for a Heterogeneous Multirobotic System. IEEE Trans. Autom. Sci. Eng. 2015, 12, 795–809. [Google Scholar] [CrossRef]
  33. Panzieri, S.; Pascucci, F.; Sciavicco, L.; Setola, R. Distributed multi-robot localization. In Robotics: Concepts, Methodologies, Tools, and Applications; IGI Global: Hershey, PA, USA, 2013; Volume 1, pp. 391–406. [Google Scholar] [CrossRef]
  34. Carrillo-Arce, L.C.; Nerurkar, E.D.; Gordillo, J.L.; Roumeliotis, S.I. Decentralized multi-robot cooperative localization using covariance intersection. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 1412–1417. [Google Scholar] [CrossRef] [Green Version]
  35. Wanasinghe, T.R.; Mann, G.K.I.; Gosine, R.G. Decentralized Cooperative Localization Approach for Autonomous Multirobot Systems. J. Robot. 2016, 2016, 2560573. [Google Scholar] [CrossRef] [Green Version]
  36. Kia, S.S.; Hechtbauer, J.; Gogokhiya, D.; Martinez, S. Server-Assisted Distributed Cooperative Localization over Unreliable Communication Links. IEEE Trans. Robot. 2018, 34, 1392–1399. [Google Scholar] [CrossRef] [Green Version]
  37. Wang, L.; Zhang, T.; Gao, F. Distributed cooperative localization with lower communication path requirements. Rob. Auton. Syst. 2015, 79, 26–39. [Google Scholar] [CrossRef]
  38. Wang, X.; Guo, Y.; Cao, J.; Wu, M.; Sun, Z.; Lv, C. Simplify Belief Propagation and Variation Expectation Maximization for Distributed Cooperative Localization. Appl. Sci. 2022, 12, 3851. [Google Scholar] [CrossRef]
  39. Shi, X.; Wang, T.; Huang, B.; Zhao, C. Cooperative multi-robot localization based on distributed UKF. In Proceedings of the 2010 3rd International Conference on Computer Science and Information Technology, Chengdu, China, 9–11 July 2010; Volume 6, pp. 590–593. [Google Scholar] [CrossRef]
  40. Kandepu, R.; Foss, B.; Imsland, L. Applying the unscented Kalman filter for nonlinear state estimation. J. Process Control 2008, 18, 753–768. [Google Scholar] [CrossRef]
  41. Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 1958. [Google Scholar] [CrossRef] [Green Version]
  42. Ponsa, D.; López, A.; Serrat, J.; Lumbreras, F.; Graf, T. Multiple vehicle 3D tracking using an unscented Kalman Filter. In Proceedings of the 2005 IEEE Conference on Intelligent Transportation Systems, Vienna, Austria, 16 September 2005; pp. 1108–1113. [Google Scholar] [CrossRef]
  43. Huang, G.; Kaess, M.; Leonard, J.J. Consistent unscented incremental smoothing for multi-robot cooperative target tracking. Robot. Auton. Syst. 2015, 69, 52–67. [Google Scholar] [CrossRef]
  44. Lyu, Y.; Pan, Q.; Lv, J. Unscented transformation-based multi-robot collaborative self-localization and distributed target tracking. Appl. Sci. 2019, 9, 903. [Google Scholar] [CrossRef] [Green Version]
  45. der Merwe, R.; Wan, E.A.; Van Der Merwe, R.; Wan, E.A. The square-root unscented Kalman filter for state and parameter-estimation. In Proceedings of the 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing, Salt Lake City, UT, USA, 7–11 May 2001; Volume 6, pp. 3461–3464. [Google Scholar] [CrossRef]
  46. Oliveros, J.C.; Ashrafiuon, H. Cooperative Localization of Vehicles in Three-Dimensional Space. In Proceedings of the 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing, Salt Lake City, UT, USA, 7–11 May 2001; Volume 2. [Google Scholar] [CrossRef]
  47. Bar-Shalom, Y.; Willett, P.K.; Tian, X. Tracking and Data Fusion: A Handbook of Algorithms; YBS Publishing: London, UK, 2011. [Google Scholar]
  48. Betsch, P.; Siebert, R. Rigid body dynamics in terms of quaternions: Hamiltonian formulation and conserving numerical integration. Int. J. Numer. Methods Eng. 2009, 79, 444–473. [Google Scholar] [CrossRef]
  49. Fetzer, K.K.L.; Nersesov, S.; Ashrafiuon, H. Nonlinear control of three-dimensional underactuated vehicles. Int. J. Robust Nonlinear Control 2020, 30, 1607–1621. [Google Scholar] [CrossRef]
  50. Zhao, F.; Van Wachem, B.G. A novel Quaternion integration approach for describing the behaviour of non-spherical particles. Acta Mech. 2013, 3109, 3091–3109. [Google Scholar] [CrossRef] [Green Version]
  51. Gros, S.; Zanon, M.; Diehl, M. Baumgarte stabilisation over the SO(3) rotation group for control. In Proceedings of the 2015 54th IEEE Conference on Decision and Control (CDC), Osaka, Japan, 15–18 December 2015; pp. 620–625. [Google Scholar] [CrossRef]
  52. Rucker, C. Integrating Rotations Using Nonunit Quaternions. IEEE Robot. Autom. Lett. 2018, 3, 2979–2986. [Google Scholar] [CrossRef]
Figure 1. Vehicle 3D Kinematic Model.
Figure 1. Vehicle 3D Kinematic Model.
Applsci 12 11805 g001
Figure 2. Pose geometry for a two-vehicle network.
Figure 2. Pose geometry for a two-vehicle network.
Applsci 12 11805 g002
Figure 3. Centralized EKF-based CL diagram.
Figure 3. Centralized EKF-based CL diagram.
Applsci 12 11805 g003
Figure 4. Vehicle actual path versus paths estimated with and without CL in the test case.
Figure 4. Vehicle actual path versus paths estimated with and without CL in the test case.
Applsci 12 11805 g004
Figure 5. Time history of vehicle position error norm in the test case.
Figure 5. Time history of vehicle position error norm in the test case.
Applsci 12 11805 g005
Figure 6. Time history of vehicle orientation error norm in the test case.
Figure 6. Time history of vehicle orientation error norm in the test case.
Applsci 12 11805 g006
Figure 7. Scenario 1—vehicle 1 mean Δ r norm.
Figure 7. Scenario 1—vehicle 1 mean Δ r norm.
Applsci 12 11805 g007
Figure 8. Scenario 1—vehicle 1 mean Δ q norm.
Figure 8. Scenario 1—vehicle 1 mean Δ q norm.
Applsci 12 11805 g008
Figure 9. Comparison of error norms when vehicle 1 measures its own position for 1 s at the 100 s mark. Dashed lines indicate CL with relative position measurements for the first 100 s, and solid lines indicate no CL.
Figure 9. Comparison of error norms when vehicle 1 measures its own position for 1 s at the 100 s mark. Dashed lines indicate CL with relative position measurements for the first 100 s, and solid lines indicate no CL.
Applsci 12 11805 g009
Figure 10. Overview of a mobile robot used for experiment.
Figure 10. Overview of a mobile robot used for experiment.
Applsci 12 11805 g010
Figure 11. Comparison of error norms when vehicle 3 measures its own pose for 1 s at the 40 s mark. Dashed lines indicate CL with relative and absolute pose measurements starting at 40 s, and solid lines indicate no CL.
Figure 11. Comparison of error norms when vehicle 3 measures its own pose for 1 s at the 40 s mark. Dashed lines indicate CL with relative and absolute pose measurements starting at 40 s, and solid lines indicate no CL.
Applsci 12 11805 g011
Figure 12. Comparison of error norms when vehicle 1 measures its own position for 1 s at the 40 s mark. Dashed lines indicate CL with relative position measurements for the first 40 s, and solid lines indicate no CL.
Figure 12. Comparison of error norms when vehicle 1 measures its own position for 1 s at the 40 s mark. Dashed lines indicate CL with relative position measurements for the first 40 s, and solid lines indicate no CL.
Applsci 12 11805 g012
Table 1. Limitations of related works.
Table 1. Limitations of related works.
AlgorithmAdvantagesLimitations
SLAMObtain localization of multiple vehicles, mapping of an environmentRequires knowledge of the environment and designated landmarks, needs assigned “leaders” and “followers” or a central processing station
Particle FilterNonlinear applications, low error in ideal situations, no linearization requiredRequires individual agent identification, not robust to false-positive detections
Decentralized EKF with Euler anglesNo designated central processing center neededProne to singularity issues due to gimbal lock, requires linearization (Jacobian)
Current MethodSingularity-free, computationally effectiveQuaternions cannot be measured directly, requires linearization
Table 2. Input linear (m/s) and angular (°/s) velocities of vehicles for all simulations.
Table 2. Input linear (m/s) and angular (°/s) velocities of vehicles for all simulations.
[ v x , v y , v z ] [ ω x , ω y , ω z ]
LinearAngular
V 1 −0.30, 0.00, 0.000.28, 0.55, 0.55
V 2 −0.35, 0.00, 0.000.35, 0.70, 0.70
V 3 0.40, 0.00, 0.000.43, 0.85, 0.85
V 4 0.35, 0.00, 0.000.50, 1.00, 1.00
V 5 −0.30, 0.00, 0.000.58, 1.15, 1.15
V 6 −0.35, 0.00, 0.000.65, 1.30, 1.30
Table 3. Simultaneous measurement descriptions for Scenario 1.
Table 3. Simultaneous measurement descriptions for Scenario 1.
Simultaneous Measurements ( b t )
Case 11 → 2
Case 21 → 2, 2 → 3
Case 31 → 2, 2 → 3, 3 → 4
Case 41 → 2, 2 → 3, 3 → 4, 4 → 5
Case 51 → 2, 2 → 3, 3 → 4, 4 → 5, 5 → 6
Table 4. Position (m) and quaternion error statistics of vehicle 1 across 100 simulations for Scenario 1.
Table 4. Position (m) and quaternion error statistics of vehicle 1 across 100 simulations for Scenario 1.
PositionOrientation
e r m e a n e r s t d e r m i n e r m a x e q m e a n e q s t d e q m i n e q m a x
Propagate2.870.012.842.910.050.000.050.05
Case 11.990.820.704.480.050.010.020.09
Case 21.480.850.384.560.040.020.010.09
Case 31.500.760.394.340.030.010.010.08
Case 41.330.690.394.340.030.010.010.07
Case 51.230.580.243.030.030.010.010.05
Case 5 [14]1.270.011.161.190.040.000.030.04
Table 5. Position (m) and quaternion error statistics across 100 simulations of all vehicles for Scenario 1.
Table 5. Position (m) and quaternion error statistics across 100 simulations of all vehicles for Scenario 1.
PositionOrientation
e r m e a n e r s t d e r m i n e r m a x e q m e a n e q s t d e q m i n e q m a x
Propagate14.970.0214.9115.030.430.000.430.44
Case 113.301.6110.8018.370.420.030.370.49
Case 210.582.587.1619.590.360.040.280.50
Case 39.332.775.4019.830.300.050.210.44
Case 47.822.913.5020.950.230.050.130.40
Case 56.152.591.6714.130.150.060.050.33
Case 5 [14]8.200.058.078.320.210.000.200.21
Table 6. Simultaneous measurement descriptions for Scenario 2.
Table 6. Simultaneous measurement descriptions for Scenario 2.
Simultaneous Measurements ( b t )
Case 11 → 2
Case 21 → 2, 2 → 3
Case 31 → 2, 2 → 3, 3 → 1
Case 41 → 2, 2 → 3, 3 → 1, 2 → 1
Case 51 → 2, 2 → 3, 3 → 1, 2 → 1, 3 → 2
Case 61 → 2, 2 → 3, 3 → 1, 2 → 1, 3 → 2,1 → 3
Table 7. Position (m) and quaternion error statistics across 100 simulations of all vehicles for Scenario 2.
Table 7. Position (m) and quaternion error statistics across 100 simulations of all vehicles for Scenario 2.
PositionOrientation
e r m e a n e r s t d e r m i n e r m a x e q m e a n e q s t d e q m i n e q m a x
Propagate18.080.0417.9818.160.360.000.360.36
Case 114.202.939.9523.070.330.050.240.47
Case 210.284.763.6924.400.230.080.090.50
Case 38.443.662.6120.940.180.070.060.37
Case 48.234.602.5623.840.170.070.060.33
Case 510.734.573.4723.460.200.070.050.39
Case 611.334.134.6624.180.210.060.100.40
Table 8. Mean position error (m) statistics across 100 simulations of all vehicles for Scenario 2 with only relative position measurements.
Table 8. Mean position error (m) statistics across 100 simulations of all vehicles for Scenario 2 with only relative position measurements.
e r m e a n e r s t d e r m i n e r m a x
Propagate1.350.011.321.39
Case 11.340.011.311.38
Case 21.300.101.252.28
Case 31.280.021.241.32
Case 41.280.021.251.32
Case 51.280.011.251.32
Case 61.270.011.251.31
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Oliveros, J.C.; Ashrafiuon, H. Application and Assessment of Cooperative Localization in Three-Dimensional Vehicle Networks. Appl. Sci. 2022, 12, 11805. https://doi.org/10.3390/app122211805

AMA Style

Oliveros JC, Ashrafiuon H. Application and Assessment of Cooperative Localization in Three-Dimensional Vehicle Networks. Applied Sciences. 2022; 12(22):11805. https://doi.org/10.3390/app122211805

Chicago/Turabian Style

Oliveros, Juan Carlos, and Hashem Ashrafiuon. 2022. "Application and Assessment of Cooperative Localization in Three-Dimensional Vehicle Networks" Applied Sciences 12, no. 22: 11805. https://doi.org/10.3390/app122211805

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop