Next Article in Journal
A Lightweight Hierarchical Activity Recognition Framework Using Smartphone Sensors
Next Article in Special Issue
Sequential and Automatic Image-Sequence Registration of Road Areas Monitored from a Hovering Helicopter
Previous Article in Journal
Optimization of an Optical Inspection System Based on the Taguchi Method for Quantitative Analysis of Point-of-Care Testing
Previous Article in Special Issue
Investigation of Matlab® as Platform in Navigation and Control of an Automatic Guided Vehicle Utilising an Omnivision Sensor
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments

1
Intelligent Vehicle Research Center, Beijing Institute of Technology, 5 South Zhongguancun Street, Haidian District, Beijing 10081, China
2
Beijing Automotive Technology Center, Beijing 101300, China
3
Department of Electrical and Computer Engineering, Brigham Young University, Provo, UT 84602, USA
*
Author to whom correspondence should be addressed.
Sensors 2014, 14(9), 16159-16180; https://doi.org/10.3390/s140916159
Submission received: 23 July 2014 / Revised: 24 August 2014 / Accepted: 26 August 2014 / Published: 1 September 2014
(This article belongs to the Special Issue Positioning and Tracking Sensors and Technologies in Road Transport)

Abstract

: This paper presents a monocular visual odometry algorithm that incorporates a wheeled vehicle model for ground vehicles. The main innovation of this algorithm is to use the single-track bicycle model to interpret the relationship between the yaw rate and side slip angle, which are the two most important parameters that describe the motion of a wheeled vehicle. Additionally, the pitch angle is also considered since the planar-motion hypothesis often fails due to the dynamic characteristics of wheel suspensions and tires in real-world environments. Linearization is used to calculate a closed-form solution of the motion parameters that works as a hypothesis generator in a RAndom SAmple Consensus (RANSAC) scheme to reduce the complexity in solving equations involving trigonometric. All inliers found are used to refine the winner solution through minimizing the reprojection error. Finally, the algorithm is applied to real-time on-board visual localization applications. Its performance is evaluated by comparing against the state-of-the-art monocular visual odometry methods using both synthetic data and publicly available datasets over several kilometers in dynamic outdoor environments.

Graphical Abstract

1. Introduction

Self-localization is arguably one of the most challenging problems in intelligent vehicle research. Traditionally, wheel speed encoder is used for auto-localization of wheeled autonomous vehicles. However, the accuracy of localization using photoelectric encoder is oblivious to external environment due to its dependency on proprioceptive sensors. The gross error is inevitable [1] when abnormal situations, such as wheel slippage or sinkage, occur. Another drawback of wheel encoder-based odometry is the non-boundary accumulated error [2]. Recently, a class of vision based localization method has been proposed to address these shortcomings. Using visual information to estimate the motion of a ground robot has been an active field of research in the past 30 years. Monocular, stereo and ominidirectional cameras have all been used in vision based motion estimation systems. A notable successful example of visual odometry is NASA's Mars rovers [3].

Literature review shows that visual odometry (VO) algorithms are divided into feature-based and appearance-based methods. Only feature-based methods are studied in this work because of their efficiency and robustness. As the term suggests, feature-based methods focus on tracking salient and reoccurring features over frames for motion estimation. Based on the principles of visual geometry, the relative camera pose or movement between frames can be estimated from a certain number of corresponding points. Various features, including point, line [4], contour [5], and hybrid [6], all have been used for this purpose. Point feature is the most prevalent of all because of its simplicity. However, because of errors in point feature detection, matching and tracking, data association or outlier removal problem becomes very critical. Feature detection and match algorithm is not the focus of this work. A good overview can be found in [7]. The focus of this paper is motion estimation.

For motion estimation with the presence of outliers, the RAndom SAmple Consensus (RANSAC) paradigm [8] based on hypothesis-and-test mechanism has been established as the standard method. The minimal algorithms mentioned above can act as the hypothesis generator in RANSAC framework. Minimal solvers, such as 5-point RANSAC, for monocular VO, and 3-point [9] RANSAC, for stereo VO, are now widely used [10].

The number of RANSAC iterations increases exponentially as the increase of the number of parameters that describe the motion. Vehicle kinematic constraints have been applied to VO in order to reduce the size of minimal solvers and hence reduces the number of iterations of RANSAC [1114]. Vatani et al. used the Ackermann steering principle and the planar assumption to constraint the motion model [11]. Two-dimensional planar motion was then estimated directly using pixel displacement from a downward-looking camera. Using the vehicle kinematic model, Zhu et al. [14] computed the motion parameters by solving a quadratic polynomial from equations established by the epipolar constraint. For general wheeled vehicles, Scaramuzza et al. [12,13] showed that, due to the existence of the Instantaneous Center of Rotation, motion can be locally described as planar and circular, and the motion model can be simplified to 1 DoF (Degree of Freedom). This simplification leads to a 1-point minimal solver. Their restrictive model is based on the assumption that the motion is locally planar and circular. This assumption is often violated in outdoor environments even when the road is perfectly flat. If we treat the camera and the vehicle as a spring-mass system, when acceleration, deceleration, or sharp turns occurs, the planar-motion hypothesis will fail due to the dynamic characteristics of wheel suspensions and tires. For these reasons, in their latest work [15], the authors relaxed the locally planar and circular motion constraints. The motion prior to using 1-point algorithm is used to compute the target distribution of the 6 DoF motions, leading to improved performance compared to their 1-point algorithm.

All the aforementioned kinematic model-based methods totally ignore the tire side slip characteristics, which cause deviation in the estimation of translation vector. The effect is evident especially in highway driving or even low-speed steering when the non-linear dynamic characteristics of vehicle are significant. In most practical situations the camera is fixed to sprung mass, which does not satisfy the planar motion hypothesis due to dynamic characteristic of wheel suspension. Hamme [16] also mentioned that the camera orientation with relation to the road plane can not remain fixed due to the compliant suspension system. In this work we consider the change of pitch angle, and make a sophisticated linearization of dynamic model in order to estimate principle motion parameters accurately and efficiently.

The remainder of this paper is organized as follows. In Section 2, we describe the problem formulation and some notations. In Section 3, we introduce both kinematic model and basic dynamic model of wheeled vehicles. In Section 4, we discuss how to involve the vehicle dynamic model in monocular visual pose estimation. In Section 5, we briefly describe the refinement approach. Finally, in Section 6, we present the experimental results on both synthetic and real data.

2. Problem Formulation and Notations

Considering a 3D world point Q and its projections q1 and q2 from two views that are all represented in homogeneous coordinates. A particular 3 × 4 camera matrix P can be defined for a view indicating the image projection q = PQ (up to a scale factor). P can be factored into P = K[Rt], where K is a calibration matrix holding the camera intrinsic parameters. R ∈ SO(3) and t ∈ R3 × 1 are the 3 × 3 rotation matrix and 3 × 1 translation vector that represent camera motion between the two views.

The fundamental matrix F is defined as:

F K 2 T [ t ] × R K 1 1
where [t]× denotes the skew symmetric matrix defined by t. The co-planarity equation or epipolar constraint can be expressed as:
q 2 T F q 1 = 0

For a calibrated camera, K1 and K2 (intrinsic parameters from two views of the same camera) are known and supposed to be the same. The projections q1 and q2 represented in image pixels can be converted to normalized vectors projected onto the image plane in the camera coordinate system as 1 and 2. Consequently the epipolar constraint can be rewritten as:

q ^ 2 T E q ^ 1 = 0
where E = [t]×R is called the essential matrix, q ^ 1 = K 1 1 q 1, q ^ 2 = K 2 1 q 2.

Given five, seven, or eight known point correspondences, E can be calculated according to the visual geometry principle. Motion parameters t and R can then be extracted from E using Singular Value Decomposition (SVD). After the camera rotation and translation between consecutive frames are estimated, the location and orientation of the camera (or vehicle) with respect to the vehicle's initial position can be calculated cumulatively.

In general, vision based motion estimation algorithms consist of two steps: inner identification and outer refinement, as depicted in Figure 1. In the inner loop, feature correspondences are separated into inliers and outliers using a hypothesis-test-validation paradigm, such as the standard RANSAC. The hypothesis generator uses a minimal number of feature correspondences to solve the relative orientation problem. The minimal algorithm is named after the size of the minimal set, typically 5-point [17], 7-point [18], 8-point [19] are widely used. For a calibrated camera, at least five point correspondences are required to determine the relative pose between two frames [20]. Considering common M-point algorithms for solving the relative pose problem, the number of iterations nRANSAC that is required to guarantee a stable solution is:

n RANSAC = log ( 1 p ) log ( 1 ( 1 ɛ ) s )
where s is the size of minimal set, ε is the percentage of outliers and p is the requested probability of success [8].

Obviously, nRANSAC will increase exponentially as s increases. A summary of the number of RANSAC iterations needed when s changes can be found in [12].

In the outer step, the winner minimal solution is refined with all the inliers found through minimizing reprojection error or other similar error metrics. The most popular method used as the final refining process is Bundle Adjustment (BA) [21], which optimizes reprojection error over both structure and motion parameters using the Levenberg-Marquardt algorithm.

The main contribution of this work is to incorporate the vehicle dynamic model in the inner loop to effectively select inliers. This approach provides more accurate result than the traditional kinematic model based methods while maintaining the same computation requirements. It demonstrates better performance even with the non-linear dynamic characteristics of vehicle when the vehicle is moving at high speeds. This method is able to handle sophisticated linearization of principle motion parameters and yield promising motion estimation results. This new method is named MYP algorithm for its use of the dynamic Model of Yaw and its consideration of Pitch as well.

Respecting the customs in photogrammetry [18], the Camera Coordinate System (CCS) defined in this work is to let z-axis point to the forward direction as shown in Figure 2, where V denotes the centroid of the vehicle as well as the origin of the Vehicle Coordinate System while C denotes the origin of CCS; OA represents the Optical Axis of the camera; u and v are pixel coordinates on the image plane; H represents the height of camera mount. The CCS at initial time t = 0 is defined as the Global Coordinate System.

3. Motion Models of Wheeled Vehicles

In order to reduce the degree of freedom in vision based motion estimation problem, a few researchers have introduced motion models of wheeled vehicle into the pose estimation process. We will present the kinematic model and dynamic model in this section. We will also explain how these models work and the analysis of the model fitness using synthetic and real data.

3.1. Nonholonomic Constraints of Ackerman Steering

As mentioned before, the classical Ackerman Steering Principles are used to ensure a smooth motion of vehicle, as well as the camera, where the camera is simply described with local circular motion. According to [12] the rotation matrix R and translation vector t can be parameterized as:

R = [ cos ω 0 sin ω 0 1 0 sin ω 0 cos ω ] t = ρ [ sin β 0 cos β ]
where ω is the yaw angle increment of the vehicle and β is the vehicle body side slip angle. The key of 1-point algorithm is the geometric relationship between ω and β, i.e., β = ω/2. The essential matrix E for the planar circular motion can then be described using a single parameter ω, and the closed-form solution of ω can be derived. Consider a set of 2D feature correspondences {iuk-1iuk} and i = 1, 2, … N, where N is the total number of feature points and k represents the frame number. For each feature correspondence iuk-1iuk, the yaw angle increment can be determined using only one correspondence as:
ω i = 2 arctan û i k 1 ( 2 ) · û i k ( 1 ) û i k 1 ( 1 ) · û i k ( 2 ) û i k 1 ( 3 ) · û i k ( 2 ) + û i k 1 ( 2 ) · û i k ( 3 )
where the hat denote the normalized vector of image projection, ik/k−1(j) denotes the j-th element of vector ik/k−1.

Then, the best estimate of yaw angle increment, denoted as ω*, can be selected through histogram voting or a RANSAC based method. Using ω* and Equation (5), the rotation matrix R1 and translation vector t1 can be determined. Here, the superscript 1 represents model estimation results using the 1-point algorithm.

Three model based motion estimation methods, specifically using the 1-point algorithm, are summarized below:

Method 1:

Uses {iω} and ω* to reject outliers whose estimation error ‖iωω*‖ is larger than some threshold.

Method 2:

Triangulates features with motion baseline R1 and t1 to reject outliers by thresholding the reprojection error.

Method 3:

Uses stereo triangulated 3D points {iX} and 2D image projection {iuk−1} to reject outliers by thresholding the reprojection error.

Basically, Method 1 can only be used as a coarse filter with lax thresholding. In monocular setting, only Method 2 can work like the original application in [12]. Method 3 works only when stereo vision is available.

3.2. Single-Track Bicycle Dynamic Model

As discussed in Section 3.1, the tire side slip characteristics are totally ignored in the kinematic model. The omission of the tire side slip causes deviation in the estimation of translation vector. In [14], the yaw increment ω and side slip angle β are treated as independent variables which lead to quadratic equations from feature correspondences. We are more interested in the inherent relationship between ω and β, which is naturally relevant in vehicle dynamics. Single-track bicycle model described by Mitschke in [22] is the simplest dynamic model of wheeled vehicles. This model combines the front and rear wheels and treats the vehicle as a bicycle. It originally describes the vehicle motion in three degrees of freedom (x-y position and yaw rate) in integral use.

The following are the differential equations of the bicycle model:

( k F + k R ) β + 1 v long ( l F k F l R k R ) ω r k F δ = m ( v ˙ lat + v long ω r )
( l F k R l R k R ) β + 1 v long ( l F 2 k F + l R 2 k R ) ω r l F k F δ = I y ω ˙ r

In these two equations, kF and kR denote the effective cornering stiffness of front and rear tires respectively and m represents the vehicle curb weight. Iy is the moment of inertia around the y-axis (defined in vehicle coordinate system, see Figure 2). lF and lR are geometric values that denote the distance from the front and rear axles to the centroid of vehicle respectively. vlong and vlat represent longitudinal and lateral velocity and ωr denotes yaw rate. β denotes the vehicle body side slip angle and δ represents the wheel steering angle. There is a difference between ωr and pre-defined ω. ωr is canonical yaw rate of which dimension is rad/second, while ω is yaw angle increment in each interval, its dimension is rad, and their relationship fulfill ω = ωrdt where dt is the acquisition interval.

3.3. Model Fitness Analysis

We observe the values of ω and β using both synthetic data and real car data. In Figure 3, we present the plots of ω and β using CarSim simulated data, and the detail of simulation configuration was proposed in our previous work [23]. It is interesting to note that there is an ideal linear proportion between ω and β when the vehicle runs at medium or low speed (typically less than 50 km/h). Whereas, when the vehicle is running at high speed, the non-linear relationship becomes noticeable. At the medium and low speed applications, the scale factor between ω and β varies with the speed. The slower is the speed, the higher is the ratio of β/ω. The plots from real car data shown in Figure 4 also confirm this observation.

We compared the angles using one KITTI sequence [24]. The yaw angle increment ω and side slip angle β for high speed (top left plot) and low speed (top right plot) and their corresponding speed over time are plotted in Figure 4. Both the non-linear proportion of β/ω at high speed (60–100 km/h on the left) and the linear proportion of β/ω in low-speed steering (around 20 km/h on the right) are far from the β/ω = 1/2 assumed in the kinematic model.

We are interested in investigating the influence of model angle bias on the performance of 1-point algorithm. In Figure 5 we give a simple demonstration using synthetic data. The motion is simulated with fixed yaw angle increment while increasing the pitch angle increment and roll angle increment respectively, then the reprojection error are calculated using the canonical pipeline of 1-Point algorithm described in Section 3.1. Figure 5 shows that the increase of pitch angle increment causes more severe damage to the distribution of reprojection error, and finally damages the inlier identification result. Therefore, it is necessary to consider pitch angle increment separately in the model estimation step.

As the inlier identification procedure can be regarded as a binary classification problem, we use sensitivity and specificity as evaluation metrics. Define True Positive (TP) as the number of true inliers found, False Negative (FN) as the number of true inliers missed, True Negative (TN) as the number of true outliers found, and False Positive (FP) as the number of true outliers identified as inliers. To compute the true inliers and true outliers, we use ground truth poses and a reprojection-error threshold of 1 pixel. Then, sensitivity (also called true-positive rate or recall rate) and specificity (or true negative rate) can be computed as:

sensitivity = TP TP + FN , specificity = TN TN + FP

In ideal cases, we would like sensitivity = 1 and specificity = 1.

The following is an exemplary evaluation of the outlier-removal performance of the original 1-point algorithm [12] and MOBRAS [15] on the publicly-available KITTI dataset. We choose Sequence 05 as it is a very complicated path, with many sharp turns as well as many accelerations and decelerations.

The relations between sensitivity, specificity, and the non-planar-angle component of the rotation (defined as pitch angle here) are depicted in Figure 6. From the distribution of sensitivity, MOBRAS finds on average more true inliers than the 1-point algorithm, in which detection ratio decreases significantly as the non-planar angle increases. However, the specificity plot indicates that both methods suffer from high false positive rate, which can undermine the inlier quality and undermine the motion estimation result. However, we will show in the following sections that by introducing the linearized pitch angle component, motion estimation accuracy can be refined and improved.

4. Dynamic Model Based MYP Algorithm

In this section we explore the relationship between yaw angle increment ω and side slip angle β, and try to represent β as a linear combination of ω and other measurable parameters. We also present how to solve ω and pitch change γ with this parameterization. The processing time of our MYP algorithm is the same as the 1-point algorithm but it improves the estimation accuracy significantly.

4.1. Exploring the Relationship between Yaw Rate and Side Slip Angle

Recall that in Equations (7) and (8), we want to derive the relationship between yaw rate ωr and side slip angle β, which are the principal components of vehicle motion. Because we do not need to estimate wheel steering angle δ, it is easy to eliminate it by multiplying Equation (7) by lF and subtracting Equation (8) from it. The result of this elimination is:

( l F + l R ) k R β l R v long ( l F + l R ) k R ω r = m · l F · v ˙ lat + m · l F · v long · ω r I y ω ˙ r

Or:

β = l R ω r v long + m k R · l F l · a lat + m k R · l F l · v long · ω r I y l · k R ω ˙ r
where l = lF + lR denotes the wheelbase of the vehicle, and alat represents lateral acceleration.

As shown in Equation (11), the side slip angle β can be represented as a linear function of yaw rate ωr. Notice that if monocular set longitudinal velocity vlong is known (provide by other sensors), lF, l are geometric parameters and can be measured manually. m, kR and Iy can be estimated off-line using an identification method.

4.2. Parameterization of the Motion

Similar to the model in [12], local road planar hypothesis is also tenable in this work because the moving trajectory relates to unsprung mass. MYP algorithm (1) represents the side slip angle β as the linear function of yaw angle change ω, like β = c1ω + c2, where c 1 = ( l R v long + m k R · l F l · v long ) / d t, c 2 = m k R · l F l · a lat I y l · k R ω ˙ r, dt is the acquisition interval; and (2) adds the rotation component caused by pitch angle change.

First, the coefficients c1 and c2 can be calculated using known geometrical parameters lF, l, measured parameters vlong, alat and off-line trained parameters m, kR and Iy as shown in Equation (11). The off-line training method for these parameters will be described later in this section. Second, as we observed in both synthetic and real car data, we found that the pitch angle change γ was maintained to be less than 2° per 100 ms, it is reasonable to assume sin γγ, cos γ ≈ 1, finally the essential matrix is parameterized as:

E = [ 0 1 0 1 0 tan β 0 tan β 0 ] · [ cos ω 0 sin ω 0 1 0 sin ω 0 cos ω ] · [ 1 0 0 0 1 γ 0 γ 1 ] E = [ 0 1 γ cos ω tan β sin ω γ ( sin ω + cos ω tan β ) ( sin ω + cos ω tan β ) 0 tan β γ tan β ]

Substituted into Epipolar Constraint Equation (3), the equation constrained by feature correspondences is:

x k 1 y k ( cos ω tan β sin ω ) x k y k 1 y k 1 y k γ ( sin ω + cos ω tan β ) + y k 1 tan β + x k γ y k ( sin ω + cos ω tan β ) γ tan β = 0
where xk−1, yk−1, xk, yk are elements of calibrated image coordinates k−1 = [xk−1 yk−1 1]T and k = [xk yk 1]T of a couple of feature points.

Equation (13) contains complex trigonometric polynomials and it is difficult to solve for ω. Practically, in both synthetic data and real data, the yaw angle increment are typically less than 10° (per 100 ms). We use Taylor expansion of trigonometric functions and neglect the term with orders higher than O(ω2).

sin ω ω cos ω 1 ω 2 2 tan β = tan ( c 1 ω + c 2 ) c 1 ω + c 2 + ( c 1 ω + c 2 ) 3 3 c 2 + c 2 3 3 + c 1 ( 1 + c 2 2 ) ω + c 1 2 c 2 ω 2

Combing the simplification in Equations (13) and (14), we obtain a quadratic of ω and γ, which can be solved using Root Formula, Sturm-sequences, or any other root-solve algorithm realized by current library, like Eigen [25]. After that, β can be also recovered. Obviously, at most, two solutions will be obtained, which is much less than 11 possible solutions of the 5-point algorithm.

4.3. Off-Line Parameters Identification

As mentioned before, we need to estimate parameters including m, kR and Iy using an optimization approach. An error function is defined as the sum of all normalized error between modeled outputs and measured values. Let β˜k denote side slip angle measured in frame k and define coefficient matrix A as:

A ( k , 1 ) = l · β ˜ i l R · ω ˙ r ( k ) / v long ( k ) A ( k , 2 ) = l F · ( v ˙ lat ( k ) + v long ( k ) · ω ˙ r ( k ) ) A ( k , 3 ) = v ˙ lat ( k )

Equation (15) is overdetermined and m/kR, Iy/kR can be solved in least square sense. We only take longitude velocity as input, and other measure parameters including lateral velocity, yaw rate, longitude acceleration, lateral acceleration are calculated using the estimation results from the previous frame.

It is noteworthy that in practice only the ratios between m and kR and Iy and kR are needed, rather than their individual values. Using only KITTI Sequence 01 for training, we obtained m/kR = 0.0073, Iy/kR = 0.0084. Empirically we assume that the vehicle mass equals to 1500 kg and the rear tire cornering stiffness is kR ≈ 205,480 N/rad, the moment of inertia around the y-axis is Iy ≈ 1726 kg·m2. The trained parameters obtained here are also used in the benchmark tests in Section 6.3.

The identification and test results are shown in Figure 7. All plots of Sequence 03, 06 and 09 demonstrate that when big turning occurred, side slip angle β calculated using the dynamic model is more accurate than ω/2, which is derived using only Ackerman Steering Principle. This characteristic will directly affect the estimation accuracy of the algorithm in sharp turning scenario.

4.4. Inliers Segmentation

As mentioned in Section 2.1, the model estimation R1 and t1 can be used to select inliers through thresholding the reprojection errors. Since MYP is essentially a model of 2 DoF, which requires two feature correspondences to compute a motion hypothesis. It can be implemented in either RANSAC scheme or a two-dimensional histogram. In this work we use the latter considering the deterministic of estimation results. The implementation of MYP can be concluded below.

(1)

For feature correspondence {iqk−1iqk}, {jqk−1jqk}, i = 1:N − 1, j = I + 1:N, solve ijω and ijγ using Equations (13) and (14) couple wise.

(2)

Conduct 2D histogram with specific resolution and range, vote with ijω and ijγ.

(3)

Select the winner model estimation {ω*, γ*} = argmax{histogram}.

(4)

Calculate β* according to Equation (11), as well as model estimation rotation matrix and translation vector:

R m = [ cos ω * 0 sin ω * 0 1 0 sin ω * 0 cos ω * ] · [ 1 0 0 0 cos γ * sin γ * 0 sin γ cos γ * ] t m = [ sin β * 0 cos β * ]

(5)

Triangulate {iqk−1iqk} with Rm and tm to obtain 3D estimation point {i}, then calculate reprojection error and distinguish inliers and outliers with some threshold.

5. Refinement

As a hypothesis generator in relative pose estimation based on hypothesis-test-validation framework, minimal set solving algorithms are not able to provide accurate solution because error exists in the coordinates of point correspondences. Consequently, all inliers that have been found should be used to obtain an accurate solution in the least square sense.

However, the winner of minimal solutions is not far from the true solution. It can work as a good starting point in some iterative non-linear algorithms for solving the minimization problem. In most optimization-based relative pose recovery algorithms, nine elements in the rotation matrix R are treated as independent variables, and the orthogonality of R cannot be preserved iteratively due to this inherent deficiency that rotation matrix has no explicit geometrical meaning. In monocular VO problem we can use three attitude angles as optimization variables, then the orthogonality problem will be natively raveled out. The computational efficiency of Euler angle parameter optimization seems low due to the trigonometric functions existing in the representation of R. Many efficient extreme search methods like conjugate gradient method, variable metric methods, and Levenberg-Marquardt method, etc., can work well, given explicit gradient expression of objective function. Then the cost function and its gradient function based on Euler angle parameterized motion expressions can be derived. Here, we refine R and t in turns, the method that compelled decoupling of R and t was also used by [26], and obtained promising result in this research.

6. Experiments

In this section, we compare our MYP algorithm with 1-point, MOBRAS and standard 5-point algorithm on both synthetic data and real car benchmark data.

6.1. CarSim Based Simulation Tests

First the inlier identification performance of our MYP algorithm was investigated with synthetic data generated by the joint simulation platform conducted using CarSim and Matlab in our previous work [23]. Real-time vehicle status parameters including velocity, pitch, roll and yaw angles were generated by CarSim [27]. Meanwhile, the positions were also recorded and used as ground truth.

In order to make the simulation more realistic, we set the simulation parameters as shown in Table 1. We varied the outlier percentage between 10% and 70% to compare the performance of mentioned algorithms in different levels of outliers.

We compared our MYP algorithm with the 1-point algorithm [12], MOBRAS algorithm [15], and standard 5-point RANSAC [17] using the generated synthetic data. We used 500 random iterations in the 5-point RANSAC method, which, according to the RANSAC statistics [1], should provide a probability of success of 99.42% (calculated assuming a fraction of outliers equal to 60%). We used the recommended value in [15] for the sampling number of MORRAS.

As defined in Section 3.3, sensitivity and specificity were used to evaluate the inlier identification performance. The resulting statistics are shown in Figures 8 and 9. Thick red bars denote the median of the errors. The higher border of the rectangles denote 75% percentiles while the lower borders represent the minimum value of the errors; the top end of the dash lines denotes 90% percentiles.

As Figure 8 illustrates, the inlier detection performance of motion model based methods (1-point, MOBRAS and MYP) were not affected by the percentage of outliers, but the detect ratio of 5-point algorithm decreased as the outlier percentage increased. It is determined by its nature that 5-point RANSAC is a probabilistic method. Whereas, model based methods grasp the main components of vehicle motion better. However, 5-point algorithm was the most robust method to the vehicle velocity when the outlier percentage was below 50%. Meanwhile the 1-point and MOBRAS algorithms were quite vulnerable to velocity level. They only worked well when the velocity was neither too low nor too high. Their performance confirmed that the ratio β/ω = 1/2 fits well when the vehicle moves at mid velocity around 50 km/h as shown in Figure 3.

Figures 10 and 11 show the accuracy of different inlier detection methods combing with the same final refinement. It is noteworthy that the absolute scale s of the translation vector is calculated by velocity vel and time interval ΔT as:

s = vel · Δ T
where ΔT equals to 0.1 s in the simulation as well as in real car test on the KITTI dataset.

The accuracy performance is consistent with the Sensitivity illustrated in Figures 8 and 9. Vehicle velocity variation did not affect the accuracy of standard 5-point RANSAC. When the vehicle velocity was around 50 km/h, 1-point algorithm as well as MORBAS that are based on kinematic models obtained the highest accuracy and the median values of their rotation errors exhibited a U-shaped distribution. The accuracy of MYP was slightly lower than 5-point RANSAC in high speed situations. On the other hand, when the percentage of outliers increased, 5-point RANSAC suffered the most while the motion model-based algorithms exhibited more robustness, wherein MYP obtained the best performance.

Table 2 shows the comparison of execution time among four monocular algorithms. The most efficient method was the 1-point algorithm, which on average took less than 2 ms. MOBRAS was more than 50 times slower than the 1-point algorithm because internally MOBRAS requires 100 1-point samplings. In our experiments, the 5-point RANSAC ran even slower than MOBRAS, around 23 times. This is consistent with the results presented in [25] which claimed that the 5-point RANSAC is 50 times slower than MOBRAS. Our MYP ran 10 times faster than MOBRAS and 240 times faster than the 5-point RANSAC. Referring to the accuracy performance shown in Figures 10 and 11, MYP algorithm had the highest performance over cost ratio.

6.2. Influence of Accuracy of the Velocity

We investigated the influence of the velocity accuracy on the performance of the proposed algorithm, since one of the inputs of MYP algorithm is the vehicle velocity. Using the same simulation platform as Section 6.1, different levels of velocity error were randomly generated and added to the input of the MYP algorithm.

As Figure 12 shows, the inlier detection performance of MYP was almost not affected by the velocity error (up to 10%). But according to Figure 13, the rotation error of MYP slightly increased as the velocity error increased. On the other hand, the translation error grew more significantly. It is true because the scale of translation is completely determined by velocity.

6.3. Test on Benchmark Data

We evaluated our MYP algorithm on the KITTI dataset, which consists of 22 video sequences. The raw image resolution of these videos is 1392 × 512 pixels, and the frame rate is on average 10 frames per second. Ground truth is provided for training Sequences 00–10. There is no ground truth for test Sequences 11–21. However, our MYP algorithm like other monocular methods requires the input of vehicle velocity. Hence we only ran on left images of Sequences 00–10 for training and the vehicle velocity was calculated using distances between positions.

We compared accuracy over Sequences 00–10. An overview of the average translation and rotation errors calculated over the eleven test sequences for three different algorithms are shown in Table 3. The results are sorted according to the average rotation error. According to this evaluation metrics, our MYP outperformed other methods.

In order to make the experiments more realistic, we also tested MYP algorithm with different velocity error on the same datasets. The results are shown in Table 4. MYP outperformed 5-point RANSAC even when velocity is measured with 5% error; MYP obtained almost the same performance as MOBRAS when velocity was error up to 10%.

A few exemplary estimated trajectories are shown in Figure 14. Of the three methods compared, MYP suffered the smallest drift. This is due to that the rotation error is the main source of drift and MYP has the smallest rotation estimation error as shown in Table 3. Another interesting finding was that both MYP and MOBRAS outperformed the 5-point RANSAC algorithm in Sequence 07. This was because in a complex scenario (marked by the yellow circle), high percentage of outliers caused by crossing vehicles undermined the estimation of the 5-point RANSAC algorithm. Figure 15 shows an example of a crossing vehicle as a moving obstacle that causes high outlier percentage and affects the performance of the 5-point RANSAC algorithm.

The execution time of our MYP algorithm pipeline including feature detection and tracking was around 60 ms. Our test platform used a PC with an Intel(R) Core(TM)2 CPU Q9500 at 2.83 GHz and 4 GB of RAM. The experimental results showed that our algorithm was capable of operating at the frame rate of 15 Hz, with the maximum execution times well below 100 ms. Our algorithm performed several times faster than other algorithms while maintaining better or at least comparable accuracy. Such a low latency system is well suited for real-time autonomous vehicles.

7. Conclusions

In this paper, we have presented a novel and fast visual odometry algorithm called MYP for wheeled vehicles moving in dynamic outdoor environments. The core of MYP is single track bicycle dynamic model and reasonable linear approximation based 2-DoF inlier selection method, which captures three most important motion components, namely yaw angle increment, side slip angle and pitch angle increment. Furthermore, MYP is fully deterministic compared to random sampling based RANSAC VO methods.

We successfully tested our algorithm on a large image dataset, spanning dozens of kilometers. Experiments showed that our MYP algorithm outperformed state-of-the-art monocular visual odometry approaches in terms of accuracy and efficiency. We also demonstrated that MYP can handle a variety of different types of moving obstacles, achieve accurate position and orientation estimations, and meet the localization and navigation requirements of intelligent vehicles in complex urban environments.

However, there is a limitation of MYP' application in higher speed situations. In the single track bicycle dynamic model, the tire stiffness characteristics exhibit good linearity. In this case, MYP can get good side slip angle fitting results when the vehicle velocity is not too high. In high-speed scenarios, typically higher than 70 km/h, the tire stiffness characteristics exhibit nonlinearity, which leads to performance degradation of the MYP algorithm. In our future work, we will extend MYP to high speed driving conditions by considering nonlinear vehicle dynamic models.

Acknowledgments

This work is supported by the National Natural Science Foundation of China (No. 91120010).

Author Contributions

Yanhua Jiang designed and implemented this new visual odometry algorithm for ground vehicles and prepares this manuscript. Guangming Xiong is the corresponding author, who is in charge of the entire project including idea generation, method development, and experiment design. Huiyan Chen advised Yanhua Jiang about the kinematic model and dynamic model of wheeled vehicles. D.J. Lee provided advice on visual technology and helped with editing the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Scaramuzza, D.; Fraundorfer, F. Visual Odometry Part I: The First 30 Years and Fundamentals. IEEE Robot. Autom. Mag. 2011, 18, 80–92. [Google Scholar]
  2. Moravec, H. Obstacle Avoidance and Navigation in the Real World by a Seeing Robot Rover. Ph.D. Thesis, University of Stanford, Stanford, CA, USA, 1980. [Google Scholar]
  3. Maimone, M.; Cheng, Y.; Matthies, Y.L. Two Years of Visual Odometry on the Mars Exploration Rovers. J. Field Robot. 2007, 24, 169–186. [Google Scholar]
  4. Taylor, C.J.; Kriegman, D.J. Structure and motion from line segments in multiple images. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 17, 1021–1032. [Google Scholar]
  5. Wong, K.-Y.K.; Mendonça, P.R.S.; Cipolla, R. Structure and motion estimation from apparent contours under circular motion. Image Vision Comput. 2002, 20, 441–448. [Google Scholar]
  6. Pradeep, V.; Lim, J. Egomotion using assorted features. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 1514–1521.
  7. Gauglitz, S.; Höllerer, T.; Turk, M. Evaluation of interest point detectors and feature descriptors for visual tracking. Int. J. Comput. Vision 2011, 3, 335–360. [Google Scholar]
  8. Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Gr. Image Process. 1981, 24, 381–395. [Google Scholar]
  9. Haralick, R.M.; Lee, C.N.; Ottenberg, K.; Nölle, M. Review and analysis of solutions of the three point perspective pose estimation problem. Int. J. Comput. Vision 1994, 13, 331–356. [Google Scholar]
  10. Nistér, D.; Naroditsky, O.; Bergen, J. Visual Odometry. Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Washington, DC, USA, 27 June–2 July 2004; pp. 652–659.
  11. Nourani-Vatani, N.; Roberts, J.; Srinivasan, M.V. Practical visual odometry for car-like vehicles. Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 3551–3557.
  12. Scaramuzza, D. 1-point-ransac structure from motion for vehicle-mounted cameras by exploiting non-holonomic constraints. Int. J. Comput. Vision 2011, 95, 74–85. [Google Scholar]
  13. Scaramuzza, D.; Fraundorfer, F.; Siegwart, R. Real-time monocular visual odometry for on-road vehicles with 1-point ransac. Proceedings of the IEEE International Conference on Robotics and Automation, Kobe, Japan, 12–17 May 2009; pp. 4293–4299.
  14. Zhu, M.; Ramalingam, S.; Taguchi, Y.; Garaas, T. Monocular Visual Odometry and Dense 3d Reconstruction for On-Road Vehicles. In Computer Vision–ECCV 2012 Workshops and Demonstrations; Springer Heidelberg: Berlin, Germany, 2012; pp. 596–606. [Google Scholar]
  15. Scaramuzza, D.; Censi, A.; Daniilidis, K. Exploiting motion priors in visual odometry for vehicle-mounted cameras with non-holonomic constraints. Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francisco, CA, USA, 25–30 September 2011; pp. 4469–4476.
  16. Hamme, D.V.; Veelaert, P.; Philips, W. Robust Monocular Visual Odometry by Uncertainty Voting. Proceedings of the IEEE Intelligent Vehicles Symposium, Baden-Baden, Germany, 5–9 June 2011; pp. 643–647.
  17. Nistér, D. An efficient solution to the five-point relative pose problem. IEEE Trans. Pattern Anal. Mach. Intell. 2004, 26, 756–770. [Google Scholar]
  18. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision, 2nd ed.; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  19. Hartley, R.I. In Defense of the Eight-Point Algorithm. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 580–593. [Google Scholar]
  20. Stéwenius, C.; Engels, H.; Nistér, D. Recent developments on direct relative orientation. ISPRS J. Photogramm. Remote Sens. 2006, 60, 284–294. [Google Scholar]
  21. Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle adjustment—A modern synthesis. In Vision Algorithms: Theory and Practice; Springer Heidelberg: Berlin, Germany, 2000; pp. 298–372. [Google Scholar]
  22. Mitschke, M. Dynamik der Kraftfahrzeuge: Band C: Fahrverhalten; Springer Heidelberg: Berlin, Germany, 1990. (in Germany) [Google Scholar]
  23. Jiang, Y.; Chen, H.; Xiong, G.; Gong, J.; Jiang, Y. Kinematic constraints in visual odometry of intelligent vehicles. Proceedings of the 2012 IEEE Intelligent Vehicles Symposium (IV), Alcala de Henares, Spain, 3–7 June 2012; pp. 1126–1131.
  24. Geiger, A. Kitti Vision Benchmark Suite. Available online: http://www.cvlibs.net/datasets/kitti/ (accessed on 22 May 2014).
  25. Eigen. Available online: http://eigen.tuxfamily.org/index.php?title=Main_Page (accessed on 15 October 2013).
  26. Oliensis, J.; Genc, Y. New algorithms for two-frame structure from motion. Proceedings of the IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 737–744.
  27. Carsim. Available online: http://www.carsim.com/ (accessed on 11 March 2012).
Figure 1. Motion estimation pipeline of two-frame feature based visual odometry algorithm.
Figure 1. Motion estimation pipeline of two-frame feature based visual odometry algorithm.
Sensors 14 16159f1 1024
Figure 2. Coordinate system definition.
Figure 2. Coordinate system definition.
Sensors 14 16159f2 1024
Figure 3. Plots of yaw angle increment ω (at 100 ms interval) and side slip angle β using CarSim simulated data. Velocity changes from 10 km/h to 70 km/h.
Figure 3. Plots of yaw angle increment ω (at 100 ms interval) and side slip angle β using CarSim simulated data. Velocity changes from 10 km/h to 70 km/h.
Sensors 14 16159f3 1024
Figure 4. Plots of yaw angle increment ω (at approximately 100 ms interval) and side slip angle β from the KITTI ground truth data at (a) high speed and (b) low speed.
Figure 4. Plots of yaw angle increment ω (at approximately 100 ms interval) and side slip angle β from the KITTI ground truth data at (a) high speed and (b) low speed.
Sensors 14 16159f4 1024
Figure 5. Plots of reprojection error against pitch angle increment and roll angle increment using synthetic data. (a) Increase the roll angle increment; (b) Increase the pitch angle increment.
Figure 5. Plots of reprojection error against pitch angle increment and roll angle increment using synthetic data. (a) Increase the roll angle increment; (b) Increase the pitch angle increment.
Sensors 14 16159f5 1024
Figure 6. Sensitivity and Specificity. Comparison between 1-Point and MOBRAS algorithms.
Figure 6. Sensitivity and Specificity. Comparison between 1-Point and MOBRAS algorithms.
Sensors 14 16159f6 1024
Figure 7. Fitted side slip angle using trained parameters: gt denotes ground truth.
Figure 7. Fitted side slip angle using trained parameters: gt denotes ground truth.
Sensors 14 16159f7 1024
Figure 8. Sensitivity and specificity of true inliers detection vs. outlier percentage. Vehicle velocity is fixed at 50 km/h.
Figure 8. Sensitivity and specificity of true inliers detection vs. outlier percentage. Vehicle velocity is fixed at 50 km/h.
Sensors 14 16159f8 1024
Figure 9. Sensitivity and specificity of true inliers detection vs. vehicle velocity. Outlier percentage is set to 50%.
Figure 9. Sensitivity and specificity of true inliers detection vs. vehicle velocity. Outlier percentage is set to 50%.
Sensors 14 16159f9 1024
Figure 10. Accuracy of final motion estimation vs. outlier percentage. Vehicle velocity is fixed at 50 km/h.
Figure 10. Accuracy of final motion estimation vs. outlier percentage. Vehicle velocity is fixed at 50 km/h.
Sensors 14 16159f10 1024
Figure 11. Accuracy of final motion estimation vehicle velocity level. Outlier percentage is set to 50%.
Figure 11. Accuracy of final motion estimation vehicle velocity level. Outlier percentage is set to 50%.
Sensors 14 16159f11 1024
Figure 12. Sensitivity and specificity of true inliers detection vs. vehicle velocity. Outlier percentage is set to 50%, velocity error varies from 0 to 10%.
Figure 12. Sensitivity and specificity of true inliers detection vs. vehicle velocity. Outlier percentage is set to 50%, velocity error varies from 0 to 10%.
Sensors 14 16159f12 1024
Figure 13. Rotation and translation error vs. vehicle velocity. Outlier percentage is set to 50%, velocity error varies from 0 to 10%.
Figure 13. Rotation and translation error vs. vehicle velocity. Outlier percentage is set to 50%, velocity error varies from 0 to 10%.
Sensors 14 16159f13 1024
Figure 14. Exemplary estimated trajectories: Comparison between MYP, MOBRAS and 5-point RANSAC.
Figure 14. Exemplary estimated trajectories: Comparison between MYP, MOBRAS and 5-point RANSAC.
Sensors 14 16159f14 1024
Figure 15. Moving obstacle causes complex situation with high outlier percentage. Yellow lines denote false feature correspondences. (a) Inliers preserved by 5-point RANSAC; (b) Inliers preserved by MYP.
Figure 15. Moving obstacle causes complex situation with high outlier percentage. Yellow lines denote false feature correspondences. (a) Inliers preserved by 5-point RANSAC; (b) Inliers preserved by MYP.
Sensors 14 16159f15 1024
Table 1. Simulation parameters.
Table 1. Simulation parameters.
ParametersValue
Simulation time step100 [ms]
Expectation velocity30:10:70 [km/h]
Outlier percentage10:10:70 [%]
Std of gaussian noise add to image points σ0.5 [pixel]
Reprojection error threshold1.0 [pixel]
Iteration number of 5-point alogrithm500
Sampling number of MOBRAS100
Table 2. Execution time of outlier rejection process.
Table 2. Execution time of outlier rejection process.
AlgorithmMean [ms]Median [ms]Min [ms]Max [ms]
5-point RANSAC2106.31936.94553513.8
MOBRAS909161108
1-point algorithm1.91.81.52.9
MYP8.68.76.410.6
Table 3. Overview of average translation and rotation error on KITTI Sequences 00–10.
Table 3. Overview of average translation and rotation error on KITTI Sequences 00–10.
MethodRotationTranslation
MYP0.0145 [deg/m]1.32%
5-Point RANSAC0.0161 [deg/m]1.36%
MOBRAS0.0172 [deg/m]1.39%
Table 4. Average performance of MYP against velocity error.
Table 4. Average performance of MYP against velocity error.
Velocity ErrorRotationTranslation
0%0.0145 [deg/m]1.32%
5%0.0156 [deg/m]1.45%
10%0.0173 [deg/m]1.65%

Share and Cite

MDPI and ACS Style

Jiang, Y.; Xiong, G.; Chen, H.; Lee, D.-J. Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments. Sensors 2014, 14, 16159-16180. https://doi.org/10.3390/s140916159

AMA Style

Jiang Y, Xiong G, Chen H, Lee D-J. Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments. Sensors. 2014; 14(9):16159-16180. https://doi.org/10.3390/s140916159

Chicago/Turabian Style

Jiang, Yanhua, Guangming Xiong, Huiyan Chen, and Dah-Jye Lee. 2014. "Incorporating a Wheeled Vehicle Model in a New Monocular Visual Odometry Algorithm for Dynamic Outdoor Environments" Sensors 14, no. 9: 16159-16180. https://doi.org/10.3390/s140916159

Article Metrics

Back to TopTop