Next Article in Journal
Unmasking of Olive Oil Adulteration Via a Multi-Sensor Platform
Next Article in Special Issue
An Improved WiFi Indoor Positioning Algorithm by Weighted Fusion
Previous Article in Journal
Label-Free Biosensor Imaging on Photonic Crystal Surfaces
Previous Article in Special Issue
Heading Estimation for Indoor Pedestrian Navigation Using a Smartphone in the Pocket
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Probabilistic Feature Map-Based Localization System Using a Monocular Camera

1
Urban Robotics Laboratory (URL), Korea Advanced Institute of Science and Technology (KAIST), 291 Daehak-ro (373-1 Guseong-dong), Yuseong-gu, Daejeon 305-701, Korea
2
Ocean System Engineering Research Division, Korea Research Institute of Ships and Ocean Engineering (KRISO), 32 1312 Beon-gil, Yuseong-daero, Yuseong-gu, Daejeon 305-343, Korea
*
Author to whom correspondence should be addressed.
Sensors 2015, 15(9), 21636-21659; https://doi.org/10.3390/s150921636
Submission received: 27 April 2015 / Revised: 23 August 2015 / Accepted: 24 August 2015 / Published: 31 August 2015
(This article belongs to the Special Issue Sensors for Indoor Mapping and Navigation)

Abstract

:
Image-based localization is one of the most widely researched localization techniques in the robotics and computer vision communities. As enormous image data sets are provided through the Internet, many studies on estimating a location with a pre-built image-based 3D map have been conducted. Most research groups use numerous image data sets that contain sufficient features. In contrast, this paper focuses on image-based localization in the case of insufficient images and features. A more accurate localization method is proposed based on a probabilistic map using 3D-to-2D matching correspondences between a map and a query image. The probabilistic feature map is generated in advance by probabilistic modeling of the sensor system as well as the uncertainties of camera poses. Using the conventional PnP algorithm, an initial camera pose is estimated on the probabilistic feature map. The proposed algorithm is optimized from the initial pose by minimizing Mahalanobis distance errors between features from the query image and the map to improve accuracy. To verify that the localization accuracy is improved, the proposed algorithm is compared with the conventional algorithm in a simulation and realenvironments.

1. Introduction

Image-based localization is an important issue in robotics communities as well as computer vision communities. Its applications include navigation of robots or pedestrians, virtual reality, and visualization of tourism cites [1]. Since Google Street view [2] and other tourism sites data sets [3] such as Dubrovnik, Rome, and Vienna are provided, image-based localization has been widely studied and the performance of the suggested approaches has been demonstrated with these data sets. As it is expected to offer enormous image data sets to the entire world on the Internet, image-based localization can be utilized for global localization as a replacement of GPS (global positioning system). As cameras are extensively equipped in everyday electronic devices such as mobile phones, image-based localization is becoming increasingly important in various fields.
In the field of computer vision, Robertson et al. [4] proposed an image-based localization method by employing an urban navigation system. They utilized a database of rectified views of building facades by extracting edges of buildings and roads to estimate the pose of a query image. Similarly, Kosecka et al. [5] introduced an indoor image-based localization method by matching histograms that are generated by detecting edges of room images. In the robotics field, Nepier et al. [6] estimated a robot’s pose by matching road information of a query image to synthetic orthographic images of the road surface produced by a stereo vision system in advance.
color1As robust features have been developed such as SIFT (scale-invariant feature transform) [7] and SURF (speeded up robust features) [8], many researchers are trying to apply the robust features for localization. Zhang et al. [9] proposed an image-based localization method in urban environments using triangulation of matched features from database images. 3D feature map-based localization methods are introduced by matching features between a query image and the 3D map which is generated from image databases [10,11]. The accuracy of the feature map is important in feature-based localization methods because errors of the map directly influence the accuracy of localization. In response, the SfM (structure-from-motion) technique [12] was proposed to estimate camera’s motion and scene’s structures from camera images while improving the accuracy of 3D feature maps. color1Bundle Adjustment [13] is one of the widely used optimization methods for minimizing residual errors of the SfM approach. With recent developments of SfM techniques, an SfM model can be constructed on a city-scale considering millions of points [14,15,16]. Accordingly, localization methods based on a 3D feature map such as the SfM model have been widely researched as it has become possible to cover large areas. In large environments, one of the crucial issues is the computation time because the 3D feature map contains millions of features. Lu et al. [10] improved computation time by employing kd-tree and nearest neighbor (FLANN; fast library for approximate nearest neighbors) algorithms to find a query feature’s correspondences. Sattler et al. [17] applied feature-based localization to a mobile phone platform and suggested a method to accelerate the matching process. Other studies of localization based on a feature map to enhance the accuracy of matching correspondences have also been reported. Irschara et al. [18] enhanced registration performance using relevant fragments of a 3D model. Other researchers also improved the performance of 3D-to-2D correct correspondences using mutual visibility information [3] and query expansion [19].
According to [14,17,18], localization methods based on a feature map solve the real-time problem while also providing high performance of matching on a large scale map. However, these studies assume that the map has high accuracy and sufficient features. It therefore remains challenging to apply the feature-based localization methods with insufficient and uncertain features. Since the most feature-based localization methods [3,9,10,11,12,13,14,15,16,17,18,19] use camera pose estimation methods [20,21,22,23,24], uncertainty of features is not considered during the camera pose estimation. color1Although Ferraz et al. [25] proposed a camera pose estimation method considering features’ uncertainties by estimating the uncertainties from a bunch of feature points, this method is unsuitable to be employed in a mobile robot application because it is usually not easy to obtain many features of the same object from various views. Thus, a novel localization method is needed to enhance localization accuracy by considering uncertainty of the feature map for mobile robot system during the camera pose estimation.
The uncertain feature map is widely used in SLAM (Simultaneous localization and mapping) techniques in robotics communities. The camera-based SLAM techniques [26,27,28,29] generate a probabilistic feature map based on a sensor system modeling and estimate the robot’s pose at the same time. Since the monocular camera is not able to estimate the full states of the robot’s pose at a single observation, it is required to incorporate additional information such as the robot’s prediction [30] or a scale-known landmark [29,31,32]. There are also researches to estimate the robot’s pose using a pre-built map from the SLAM algorithm, which is called a robot relocation problem. Jeong et al. [33] proposed a relocation method of a cleaning robot installed with an upward-looking camera using Harris corners and their orientation information from a ceiling and side walls. This algorithm estimates the robot’s pose using Hough clustering [34] exploiting orientation information of features. Since the uncertainty of the individual feature is not considered, large uncertainties of the features directly influence the accuracy of localization. Lee et al. [35] proposed a recursive pose estimation algorithm in kidnapping situation of a cleaning robot using EKF (Extended Kalman Filtering)-based SLAM. The robot’s pose is estimated by recursive EKF updates of the corresponding past robot pose of the best feature-matched frame with a current frame. However, it is hard to estimate an optimized position since this algorithm just matches features based on the single frame. Moreover, these algorithms are designed based on the assumption that the robot is moving on a flat ground, so the robot’s pose can only be estimated in 3-DoF (x, y, yaw).
Most feature map-based localization methods are performed in a feature map which has sufficient number of features. However, it is challenging to apply these methods in real environment because the feature map from real environment also contains some areas which have inaccurate and insufficient features. The main contribution of this paper is the suggestion of the novel image-based localization method that can cover extensive areas containing inaccurate and insufficient features. In addition, as the proposed method is able to estimate 3D camera pose, it can deal with a relocation situation in a 3D SLAM system. In our approach, a probabilistic feature map is generated from an insufficient data set using sensor modeling. An initial camera pose is estimated by the PnP (Perspective-n-Point) algorithm [20,22] using 3D-to-2D matching correspondences between the map and the query image. The camera pose is further enhanced by minimizing Mahalanobis distance error between the matching correspondences on its image plane.
The rest of this paper is structured as follows. Section 2 explains the method of generating a probabilistic feature map. In Section 3, the proposed localization method is explained in detail. To verify the effectiveness of the proposed method, the results of simulations and experiments are shown in Section 4. In Section 5, a conclusion and directions for future work are provided.

2. Generation of Probabilistic Feature Map

To generate a probabilistic feature map, camera-based sensor data are necessary in advance. If the sensor data and their corresponding poses are provided, a 3D map can be constructed. After features are extracted from each image, it is possible to express a probabilistic feature map in the 3D space by referring to its sensor system modeling. Any camera-based sensor system such as a stereo camera, ToF (time-of-flight) camera, and Kinect can be modeled. In this section, the method to generate a 3D probabilistic feature map is introduced using a camera based on bearing-only landmark initialization [36,37].

2.1. Definition of Probabilistic Feature Map

The probabilistic feature map has m features that are assumed to have Gaussian distributions comprising position and covariance values. The probabilistic feature map is expressed on the global coordinates of the 3D space as follows:
Map = [ F 1 , F 2 , , F m ]
F i = [ X f i , C f i ] ( i = 1 , , m )
where F i is the i-th feature composed of a position, X f i R 3 , and its covariance, C f i R 3 × 3 , and m is the total number of features. The feature also contains additional information for feature matching such as a descriptor vector and an index of the image. The method of constructing the probabilistic feature map is explained below.

2.2. Probabilistic Representation of Features

A probabilistic representation of features is related to a sensor system modeling. Although various image-based sensor systems are available for use, this paper deals with a stereo vision system without loss of generality. Figure 1 shows the steps of probabilistic representation considering the stereo vision system. Let us consider a point on the 3D space represented by X = ( x , y , z ) T in Cartesian coordinates and X ˜ = ( ρ , θ , ϕ ) T in spherical coordinates. The covariance of each coordinate system is represented by C and C ˜ as follows:
C = σ x 2 σ x y 2 σ x z 2 σ x y 2 σ y 2 σ y z 2 σ x z 2 σ y z 2 σ z 2 , C ˜ = σ ρ 2 σ ρ θ 2 σ ρ ϕ 2 σ ρ θ 2 σ θ 2 σ θ ϕ 2 σ ρ ϕ 2 σ θ ϕ 2 σ ϕ 2
where each element denotes the covariance between corresponding axes. By measuring intrinsic parameters and the baseline of two cameras, the position of a feature in the 3D space can be calculated as X 1 and X ˜ 1 on each coordinate system by a triangulation algorithm, as shown in Figure 1a. For estimating the covariance based on the sensor system modeling, a spherical coordinate system is employed, as shown in Figure 1b. As the relationship of covariances between these two coordinate systems is expressed by a Jacobian transformation, the covariance of the spherical coordinate system at X 1 can be converted to the Cartesian coordinate system as follows (refer to Figure 1b):
C 1 = J F ( ρ , θ , ϕ ) C ˜ 1 J F ( ρ , θ , ϕ ) T
where J F ( ρ , θ , ϕ ) is the Jacobian matrix at X ˜ 1 for converting the spherical coordinate system to the Cartesian coordinate system; ρ, θ, and ϕ denote the radial distance, inclination, and azimuth, respectively. Since the covariance of features on the 3D space is influenced by the depth information, C ˜ 1 is formulated as follows:
C ˜ 1 = z 2 σ 1 2 0 0 C ¯ 1
where σ 1 is the standard deviation of the depth error and C ¯ 1 is a submatrix of C ˜ 1 regarding to θ and ϕ. As the uncertainty in the radial direction on the spherical coordinate system is dependent on the depth value, the covariance of radial distance is set to z 2 σ 1 2 according to the sensor modeling of a stereo vision system [38]. C ¯ 1 is influenced by pixel errors of the feature on an image. rcolor1 X ˜ 2 and C ˜ 2 are the projected point of X 1 and its covariance on a virtual image plane where z = 1 , as shown in Figure 1b. Since the inclination and azimuth of X ˜ 1 and X ˜ 2 are same, the inclination and azimuth of C ˜ 1 are also same as the ones of C ˜ 2 . Therefore, C ¯ 1 is equal to the inclination and azimuth covariance of C ˜ 2 which is the covariance at X ˜ 2 . As the uncertainty of features on the image is usually set to a fixed value, the covariance of features on the virtual image plane is expressed as
C 2 = diag ( σ 2 2 , σ 2 2 , 0 )
where σ 2 is the standard deviation of the pixel value. The uncertainty in z-axis is set to zero because the virtual image plane does not have z-axis values. C ¯ 1 is converted from C 2 by approximation of the Jacobian matrix on the virtual image plane as follows:
C ¯ 1 = B J F ( x , y , z ) C 2 J F ( x , y , z ) T B T
B = 0 1 0 0 0 1
where J F ( x , y , z ) is a Jacobian matrix at X 2 for converting Cartesian to spherical coordinates on the virtual image plane. To obtain a sub-matrix for the inclination and azimuth, matrix B is applied.
Through Equations (3) to (7), the covariance of features from Camera 1, C 1 in Figure 1b, is estimated. Since the camera pose also contains uncertainty, the uncertainties of the sensor measurement and the camera pose, C 1 and C 3 in Figure 1c, are combined by compound approximate transformations [39] as follows:
C 4 = T 1 C 3 T 1 T + C 1
T 1 = 1 0 0 0 z 1 - y 1 0 1 0 - z 1 0 x 1 0 0 1 y 1 - x 1 0
where T 1 denotes approximate transformations to uncertainty of the 3D feature point, X 1 , from the covariance of the camera pose, C 3 . color1 x 1 , y 1 , and z 1 denote the coordinates of the 3D feature point, X 1 . Thus, the covariance from Camera 1 is represented as C 4 as in Equation (9). Let the rotational components (roll, pitch, and yaw) of the 3D pose be defined as θ roll , θ pitch , and θ yaw . C 3 can be generally set to diag ( σ x 2 , σ y 2 , σ z 2 , σ θ roll 2 , σ θ pitch 2 , σ θ yaw 2 ) from the uncertainty of each element (x, y, z, roll, pitch, and yaw).
Figure 1. Generation steps of probabilistic features in a stereo system. (a) Matched feature points on each image plane; (b) Conversion from Cartesian coordinate to spherical coordinate system; (c) Estimating covariance from a single camera; (d) Estimating initial covariance of the stereo system by merging covariance of each camera.
Figure 1. Generation steps of probabilistic features in a stereo system. (a) Matched feature points on each image plane; (b) Conversion from Cartesian coordinate to spherical coordinate system; (c) Estimating covariance from a single camera; (d) Estimating initial covariance of the stereo system by merging covariance of each camera.
Sensors 15 21636 g001
By estimating the covariance of features, the features can now be expressed as Gaussian distributions. Since there are two Gaussian distributions of the features from two cameras in the stereo system, a merging method of multivariate Gaussian distributions is required for estimating one Gaussian distribution for a single observation from this system. When two Gaussian distributions are N ( μ 1 , Σ 1 ) and N ( μ 2 , Σ 2 ) , the merged Gaussian distribution, N ( μ 3 , Σ 3 ) , can be estimated using the merging method [39] as follows:
Σ 3 = ( Σ 1 - 1 + Σ 2 - 1 ) - 1
μ 3 = Σ 3 ( Σ 1 - 1 μ 1 + Σ 2 - 1 μ 2 )
In Figure 1d, C 4 and C 5 are covariances of the features from Camera 1 and 2, respectively. C 5 seen from Camera 1 is represented as C 5 = R b C 5 R b T where R b is the rotation matrix from Camera 1 to Camera 2. Letting the probabilistic feature from each camera be N ( μ 1 = X 1 , Σ 1 = C 4 ) and N ( μ 2 = X 1 , Σ 2 = R b C 5 R b T ) in Equations (11) and (12), the probabilistic feature is estimated from a single observation of the stereo vision system, as shown in Figure 1d. Through this process, the position and covariance of each feature from a single observation, denoted by X o and C o , respectively, can be obtained.

2.3. Constructing a Probabilistic Feature Map

To construct a probabilistic feature map, the features from a single observation require transformation to the global coordinate system and merging of the same features from other observations. color1According to [40], a probabilistic feature from a single observation can be transformed to the global coordinate system as follows:
F o = [ R o X o + t o , R o C o R o T ]
where R o and t o are rotation and translation of Camera 1 to the global coordinate system.
After all features are converted to the global coordinate system, the probabilistic feature map is constructed using clustering and merging methods of the features. The clustering method has two process steps for recognition of identical features. The first process is the feature descriptor matching. color1The descriptor matching method is dependent on the feature extraction algorithm, and the approximate best-bin-first search method [41] is employed as a matching method for SURF and SIFT descriptors. However, feature matching includes many outliers. Thus, the geometric constraint is employed to remove outliers of matching in the next step. To check the geometric constraint of each feature, the Bhattacharyya distance metric [42] is used by measuring the geometric distance between the probabilistic representations of two features. As the Bhattacharyya distance measures a distance between two probability distributions, the Bhattacharyya distance for two multivariate Gaussian distributions, N ( μ 1 , Σ 1 ) and N ( μ 2 , Σ 2 ) , is
D B = 1 8 ( μ 1 - μ 2 ) T Σ - 1 ( μ 1 - μ 2 ) + 1 2 ln | Σ | | Σ 1 | | Σ 2 |
Σ = Σ 1 + Σ 2 2
where | · | denotes determinant. If the Bhattacharyya distance of two features is lower than a certain threshold, those features are regarded as the same features. After finding the same features by checking the descriptor matching and the geometric constraint, probabilistic features are estimated by the merging method described in Equations (11) and (12). Through the merging method of the same features, the probabilistic feature map is constructed.

3. Localization Method Using Probabilistic Feature Map

Using prior collected data, the probabilistic feature map can be generated in advance. With this pre-built map, the localization problem is solved using only a monocular camera. The localization of the camera is performed by matching a query image of the camera to the probabilistic feature map.

3.1. Generation of Matching Correspondences

It is difficult to find matching correspondences between a query image and the map with robustness, speediness, and accuracy in a large scale area. Many studies have attempted to solve this problem [1,10,17,18,19]. This paper therefore does not deal with matching correspondences deeply and instead applies a simple matching method.
To find matching correspondences between the query image and the probabilistic feature map, features are extracted from the query image. First, candidates of matching correspondences are generated from matched features between the query image and the feature map by descriptor matching. Although descriptor matching of features using SURF or SIFT is still reliable, there are many outliers due to the existence of identical descriptors with different features. Thus, the epipolar constraints-based RANSAC algorithm [43] is utilized to remove mismatched correspondences. As a result, the matching correspondence of features on the query image and the map is expressed as follows:
pairs = [ q j , F j ] ( j = 1 , , n )
where q j R 2 denotes the j-th feature position from a query image and F j R 3 denotes the matched probabilistic feature consisting of its position X f j and covariance C f j , and n is the total number of the matching correspondences.

3.2. Projection of Probabilistic Feature onto Image Plane

The probabilistic features on the map can be projected onto the image plane by utilizing intrinsic and extrinsic parameters of the camera, as shown in Figure 2a. A point in the 3D space can be easily projected to the 2D image plane [43] whereas the covariance in the 3D space cannot. Since the projection onto the image plane is a non-linear transformation, the Jacobian approximation method is employed to project the covariance, in a similar manner to that described in Equations (4) and (7) in Section 2.
Figure 2. (a) The projection results of probabilistic features on the map; (b) The projection of the probabilistic feature on the map onto the image plane using spherical coordinates.
Figure 2. (a) The projection results of probabilistic features on the map; (b) The projection of the probabilistic feature on the map onto the image plane using spherical coordinates.
Sensors 15 21636 g002
The projection process of a probabilistic feature is shown in Figure 2b. When the camera pose is located at the origin, X f = ( x , y , z ) T and X ˜ f = ( ρ , θ , ϕ ) T are easily projected into the virtual image plane as follows:
X f = ( x , y , z ) T X p = ( x p , y p , z p ) T = x z , y z , 1 T
X ˜ f = ( ρ , θ , ϕ ) T X ˜ p = ( ρ p , θ p , ϕ p ) T = ρ z , θ , ϕ T
where X p and X ˜ p are the corresponding points on the virtual image plane in Cartesian and spherical coordinates, respectively; ρ denotes the distance between the camera origin and X f ; and ρ p denotes the distance between the camera origin and X p as shown in Figure 2b. When projected onto the image plane, the inclination and azimuth are not changed in spherical coordinates. Using this property, the covariance in the 3D space can be projected onto the image plane approximately. The projection steps for the covariance of a feature are C f C ˜ f C ˜ p C p . First, the 3D covariance of a feature on the Cartesian coordinate system is converted into the spherical coordinate system by using a Jacobian matrix as follows:
C ˜ f = J F ( x , y , z ) C f J F ( x , y , z ) T
where C f and C ˜ f are covariance matrices on the Cartesian and spherical coordinates; and J F ( x , y , z ) is a Jacobian matrix for converting Cartesian to spherical coordinates at X f . In the projection step, C ˜ f C ˜ p , in the spherical coordinate system, it can be seen that C ˜ p = C ˜ f since the inclination and azimuth are not changed in the spherical coordinate system and the radial distance can be neglected due to the fact that depth values of points in the image plane is unity ( z = 1 ). Next, the covariance in spherical coordinates is converted to Cartesian coordinates on the virtual image plane as follows:
C p = J F ( ρ , θ , ϕ ) C ˜ p J F ( ρ , θ , ϕ ) T
where J F ( ρ , θ , ϕ ) is a Jacobian matrix at X ˜ p . As X p and C p denote the position and the covariance on the virtual image plane in the 3D space, they are converted to image coordinates by utilizing the intrinsic parameter of the camera as follows [43]:
X ¯ p = B K X p , C ¯ p = B K C p ( B K ) T
B = 1 0 0 0 1 0
where K denotes the intrinsic parameter of a camera. To obtain sub-matrices of X p and C p for x and y, the matrix B is applied.
Through these steps, the points and covariances of features in the map are projected onto the image. Let us define the projection operator P of a point and covariance as follows:
P ¯ f = ( X ¯ p , C ¯ p ) = P ( X f , C f )
where ( X f , C f ) R 3 × R 3 × 3 and ( X ¯ p , C ¯ p ) R 2 × R 2 × 2 denote probabilistic features on the 3D map and on the image, respectively. Therefore, all probabilistic features on the 3D map can be represented probabilistically on a virtual image plane through the projection operator P without considering the camera’s extrinsic parameter.

3.3. Estimating Camera Pose Based on Probabilistic Map

The camera pose of the query image is estimated using 3D-to-2D matching correspondences. After features are extracted from the query image, correspondences between the probabilistic feature map and the query image are acquired by Equation (16). Once color13D-to-2D matching correspondences are offered, the PnP algorithm [20,21] which is one of the most widely used algorithms for the camera pose estimation, is employed for comparison with the proposed algorithm. The PnP algorithm is also used for the initial point generation of the proposed algorithm. The PnP algorithm estimates R and t of the camera pose by minimizing the error of 3D-to-2D matching correspondences from the relation equation as follows:
{ R , t } = arg min R , t j = 1 n e j ( q j , K R X f j + t )
where q j and X f j are defined in Equation (16), K is the camera’s intrinsic parameter, e j denotes the distance error of each pair, n is the total number of the matching correspondences. The proposed algorithm considers the probabilistic information of the map during localization by minimizing Mahalanobis distance errors between the correspondences. Since a point on the image only represents a ray on the 3D space, it is difficult to employ the Mahalanobis distance in the 3D space. Therefore, we propose a method to employ the Mahalanobis distance on a 2D image where the probabilistic features are projected. Since the camera pose is required for performing projection of probabilistic features onto the 2D image, the initial pose is necessary. Thus, the conventional PnP algorithm is utilized for the initial pose estimation.
All matched features from the probabilistic feature map, F j from Equation (16) can be projected onto the image plane by Equation (23). However, Equation (23) does not consider the camera’s pose. Considering the camera pose, probabilistic features are projected onto the image plane at certain R and t as follows:
P ¯ F j | R , t = ( X ¯ p , C ¯ p ) | R , t = P ( R T X f j - R T t , R T C f j R )
where P ¯ F j | R , t indicates the projected probabilistic feature on the image. Using Equations (16) and (25), the Mahalanobis distance, D M , on the image plane between the projected probabilistic feature at certain R and t and the extracted feature from the query image is expressed as:
D M ( q j , P ¯ F j | R , t ) = ( q j - X ¯ p | R , t ) T ( C ¯ p | R , t ) - 1 ( q j - X ¯ p | R , t )
Minimizing the Mahalanobis distance errors for all correspondences, the optimal solution can be obtained as follows:
{ R , t } = arg min R , t 1 N j = 1 N min ( D M ( q j , P ¯ F j | R , t ) , τ )
where F j and q j are the matching correspondences from Equation (16); and τ denotes a certain threshold to restrict the maximum Mahalanobis distance. color1Although mismatched pairs are rejected using the epipolar constraints-based RANSAC algorithm [43], there might still be mismatches. The localization accuracy is improved by the maximum restricted Mahalanobis distance that reduces influence from outlier matching during optimization. Since Equation (27) is a nonlinear problem with respect to R and t, various nonlinear optimization algorithms can be employed. Since it is important to set the initial point in the nonlinear optimization, the estimation from the conventional PnP method is utilized.

4. Simulation and Experiments

The proposed algorithm is validated through comparison with the conventional algorithm in simulations and real experiments. The proposed algorithm is demonstrated in a virtual probabilistic map in the simulation environment. Through testing in the real environment, the robustness of the proposed algorithm is also verified.

4.1. Simulation

The camera pose is estimated using the proposed algorithm based on the probabilistic feature map. It is assumed that there are feature maps expressed by Gaussian distributions in a virtual 3D space. The features are randomly generated in the virtual 3D space sized 50 m × 50 m × 10 m, similar to a real test environment. Their covariances are set randomly from 1 m to 3 m for each axis. The circular path for a 6-DoF (Degree-of-freedom) camera pose is generated with diameter of 25 m. Figure 3 shows the top view of the virtual environment. The blue ellipses denote the randomly generated covariances of features. The red line is the trajectory of the camera pose. Considering the intrinsic camera parameter and image size, the feature map is projected to the image plane at the virtual path. The feature map contains Gaussian noises based on the covariance of each feature. The features on the image plane have fixed Gaussian noises considering pixel values. 10% of mismatching is also implemented to the matching correspondence between the map and the image plane, similar to real environments.
Figure 3. Top view of the simulation environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Figure 3. Top view of the simulation environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Sensors 15 21636 g003
As conventional PnP methods, P3P [20] and OPnP [22] methods are employed as a traditional and state-of art methods using these matching correspondences between features on the map and the image plane, respectively. The proposed algorithm minimizes errors between the probabilistic map features and extracted features on the image plane. The active-set algorithm [44] is employed as a nonlinear optimization method in the proposed algorithm. Since it is important to set an initial point in the nonlinear optimization method, the result of the conventional PnP method is set as the initial point for optimization. Figure 4a shows the top view of the simulation results. The Euclidean distance errors for the PnP and the proposed algorithms are shown in Figure 4b. The mean and standard deviation of the error for each 6-DoF are shown in Table 1.
Figure 4. (a) Top view of the simulation results. Red triangles, green triangles, and blue circles denote results of PnP, OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the ground truth data.
Figure 4. (a) Top view of the simulation results. Red triangles, green triangles, and blue circles denote results of PnP, OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the ground truth data.
Sensors 15 21636 g004
Table 1. Comparison of the error statistics of the conventional and the proposed algorithms in the simulation environment (Unit: m, deg).
Table 1. Comparison of the error statistics of the conventional and the proposed algorithms in the simulation environment (Unit: m, deg).
P3P AlgorithmOPnP AlgorithmProposed Algorithm
MeanStdevMeanStdevMeanStdev
x0.6340.3530.4860.1320.2920.085
y0.7140.3510.6650.3070.2790.076
z0.8010.4440.6630.2450.7060.368
θ roll 1.3681.7950.9810.7051.2271.569
θ pitch 1.5272.1411.2921.4931.2730.658
θ yaw 1.0700.9341.1880.7740.4930.257
As can be seen in Table 1, the accuracy of the proposed algorithm seems to be higher than that of the conventional PnP algorithm. x, y, and θ yaw in the proposed algorithm especially have better performance because the simulation is made for a mobile robot moving on a flat ground. To be clear, a paired t-Test is performed to analyze the performance statistically. As the p-value is 1 . 1527 × 10 - 57 , the superiority of the proposed algorithm is assured.

4.2. Experiment in Indoor Environment

To validate the performance of the proposed algorithm, experiments with a mobile robot were performed. The mobile robot system, Pioneer 3-AT model [45], is equipped with a stereo vision, Bumblebee XB3 model [46], and a marker system for the ground truth as shown in Figure 5. The XB3 provides the depth value of each pixel. Thus, the sensor produces a 2D image as well as per-pixel depth data with 1920 × 1080 resolution. A camera is installed on the ceiling for measuring the ground truth pose of the mobile robot as shown in Figure 6a and its sample captured image is shown in Figure 6b. The ground truth system is only able to estimate 3-DoF pose (x, y, θ yaw ). Thus, other elements of robot pose (z, θ roll , θ pitch ) for ground truth are set to zero because the mobile robot is assumed to move on the flat ground. As the resolution of ceiling camera is 640 × 480 pixels and the camera covers 4.4 m × 3.3 m area, the resolution of the ground truth system is about 0.7 cm per pixel. The indoor experiment was performed in the hall as shown in Figure 7. The trajectory of the mobile robot was composed of a circle with a diameter of 2.0 m repeated 5 times. The probabilistic feature map is generated from the first and second laps of the experiment and the proposed localization is demonstrated from third to fifth laps using the pre-generated probabilistic feature map.
Figure 8 shows the experimental environment where the probabilistic feature map was generated in advance. The shapes of covariance ellipses are mostly narrow since the features were observed only a few times. Figure 9a shows the top view of the results in the indoor experiment. Euclidean distance errors from the ground truth for each algorithm are shown in Figure 9b. The error results of the 6-DoF robot’s pose are presented in Table 2. Similar to the simulation results, the accuracy of the proposed algorithm is higher than that of the conventional algorithm, particularly for the x, y, and θ y a w values. As the p-value of the paired t-Test is 1 . 5353 × 10 - 34 , the superiority of the proposed method is confirmed statistically.
Figure 5. A mobile robot system equipped with a stereo camera. A patterned marker is used for providing ground truth.
Figure 5. A mobile robot system equipped with a stereo camera. A patterned marker is used for providing ground truth.
Sensors 15 21636 g005
Figure 6. Ground truth system based on global vision sensor. (a) Installed camera on the ceiling for the ground truth system; (b) The image processing result for ground truth system.
Figure 6. Ground truth system based on global vision sensor. (a) Installed camera on the ceiling for the ground truth system; (b) The image processing result for ground truth system.
Sensors 15 21636 g006
Figure 7. Indoor experiment site for demonstrating the performance of the proposed algorithm.
Figure 7. Indoor experiment site for demonstrating the performance of the proposed algorithm.
Sensors 15 21636 g007
Figure 8. The top view of the indoor environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Figure 8. The top view of the indoor environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Sensors 15 21636 g008
Figure 9. (a) Top view of the results of the indoor experiment. Red triangles, color1green triangles, and blue circles denote results of PnP, color1OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the ground truth data.
Figure 9. (a) Top view of the results of the indoor experiment. Red triangles, color1green triangles, and blue circles denote results of PnP, color1OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the ground truth data.
Sensors 15 21636 g009
Table 2. Comparison of the error statistics of the conventional and the proposed algorithms in the indoor environment (Unit: m, deg).
Table 2. Comparison of the error statistics of the conventional and the proposed algorithms in the indoor environment (Unit: m, deg).
P3P AlgorithmOPnP AlgorithmProposed Algorithm
MeanStdevMeanStdevMeanStdev
x0.28120.02180.27520.03730.2240.0152
y0.25430.02290.28130.03620.20790.0131
z0.0220.00030.06410.00260.02440.0003
θ roll 0.58570.25211.47940.32130.78750.2246
θ pitch 0.65020.27171.35150.29140.67940.2816
θ yaw 4.97755.05885.76474.58452.04172.8283

4.3. Experiment in Outdoor Environment

The outdoor experiments have been conducted at Korea Advanced Institute of Science and Technology in Daejeon, South Korea as shown in Figure 10. color1The robot system for experiments is equipped with the stereo vision same as the previous indoor experiments and we added Huace X90 RTK-GPS receiver [47] and E2BOX IMU 9DOFV2 [48] for the reference which have 2 cm and 1 accuracy, respectively. The probabilistic feature map was collected from 383 positions in 177 m trajectory as shown in Figure 11. The maximum valid depth information from the stereo vision system was restricted to 20 m because the uncertainty of depth data rapidly increases with the distance beyond 20 m.
Figure 10. Outdoor experiment site for demonstrating the performance of the proposed algorithm.
Figure 10. Outdoor experiment site for demonstrating the performance of the proposed algorithm.
Sensors 15 21636 g010
Figure 11. Top view of the outdoor environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Figure 11. Top view of the outdoor environment. Blue circles denote the covariance of features and the red dots denote the ground truth of the robot’s trajectory.
Sensors 15 21636 g011
The experiment for localization was performed at 383 positions around the probabilistic feature map. Figure 12a shows the top view of the results in the outdoor experiment. Euclidean distance errors from the reference for each algorithm are shown in Figure 12b. The error results of the 6-DoF robot’s pose are presented in Table 3. The accuracy of the proposed algorithm is also higher than that of the conventional algorithm particularly for the x, y, and θ y a w values similar to the simulation and the indoor experiment. It is natural that the accuracy of z and θ roll values does not show improvement since the z and θ roll values are not compensated well because the experiment is performed on a flat ground. The experimental results are similar to the simulation results. The p-value of the paired t-Test is 2 . 339 × 10 - 23 , which statistically confirms the superiority of the proposed method.
Figure 12. (a) Top view of the outdoor results. Red triangles, color1green triangles, and blue circles denote results of PnP, color1OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the reference data.
Figure 12. (a) Top view of the outdoor results. Red triangles, color1green triangles, and blue circles denote results of PnP, color1OPnP, and the proposed algorithm, respectively; (b) Euclidean distance errors relative to the reference data.
Sensors 15 21636 g012
Table 3. Comparison of the error statistics of the conventional and the proposed algorithms in the outdoor environment (Unit: m, deg).
Table 3. Comparison of the error statistics of the conventional and the proposed algorithms in the outdoor environment (Unit: m, deg).
P3P AlgorithmOPnP AlgorithmProposed Algorithm
MeanStdevMeanStdevMeanStdev
x1.08650.53742.10111.25460.74730.1549
y0.89080.43341.59472.05420.69350.2382
z0.16070.03760.20560.05410.17540.0356
θ roll 1.4091.97770.44891.15412.17613.9504
θ pitch 1.43472.45321.20552.51321.34892.1354
θ yaw 2.1554.2643.12185.12351.56891.9399

5. Conclusions

This paper sought to enhance the accuracy of monocular camera localization using a probabilistic feature map that is generated in advance with a prior data set by adopting probabilistic sensor system modeling. When the map is generated, the probabilistic feature map is estimated not only by the sensor system modeling but also by considering the uncertainty of the robot’s pose. In the conventional PnP method, the camera pose is estimated by minimizing the Euclidean distance between 2D-to-3D matching correspondences. The proposed algorithm is optimized based on the Mahalanobis distance error in the image plane between the matching correspondences. The main contribution of this paper is that the proposed method enhances the accuracy of the conventional camera pose estimation algorithm by providing probabilistic sensor modeling. The performance of the proposed algorithm is demonstrated by comparing with the conventional PnP algorithm in simulations and real experiments. By the experiments conducted in indoor and outdoor environments, the superiority of the proposed algorithm is proved.
Although the average computation of the PnP algorithm takes less than 5 ms per one frame, the proposed algorithm takes about 100 ms per one frame in the computing platform of Intel i7 3.4 GHz with 8 GB RAM. The reason is that the proposed algorithm solves a complex nonlinear optimization problem. color1The average computation times of subtasks are: 187 ms for feature extraction, 231 ms for matching, 5 ms for initial pose estimation using conventional PnP algorithm, and 95 ms for optimization in our proposed method. Therefore, a simple linearization method for the proposed algorithm and fast feature management should be researched for real time operation in the future. The proposed algorithm will be also applied to vast data sets and various environments in the future.

Acknowledgments

This work was supported in part by the Technology Innovation Program, 10045252, Development of robot task intelligence technology, supported by the Ministry of Trade, Industry, and Energy (MOTIE, Korea). This research was also supported by the project, “Development of basic SLAM technologies for autonomous underwater robot and software environment for MOSS-IvP” sponsored by Korea Research Institute of Ships & Ocean Engineering (KRISO), Korea. The students are supported by Ministry of Land, Infrastructure and Transport (MoLIT) as U-City Master and Doctor Course Grant Program.

Author Contributions

Hyungjin Kim and Donghwa Lee conducted algorithm design, experiments, and analysis under the supervision of Hyun Myung. Taekjun Oh and Hyun-Taek Choi conducted experiments and analysis. All authors were involved in writing paper, literature review, and discussion of results.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Snavely, N.; Seitz, S.M.; Szeliski, R. Photo tourism: Exploring photo collections in 3D. ACM Trans. Graph. 2006, 25, 835–846. [Google Scholar] [CrossRef]
  2. Anguelov, D.; Dulong, C.; Filip, D.; Frueh, C.; Lafon, S.; Lyon, R.; Ogale, A.; Vincent, L.; Weaver, J. Google street view: Capturing the world at street level. Computer 2010, 43, 32–38. [Google Scholar] [CrossRef]
  3. Li, Y.; Snavely, N.; Huttenlocher, D.P. Location Recognition Using Prioritized Feature Matching. In Proceedings of 11th European Conference on Computer Vision (ECCV), Heraklion, Rethymnon, Greece, 5–11 September 2010; pp. 791–804.
  4. Robertson, D.P.; Cipolla, R. An Image-Based System for Urban Navigation. In Proceedings of the British Machine Vision Conference (BMVC), Nottingham, UK, 1–5 Sptember 2004; pp. 1–10.
  5. Kosecka, J.; Zhou, L.; Barber, P.; Duric, Z. Qualitative Image Based Localization in Indoors Environments. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Fairfax, VA, USA, 18–20 June 2003; Volume 2, pp. 3–8.
  6. Napier, A.; Newman, P. Generation and Exploitation of Local Orthographic Imagery for Road Vehicle Localisation. In Proceedings of the IEEE Symposium on Intelligent Vehicles (IV), Madrid, Spain, 3–7 June 2012; pp. 590–596.
  7. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  8. Bay, H.; Tuytelaars, T.; van Gool, L. SURF: Speeded up Robust Features. In Proceedings of the European Conference on Computer Vision (ECCV), Graz, Austria, 7–13 May 2006; pp. 404–417.
  9. Zhang, W.; Kosecka, J. Image Based Localization in Urban Environments. In Proceedings of the IEEE Third International Symposium on 3D Data Processing, Visualization, and Transmission, Chapel Hill, NC, USA, 14–16 June 2006; pp. 33–40.
  10. Lu, G.; Kambhamettu, C. Image-Based Indoor Localization System Based on 3D SfM Model. In Proceedings of the International Society for Optics and Photonics on IS&T/SPIE Electronic Imaging, San Francisco, CA, USA, 2 February 2014; pp. 90250H-1–90250H-8.
  11. Kim, H.; Lee, D.; Oh, T.; Lee, S.W.; Choe, Y.; Myung, H. Feature-Based 6-DoF Camera Localization Using Prior Point Cloud and Images. In Proceedings of the Robot Intelligence Technology and Applications (RiTA), Beijing, China, 6–8 November 2014; pp. 3–11.
  12. Jebara, T.; Azarbayejani, A.; Pentland, A. 3D structure from 2D motion. IEEE Signal Process. Mag. 1999, 16, 66–84. [Google Scholar] [CrossRef]
  13. Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle Adjustment–A Modern Synthesis. In Proceedings of the International Workshop on Vision Algorithms: Theory and Practice, Corfu, Greece, 21–22 September 2000; pp. 298–372.
  14. Frahm, J.M.; Fite-Georgel, P.; Gallup, D.; Johnson, T.; Raguram, R.; Wu, C.; Jen, Y.H.; Dunn, E.; Clipp, B.; Lazebnik, S.; et al. Building Rome on a Cloudless Day. In Proceedings of European Conference on Computer Vision (ECCV), Heraklion, Crete, Greece, 5–11 September 2010; pp. 368–381.
  15. Strecha, C.; Pylvanainen, T.; Fua, P. Dynamic and Scalable Large Scale Image Reconstruction. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Francisco, CA, USA, 13–18 June 2010; pp. 406–413.
  16. Pollefeys, M.; Nistér, D.; Frahm, J.M.; Akbarzadeh, A.; Mordohai, P.; Clipp, B.; Engels, C.; Gallup, D.; Kim, S.J.; Merrell, P.; et al. Detailed real-time urban 3D reconstruction from video. Int. J. Comput. Vis. 2008, 78, 143–167. [Google Scholar] [CrossRef]
  17. Sattler, T.; Leibe, B.; Kobbelt, L. Fast Image-Based Localization Using Direct 2D-to-3D Matching. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011; pp. 667–674.
  18. Irschara, A.; Zach, C.; Frahm, J.M.; Bischof, H. From Structure-from-Motion Point Clouds to Fast Location Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 2599–2606.
  19. Lu, G.; Ly, V.; Shen, H.; Kolagunda, A.; Kambhamettu, C. Improving Image-Based Localization through Increasing Correct Feature Correspondences. In Proceedings of the International Symposium on Advances in Visual Computing, Crete, Greece, 29–31 July 2013; pp. 312–321.
  20. Gao, X.S.; Hou, X.R.; Tang, J.; Cheng, H.F. Complete solution classification for the perspective-three-point problem. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 930–943. [Google Scholar]
  21. Lepetit, V.; Moreno-Noguer, F.; Fua, P. EPnP: An accurate O(n) solution to the PnP problem. Int. J. Comput. Vis. 2009, 81, 155–166. [Google Scholar] [CrossRef] [Green Version]
  22. Zheng, Y.; Kuang, Y.; Sugimoto, S.; Astrom, K.; Okutomi, M. Revisiting the pnp Problem: A Fast, General and Optimal Solution. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Sydney, NSW, Australia, 1–8 December 2013; pp. 2344–2351.
  23. Garro, V.; Crosilla, F.; Fusiello, A. Solving the pnp Problem with Anisotropic Orthogonal Procrustes Analysis. In Proceedings of the Second International Conference on 3D Imaging, Modeling, Processing, Visualization & Transmission, Zurich, Switzerland, 13–15 October 2012; pp. 262–269.
  24. Ferraz, L.; Binefa, X.; Moreno-Noguer, F. Very Fast Solution to the PnP Problem with Algebraic Outlier Rejection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 501–508.
  25. Ferraz, L.; Binefa, X.; Moreno-Noguer, F. Leveraging Feature Uncertainty in the PnP Problem. In Proceedings of the British Machine Vision Conference (BMVC), Nottingham, UK, 1–5 September 2014.
  26. Henry, P.; Krainin, M.; Herbst, E.; Ren, X.; Fox, D. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. Int. J. Robot. Res. 2012, 31, 647–663. [Google Scholar] [CrossRef]
  27. Konolige, K.; Agrawal, M. FrameSLAM: From bundle adjustment to real-time visual mapping. IEEE Trans. Robot. 2008, 24, 1066–1077. [Google Scholar] [CrossRef]
  28. Lee, D.; Myung, H. Solution to the SLAM problem in low dynamic environments using a pose graph and an RGB-D sensor. Sensors 2014, 14, 12467–12496. [Google Scholar] [CrossRef] [PubMed]
  29. Davison, A.J. Real-Time Simultaneous Localisation and Mapping with a Single Camera. In Proceedings of the 9th IEEE International Conference on Computer Vision, Nice, France, 13–16 October 2003; pp. 1403–1410.
  30. Lemaire, T.; Lacroix, S.; Sola, J. A practical 3D Bearing-Only SLAM Algorithm. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Edmonton, AB, Canada, 2–6 August 2005; pp. 2449–2454.
  31. Davison, A.J.; Cid, Y.G.; Kita, N. Real-Time 3D SLAM with Wide-Angle Vision. In Proceedings of the IFAC/EURON Symposium on Intelligent Autonomous Vehicles, Lisbon, Portugal, 5–7 July 2004.
  32. Davison, A.J.; Reid, I.D.; Molton, N.D.; Stasse, O. MonoSLAM: Real-time single camera SLAM. IEEE Trans. Pattern Anal. Mach. Intell. 2007, 29, 1052–1067. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Jeong, W.; Lee, K.M. CV-SLAM: A New Ceiling Vision-Based SLAM Technique. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Edmonton, AB, Canada, 2–6 August 2005; pp. 3195–3200.
  34. Bhandarkar, S.M.; Suk, M. Hough Clustering Technique for Surface Matching. In Proceedings of the International Association of Pattern Recognition (IAPR) Workshop on Computer Vision, Tokyo, Japan, 12–14 October 1988; pp. 82–85.
  35. Lee, S.; Lee, S.; Baek, S. Vision-based kidnap recovery with SLAM for home cleaning robots. J. Intell. Robot. Syst. 2012, 67, 7–24. [Google Scholar] [CrossRef]
  36. Costa, A.; Kantor, G.; Choset, H. Bearing-Only Landmark Initialization with Unknown Data Association. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), New Orleans, LA, USA, 26 April–1 May 2004; Volume 2, pp. 1764–1770.
  37. Bailey, T. Constrained Initialisation for Bearing-Only SLAM. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Taipei, Taiwan, 14–19 September 2003; Volume 2, pp. 1966–1971.
  38. Stereo Accuracy and Error Model of XB3. Available online: http://www.ptgrey.com/support/downloads/10403? (accessed on 27 April 2015).
  39. Smith, R.C.; Cheeseman, P. On the representation and estimation of spatial uncertainty. Int. J. Robot. Res. 1986, 5, 56–68. [Google Scholar] [CrossRef]
  40. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  41. Beis, J.S.; Lowe, D.G. Shape Indexing Using Approximate Nearest-Neighbour Search in High-Dimensional Spaces. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), San Juan, Puerto Rico, 17–19 June 1997; pp. 1000–1006.
  42. Bhattacharyya, A. On a measure of divergence between two multinomial populations. Sankhyā: Indian J. Stat. 1946, 7, 401–406. [Google Scholar]
  43. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  44. Murty, K.G.; Yu, F.T. Linear Complementarity, Linear and Nonlinear Programming; Heldermann Verlag: Berlin, Germany, 1988. [Google Scholar]
  45. Pioneer 3-AT. Available online: http://www.mobilerobots.com/ResearchRobots/P3AT.aspx (accessed on 27 April 2015).
  46. Bumblebee XB3. Available online: http://www.ptgrey.com/bumblebee-xb3-1394b-stereo-vision-camera-systems-2 (accessed on 27 April 2015).
  47. Huace X90 RTK-GPS. Available online: http://geotrax.in/assets/x90GNSS_Datasheet.pdf (accessed on 27 April 2015).
  48. E2BOX IMU. Available online: http://www.e2box.co.kr/124 (accessed on 27 April 2015).

Share and Cite

MDPI and ACS Style

Kim, H.; Lee, D.; Oh, T.; Choi, H.-T.; Myung, H. A Probabilistic Feature Map-Based Localization System Using a Monocular Camera. Sensors 2015, 15, 21636-21659. https://doi.org/10.3390/s150921636

AMA Style

Kim H, Lee D, Oh T, Choi H-T, Myung H. A Probabilistic Feature Map-Based Localization System Using a Monocular Camera. Sensors. 2015; 15(9):21636-21659. https://doi.org/10.3390/s150921636

Chicago/Turabian Style

Kim, Hyungjin, Donghwa Lee, Taekjun Oh, Hyun-Taek Choi, and Hyun Myung. 2015. "A Probabilistic Feature Map-Based Localization System Using a Monocular Camera" Sensors 15, no. 9: 21636-21659. https://doi.org/10.3390/s150921636

APA Style

Kim, H., Lee, D., Oh, T., Choi, H. -T., & Myung, H. (2015). A Probabilistic Feature Map-Based Localization System Using a Monocular Camera. Sensors, 15(9), 21636-21659. https://doi.org/10.3390/s150921636

Article Metrics

Back to TopTop