Next Article in Journal
Electrochemical Immunosensor Based on Polythionine/Gold Nanoparticles for the Determination of Aflatoxin B1
Next Article in Special Issue
Clock Synchronization in Wireless Sensor Networks: An Overview
Previous Article in Journal
Eliminating the Interference of Oxygen for Sensing Hydrogen Peroxide with the Polyaniline Modified Electrode
Previous Article in Special Issue
High Precision Signal Processing Algorithm for White Light Interferometry
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Objective Error Criterion for Evaluation of Mapping Accuracy Based on Sensor Time-of-Flight Measurements

Department of Electrical and Electronics Engineering, Bilkent University, Bilkent 06800 Ankara, Turkey
Sensors 2008, 8(12), 8248-8261; https://doi.org/10.3390/s8128248
Submission received: 24 September 2008 / Revised: 2 December 2007 / Accepted: 2 December 2008 / Published: 15 December 2008
(This article belongs to the Special Issue Sensor Algorithms)

Abstract

:
An objective error criterion is proposed for evaluating the accuracy of maps of unknown environments acquired by making range measurements with different sensing modalities and processing them with different techniques. The criterion can also be used for the assessment of goodness of fit of curves or shapes fitted to map points. A demonstrative example from ultrasonic mapping is given based on experimentally acquired time-of-flight measurements and compared with a very accurate laser map, considered as absolute reference. The results of the proposed criterion are compared with the Hausdorff metric and the median error criterion results. The error criterion is sufficiently general and flexible that it can be applied to discrete point maps acquired with other mapping techniques and sensing modalities as well.

1. Introduction

A truly autonomous robot should be able to build an accurate spatial model of its unknown physical environment based on the sensory data it acquires. This problem is addressed by the field of robotic mapping. The environment in which the robot operates could be a static one, or more realistically, a dynamically changing environment. Therefore, the selection of a suitable mapping scheme is an important task [1]. The mapping approaches used in the 1980s and the early 1990s can be categorized into two, as metric and topological approaches. Metric maps capture the geometric properties of the environment, whereas topological maps describe the connectivity of its different parts. Since the 1990s, probabilistic techniques have dominated the field of robotic mapping. In particular, the simultaneous localization and mapping (SLAM) problem, also referred as concurrent mapping and localization (CML), has received considerable interest and has been a highly active research area over the past two decades. Consequently, a large family of algorithms for concurrent estimation of the map and the robot pose have emerged [2, 3].
Since robot mapping is characterized by sensor measurement noise and uncertainty, most of the current robotic mapping algorithms in the literature are probabilistic in nature [4]. These algorithms employ probabilistic models of the robot and its surroundings and rely on probabilistic inference to extract the map of the environment based on sensory measurements. The different approaches can be classified as Kalman filtering techniques [57], approaches based on Dempster's expectation-maximization algorithm [8, 9], occupancy grid techniques [10, 11], techniques for learning object models or features in the environment [1214], and hybrid approaches [15].
Feature-based approaches are based on extracting the geometry of the environment from sensor data as the first step in data interpretation process (e.g., edge detection, fitting straight-lines or curves to obstacle boundaries, differentiating features such as planar walls, corners, and edges, open and closed doors, and corridors). Important issues to consider are the representation of uncertainty, suitability of the selected feature to the environment and the type of data, the reliability of the feature extraction process, and the speed with which the model can be constructed. An alternative representation is the certainty or occupancy grid where the environment is discretized into an array of cells. The difficulty of line segment-based feature extraction has been an important factor in the development of the grid concept proposed by Elfes [10, 11]. [Reference 1] provides a comprehensive survey and further references on robotic mapping, with a focus on indoor environments.
The map of an unknown environment can be acquired by making range measurements with a variety of sensing modalities, such as ultrasonic sensors, laser range finders, or millimeter wave radar. However, the acquired map is usually erroneous, consisting of outliers, gaps, noisy or incorrect measurements. Some of the measurements are caused by multiple or higher-order reflections from specular (mirror-like) surfaces, or crosstalk between multiple sensors, making the data difficult to interpret. This is particularly a problem with airborne ultrasonics and underwater sonar sensing. Researchers have identified regions of constant depth (RCDs) [16, 17], and defined and processed ultrasonic arc maps that represent the angular uncertainty of ultrasonic sensors [1820]. Line segments [21], polynomials [22], and snake curves [23] have been fitted to acquired ultrasonic maps to represent the map points more compactly and efficiently. However, to the best of our knowledge, there is no well-defined objective criterion or systematic method for the evaluation of map accuracy. Lack of such criteria also makes it difficult to compare maps acquired by different mapping techniques and sensor types. Although the areas of sensing, measurement technology, and mapping have developed considerably and have been extended to 3-D over the recent years [9, 2426], assessment of the accuracy of the acquired maps and comparison between different mapping techniques is an important issue not extensively studied. In most mapping studies, the map accuracy is assessed by graphically displaying the acquired map together with the true map, and a subjective, qualitative judgment is made by visual comparison. The main contribution of this paper is the proposition of an objective and quantitative error criterion for the accuracy assessment and comparative evaluation of acquired maps.
In Section 2., we give a description of the proposed error criterion and provide two other criteria for comparison: the Hausdorff metric and the median error. The use of the criterion is demonstrated through an example from ultrasonic and laser sensing in Section 3. Section 4. provides details of the experimental procedure and compares the results of the proposed criterion with the Hausdorff metric and the median error. Section 5. discusses the limiting circumstances for the criterion that may arise when there are temporal or spatial differences in acquiring the maps. The last section concludes the paper by indicating some potential application areas and providing directions for future research.

2. The Error Criterion

Let P ⊂ ℝ3 and Q ⊂ ℝ3 be two finite sets of arbitrary points with N1 points in set P and N2 points in set Q. We do not require the correspondence between the two sets of points to be known. Each point set could correspond to either (i) an acquired set of map points, (ii) discrete points corresponding to an absolute reference (the true map), or (iii) some curve (2-D) or shape (3-D) fit to the map points. The absolute reference could be an available true map or plan of the environment or could be acquired by making range or time-of-flight measurements through a very accurate sensing system.
The well-known Euclidean distance d(pi, qj) : ℝ3 → ℝ0 of the i'th point in set P with position vector pi = (pxi, pyi, pzi)T to the j'th point qj = (qxj, qyj, qzj)T in set Q is given by:
d ( p i , q j ) = ( p x i q x j ) 2 + ( p y i q y j ) 2 + ( p z i q z j ) 2 i 1 , , N 1 j 1 , , N 2
There is a choice of metrics to measure the similarity between two sets of points, each with certain advantages and disadvantages:
A very simple metric is to take the minimum of the distances between any point of set P and any point of Q. This corresponds to a minimin function and is defined as:
D ( P , Q ) = min p i P { min q i Q { d ( p i , q i ) }
In other words, for every point pi of set P, we find its minimum distance to any point qj of Q and we keep the minimum distance found among all points pi. This metric has very low informative content since the minimum value is not representative of the whole point set. For the same value of D(P, Q), the shape, position, and orientation of the two point sets can be very different with respect to each other. Even if one single point of the two point sets overlap, the metric will take the value zero. Since this metric has obvious shortcomings, we did not not consider it for our purpose of comparing maps.
The error criterion we propose for measuring the closeness or similarity between sets P and Q overcomes the shortcomings of the minimin function in Equation (2) by taking into account all of the points in the two sets:
ε mean = 1 2 ( 1 N 1 i = 1 N 1 min q i Q { d ( p i , q j ) } + 1 N 2 j = 1 N 2 min p j P { d ( p i , q j ) } )
The distance of every point in set P to the nearest point in set Q is found and averaged, and vice versa. The two terms in Equation (3) are averaged so that the criterion is symmetric with respect to P and Q. For the simple example in Figure 1(a) where N1 = 3 and N2 = 2, the error is ε mean = d ( p 1 , q 1 ) + d ( p 2 , q 2 ) + d ( p 3 , q 2 ) 6 + d ( p 1 , q 1 ) + d ( p 2 , q 2 ) 4. If the two sets of points are completely coincident, the average distance between the two sets will be zero. If one set is a subset of the other, there will be some error. Had an asymmetric criterion been employed, say including only the first (or the second) term in Equation (3), the error would have been zero when PQ (or QP). The case when PQ is illustrated in Figure 1(b) where the major contribution to the resulting error comes from the second term in Equation (3). Gaps occurring in the maps and sparsity are penalized by the error criterion, resulting in larger errors on the average.
The error criterion we propose is sufficiently general that it can be used to compare any two arbitrary sets of points with each other. This makes it possible to compare the accuracy of discrete point maps acquired with different techniques or sensing modalities with an absolute reference, as well as among themselves, both in 2-D and 3-D. When curves or shapes (e.g., lines, polynomials, snakes, spherical or elliptical caps) are fitted to the map points, the criterion proposed here also enables us to make an assessment of goodness of fit of the curve or shape to the map points. In other words, a fitted curve or shape comprised of a finite number of points can be treated in exactly the same way.
A widely used measure for comparing point sets in the computer vision area is the Hausdorff metric [27]. We included this metric in our study for comparison purposes. Hausdorff distance is commonly used for matching images in applications such as image analysis [28], visual robot navigation [29], target indexing in SAR images [30], computer-assisted surgery, etc. Basically, the Hausdorff metric is used to check if a template image is present in a test image so that the lower the value of the metric, the better is the match. It works well even in the presence of noise or occlusion of the target. Hausdorff distance can be defined as the maximum of the minimum distances of a set of points to the points in the other set.
More formally, the directed Hausdorff distance from set P to set Q is a maximin function, defined as:
h ( P , Q ) = min p i P { min q i Q { d ( p i , q j ) } }
where d(pi, qj) can, in fact, be any metric between these points. For simplicity, we take d(pi, qj) as the Euclidean distance between pi and qj as defined in Equation (1). Note that the Hausdorff distance, defined as above, is directed or asymmetric, meaning that most of the time, h(P, Q) ≠ h(Q, P). This asymmetry is a property of maximin functions, while minimin functions are symmetric.
A more general definition of Hausdorff distance is
H ( P , Q ) = max { h ( P , Q ) , h ( Q , P ) }
which defines the Hausdorff distance between sets P and Q, while Equation (4) defines the Hausdorff distance from P to Q (the directed Hausdorff distance). Referring to the example in Figure 1(a), h(P,Q) = d(p3, q2), h(Q, P) = d(p2, q2), and H(P, Q) = max{d(p3, q2), d(p2, q2) } = d(p3,q2) since d(p3, q2) > d(p2, q2).
Instead of taking the average or the maximum value among the minimum distance values, another possibility is to take the median of the minimum distances. In this case, a suitable error measure can be defined as:
ε median = 1 2 ( median p i P [ min q i Q { d ( p i , p j ) } ] + median q i Q [ min p i P { d ( p i , q j ) } ] )
Referring to our example in Figure 1(a), the median error is ε median = 1 2 [ d ( p 2 , q 2 ) + d ( p 1 , q 1 ) + d ( p 2 , q 2 ) 2 ]. When there is an even number of minimum distances in the first or the second term of Equation (6), their median is found by averaging the two distances that fall in the middle.
For all of the criteria above, the two sets of points can be chosen as (i) two different sets of map points acquired with different mapping techniques or different sensing modalities (e.g., ultrasonic and laser, laser being the accurate reference), or (ii) two sets of fitted curve points (2-D) or shapes (3-D) (e.g., two polynomials, snake curves or spherical caps) to maps extracted by different mapping techniques or sensing modalities, or (iii) a set of extracted map points and a set of curve points fitted to them (i.e., assessment of goodness of fit).

3. Mapping Techniques

To demonstrate the use of the criterion through an example, we consider three ultrasonic mapping techniques. Most commonly, the large beamwidth of the transducer is accepted as a device limitation which determines the angular resolving power of the system. In this naive approach (method 1), a range reading of r from a transmitting/receiving transducer is taken to imply that the echo-producing object lies along the line-of-sight (LOS) of the transducer at the measured range. This is the simplest and the earliest approach, where a mark is placed along the LOS of the sensor at the measured range. It produces reasonable estimates for the positions of objects at nearby ranges or high frequencies of operation where the corresponding sensor beamwidth is small. If the sensor beamwidth is large, the angular position of the object is highly uncertain. Furthermore, this technique cannot eliminate any of the outliers or erroneous measurements.
The Arc-Transversal Median Algorithm (method 2) was developed by Choset and his co-workers, and requires both extensive bookkeeping and considerable amount of processing [18]. The algorithm can be summarized as follows: First, an ultrasonic arc map (UAM), representing the angular uncertainty of each range measurement is constructed. For each arc in the UAM, the positions of the intersection(s) with other arcs, if they exist, are recorded. For arcs without any intersections, the mid-point of the arc is taken to represent the actual point of reflection (as in method 1) which corresponds to the intersection point of the arc with the LOS. If the arc has a single intersection, the algorithm uses the intersection point as the location of the reflecting object. For arcs with more intersections, the median of the positions of the intersection points with other arcs is chosen to represent the actual point of reflection. In the work reported in [18], the median operation is applied when an arc has three or more intersection points. If there is an even number of intersections, the algorithm uses the mean of the two middle values (except that arcs with two intersections are ignored). It can be considered as a much improved version of method 1, where a single mark is placed along the LOS at the measured range.
Method 3 is based on directional processing of ultrasonic data and is detailed in [31]. This technique is based on the idea that in processing the acquired range data, there is a direction-of-interest (DOI) associated with each detected echo. Ideally, the DOI corresponds to the direction of a perpendicular line drawn from the sensor to the nearest surface from which an echo is detected. However, in practice, due to the angular uncertainty of the object position, the DOI can be approximated as the LOS of the sensor when an echo is detected. Since prior information on the environment is usually unavailable, the DOI needs to be updated while sensory data are being collected and processed on-line. It may also be possible to determine the DOI by post processing, based on the distribution of the acquired data by choosing it perpendicular to the direction along which the spread of the collected data over a given region is maximum. It should be noted that ideally, the DOI is different for each detected echo and sensor position.
The latter two methods are the most recently reported techniques for processing ultrasonic range measurements. All three methods use the same set of time-of-flight measurements acquired from the same environment. However, due to differences in processing the measurements, each one results in a different set of map points.

4. Experimental Procedure and Results

The experiments are performed using the Polaroid 6500 series ultrasonic transducers [32] available on the Nomad 200 mobile robot. The Nomad 200, shown in Figure 2, is an integrated mobile robot including tactile, infrared, ultrasonic, and structured-light sensing systems. The Polaroid 6500 has a resonance frequency of fo = 49.4 kHz and aperture radius a = 2.0 cm, corresponding to a half-beamwidth angle of θ o = sin 1 [ 0.61 c a f o ] = 12.2 °. It measures time-of-flight to, which is the round-trip travel time of the transmitted pulse between the transducer and the point of reflection, from which the range r can be calculated as r = cto/2, where c is the speed of sound in air [33]. Results from three sample experiments are provided where ultrasonic time-of-flight and laser triangulation measurements are collected simultaneously. Both systems are rigidly fixed to the turret of the mobile robot so that the correspondence between them is never altered. The laser system is much more expensive, complex, and heavy, also requiring higher-power. Furthermore, this mode of sensing does not work in all environments, such as those with dark-colored upholstery or glass. Since it reveals a very accurate surface profile, the profile detected by the laser is taken as an absolute reference in the evaluation of the maps. In all of the experiments, the environment is divided into square elements or pixels of size 1 cm.
In the first experiment, we considered a 90° corner which is a typical feature of indoor environments. Due to its high curvature at the corner point, accurate mapping of this feature is usually difficult with ultrasonic sensors. The laser data obtained from the 90° corner and the locations of the five ultrasonic sensors are presented in Figure 3(a). Ultrasonic time-of-flight measurements were collected by rotating the ultrasonic sensors located on an arc from –30° to 30° in 1° steps. The results of the three mapping methods are shown in Figure 3(b)-(d).
For the second experiment, we have constructed curved surfaces of varying curvature and dimensions in our laboratory, using thin cardboard (see Figure 4(a) for an example). Again, only the front five ultrasonic sensors of the robot have been fired. Comparison is made with a 9th-order polynomial fit to the laser data (Figure 3(e)) obtained from one of the curved surfaces of height 1.05 m and length 3.65 m. In this experiment, the mobile robot simply translates along a straight path from (–75, 0) cm to (75, 0) cm alongside the surface at an average distance of 1 m and collects time-of-flight measurements by firing the front five ultrasonic transducers at every 2.5 cm. The turret is oriented such that both the laser and the ultrasonic sensors are directed towards the surface and it does not rotate throughout the translational movement. In the maps of methods 1 and 2 shown in Figure 3(f) and (g), there are some outlier points due to higher-order reflections, crosstalk, reflections from other objects in the environment, or totally erroneous measurements.
We also present experimental results from the indoor environment in Figure 4(b), comprising smooth wooden (left) and painted walls (middle) and a window shade (right) with vertical slats of 15 cm width. Some of the corners of the room are not perfect (e.g., where the shade and the painted wall make a corner on the right). There is a cylindrical object of radius 15 cm at a distance of 30 cm from the center of the painted wall. In this experiment, we use a simple rule-based wall-following algorithm for motion planning while collecting data. The mobile robot simply tries to follow the walls of the indoor environment trying to maintain a fixed distance to the walls, using sensory feedback from the front three sensors. The resulting laser map is illustrated in Figure 3(i) and the ultrasonic maps are shown in Figure 3(j)-(l). Again, the maps acquired with methods 1 and 2 contain many artifacts, especially exterior to the surrounded region in which the robot moves.
The errors are of the order of several pixels and are presented in Table 1. Obviously, the smaller the error, the better is the match between the ultrasonic map and the absolute reference. Overall, method 3 results in less artifacts and consequently smaller error in general, and is therefore considered to be superior to the other two methods. As we demonstrated with this example, the proposed error criterion allows us to make a quantitative comparison between the mapping results of different techniques or sensing modalities.
In Table 2, the directed Hausdorff distances h(P, Q) and h(Q, P) are tabulated for the three experiments where the set P corresponds to the resulting ultrasonic map points, and the set Q corresponds to the laser data. The Hausdorff distance H(P, Q) between sets P and Q, which is the larger of h(P, Q) and h(Q, P), is indicated by using boldface characters in the table. The Hausdorff distance results are much larger than εmean since this metric takes the maximum among the minimum values, whereas our proposed metric averages the minimum distance values. Consequently, the Hausdorff distance severely penalizes outlier points in the map. For this reason, method 3, which is successful at eliminating the outliers in the second and third experiments results in smaller errors in these experiments.
Since we are comparing ultrasonic map points with much more accurate laser data in this example, h(P, Q) is larger than h(Q, P) most of the time. This is explained by the fact that the points of ultrasonic maps are typically more distributed, fuzzy, and erroneous compared to laser maps so that the maximin distance h(P, Q) usually corresponds to the distance between one of the outliers of the ultrasonic map points and a laser data point. On the other hand, when h(Q, P) is calculated, there is usually an ultrasonic map point nearer to each laser data point than the outliers so that the effect of outliers on the error is reduced.
The results of calculating the median error are tabulated in Table 3. Note that most of the errors are smaller than the mean error εmean presented in Table 1, and all of the errors are much smaller than the Hausdorff distances given in Table 2. Since this approach can filter out the outlier points successfully, outlier map points are not penalized as much as in the first two approaches.
In summary, it can be concluded that the Hausdorff metric severely penalizes outlier measurements since it takes the maximum value among the minimum distances for the set points. The median error filters out the outliers and under-penalizes them because of the nonlinear filtering involved in taking the median. Our proposed metric provides a reasonable balance between these two metrics by weighting each of the minimum distances equally and calculating their average value bi-directionally.

5. Discussion

In this work, we have assumed that the robot or the sensing mechanism is well localized so that sufficiently accurate estimates of its pose are available. If this assumption is not true so that the localization errors are large, the acquired maps may be shifted and rotated with respect to the true map, resulting in increased error values. In this case, localization errors could be dominating the map error. If we are comparing simultaneously acquired maps not with the true map but among themselves, the relative error between the maps will not be much affected from the localization errors. Note that our criterion still remains applicable under these circumstances.
We have also assumed that the maps are acquired with the different sensing modalities under similar conditions. In the case of a dynamic environment, maps should be acquired simultaneously. For a static environment, the timing could be different. In the example that we have given, the two data sets were acquired simultaneously from the ultrasonic and structured-light laser systems on the mobile robot.
Another possibility that may result in large errors between the two point sets can occur when the two sensing modalities compared are so different in nature or at completely different positions in the environment that the corresponding views of the environment and the occluded or “shadowed” parts of the environment may be very different.
In case the sensing modalities take partial views or scans of the environment, we assume that they correspond to the same sectoral scan. Otherwise, the map points would have to be matched and their correspondence be found, which is beyond the scope of this work. It is worth emphasizing once again that in this work, we do not require the correspondence between the two sets of points to be known and do not make distinctions between the points of the same data set.
To summarize, the limiting circumstances for the proposed criterion could be caused by spatial or temporal differences in the acquirement of the two sets of map points.

6. Conclusion

We have presented an objective criterion to compare maps obtained with different techniques with an absolute reference as well as among themselves. This criterion can be used to compare two arbitrary sets of points with each other without requiring the correspondence between the two sets of points to be known. The two sets of points can correspond to: (i) two different sets of map points acquired with different mapping techniques or different sensing modalities, or (ii) two sets of fitted curve points to maps extracted by different mapping techniques or sensing modalities, or (iii) a set of extracted map points and a set of curve points fitted to them. We have demonstrated how the criterion can be used through an example from ultrasonic sensing based on time-of-flight measurements, considering accurate laser data as absolute reference and have compared the results with the Hausdorff metric and the median error results. We believe that the error criterion is sufficiently general so that it can be applied to both 2-D and 3-D range measurements of different scales, acquired with different mapping techniques and sensing modalities. Potential application areas for this work include millimeter wave radar for unmanned ground vehicles, mobile robotics, underwater sonar, optical sensing and metrology, remote sensing, ocean surface exploration, geophysical exploration, and acoustic microscopy. As part of the future work, it would be beneficial to demonstrate the applicability of the criterion to these sensing modalities at different scales.

Acknowledgments

This work is supported by The Scientific and Technological Research Council of Turkey (TÜBİTAK) under grant number EEEAG-105E065.

References and Notes

  1. Lakemeyer, G.; Nebel, B. Robotic mapping: A survey. In Exploring Artificial Intelligence in the New Millennium; Lakemeyer, G.; Nebel, B. Morgan Kaufman, 2002; pp. 1–34, 1. [Google Scholar]
  2. Durrant-Whyte, H. F.; Bailey, T. Simultaneous Localisation and Mapping (SLAM): Part I The Essential Algorithms. IEEE Robot. Autom. Mag. 2006, 13, 99–108, 1. [Google Scholar]
  3. Bailey, T.; Durrant-Whyte, H. F. Simultaneous Localisation and Mapping (SLAM): Part II State of the Art. IEEE Robot. Autom. Mag. 2006, 13, 108–117, 1. [Google Scholar]
  4. Thrun, S.; Burgard, W.; Fox, D. Probabilistic Robotics; MIT Press: Cambridge, MA, 2005; 1. [Google Scholar]
  5. Bar-Shalom, Y.; Fortmann, T. E. Tracking and Data Association; Academic Press: New York, 1988; 1. [Google Scholar]
  6. Smith, R.; Self, M.; Cheeseman, P. Estimating uncertain spatial relationships in robotics. In Autonomous Robot Vehicles; Springer-Verlag: New York, 1990; pp. 167–193. [Google Scholar]
  7. Tardos, J. D.; Neira, J.; Newman, P. M.; Leonard, J. J. Robust mapping and localization in indoor environments using sonar data. Int. J. Robot. Res. 2002, 21, 311–330, 1. [Google Scholar]
  8. McLachlan, G. J.; Krishnan, T. The EM Algorithm and Extensions; Series in Probability and Statistics; Wiley: New York, 1997; 1. [Google Scholar]
  9. Thrun, S.; Martin, C.; Yufeng, L.; Hahnel, D.; Emery-Montemerlo, R.; Chakrabarti, D.; Burgard, W. A real-time expectation-maximization algorithm for acquiring multiplanar maps of indoor environments with mobile robots. IEEE T. Robotic. Autom. 2004, 20, 433–443, 1. [Google Scholar]
  10. Elfes, A. Sonar based real-world mapping and navigation. IEEE T. Robotic. Autom. 1987, RA-3, 249–265, 1. [Google Scholar]
  11. Elfes, A. Using occupancy grids for mobile robot perception and navigation. IEEE Computer 1989, 22, 46–57, 1. [Google Scholar]
  12. Wijk, O.; Christensen, H. I. Triangulation-based fusion of sonar data with application in robot pose tracking. IEEE T. Robotic. Autom. 2000, 16, 740–752, 1. [Google Scholar]
  13. Wijk, O.; Christensen, H. I. Localization and navigation of a mobile robot using natural point landmarks extracted from sonar data. Robot. Auton. Syst. 2000, 31, 31–42. [Google Scholar]
  14. Utete, S. W.; Barshan, B.; Ayrulu, B. Voting as validation in robot programming. Int. J. Robot. Res. 1999, 18, 401–413, 1. [Google Scholar]
  15. Thrun, S. A probabilistic online mapping algorithm for teams of mobile robots. Int. J. Robot. Res. 2001, 20, 335–363, 1. [Google Scholar]
  16. Leonard, J. J.; Durrant-Whyte, H. F. Directed Sonar Sensing for Mobile Robot Navigation; Kluwer Academic Publishers: Boston, MA, 1992; 1. [Google Scholar]
  17. Kuc, R.; Siegel, M. W. Physically-based simulation model for acoustic sensor robot navigation. IEEE T. Pattern Anal. 1987, PAMI-9, 766–778, 1. [Google Scholar]
  18. Choset, H.; Nagatani, K.; Lazar, N. A. The arc-transversal median algorithm: A geometric approach to increasing ultrasonic sensor azimuth accuracy. IEEE T. Robotic. Autom. 2003, 19, 513–522, 1., 3. [Google Scholar]
  19. Barshan, B.; Ba şkent, D. Morphological surface profile extraction with multiple range sensors. Pattern Recogn. 2001, 34, 1459–1467. [Google Scholar]
  20. Silver, D.; Morales, D.; Rekleitis, I.; Lisien, B.; Choset, H. Arc carving: obtaining accurate, low latency maps from ultrasonic range sensors. Proc. IEEE Int. Conf. Robotics Autom., 26 Apr – 1 May 2004; Vol. 2, pp. 1554–1561, 1.
  21. Ribas, D.; Ridao, P.; Neira, J.; Tardós, J. D. Line extraction from mechanically scanned imaging sonar. Pattern Recognition and Image Analysis, Lect. Notes Comput. Sc. 2007, 4477, 322–329, 1. [Google Scholar]
  22. Başkent, D.; Barshan, B. Surface profile determination from multiple sonar data using morphological processing. Int. J. Robot. Res. 1999, 18, 788–808, 1. [Google Scholar]
  23. Altun, K.; Barshan, B. Performance evaluation of ultrasonic arc map processing techniques by active snake contours. European Robotics Symp. 2008; Springer: Berlin, Germany, 2008; Vol. 44, pp. 185–194, 1. [Google Scholar]
  24. Borrmann, D.; Elseberg, J.; Lingemann, K.; Nüchter, A.; Hertzberg, J. Globally consistent 3D mapping with scan matching. J. Robotics Autonom. Syst. 2008, 56, 130–142, 1. [Google Scholar]
  25. Nüchter, A.; Lingemann, K.; Hertzberg, J.; Surmann, H. 6D SLAM–3D Mapping Outdoor Environments. J. Field Robot. 2007, 24, 699–722. [Google Scholar]
  26. Auran, P. G.; Silven, O. Underwater sonar range sensing and 3-D image formation. Control Eng. Pract. 1996, 4, 393–400, 1. [Google Scholar]
  27. Huttenlocher, D. P.; Klanderman, G. A.; Rucklidge, W. J. Comparing images using the Hausdorff distance. IEEE T. Pattern Anal. 1993, 15, 850–862, 2. [Google Scholar]
  28. Rucklidge, W. J. Efficiently locating objects using the Hausdorff distance. Int. J. Comput. Vision 1997, 24, 251–270, 2. [Google Scholar]
  29. Donoso-Aguirre, F.; Bustos-Salas, J.-P.; Torres-Torriti, M.; Guesalaga, A. Mobile Robot Localization using the Hausdorff distance. Robotica 2008, 26, 129–141, 2. [Google Scholar]
  30. Yi, J. H.; Bhanu, B.; Li, M. Target indexing in SAR images using scattering centers and the Hausdorff distance. Pattern Recogn. Lett. 1996, 17, 1191–1198, 2. [Google Scholar]
  31. Barshan, B. Directional processing of ultrasonic arc maps and its comparison with existing techniques. Int. J. Robot. Res. 2007, 26, 797–820, 3. [Google Scholar]
  32. Ultrasonic Components Group, Polaroid Corp. Polaroid Manual; Polaroid Corp.: Cambridge, MA, 1997; 4. [Google Scholar]
  33. Barshan, B. Fast processing techniques for accurate ultrasonic range measurements. Meas. Sci. Technol. 2000, 11, 45–50, 4. [Google Scholar]
Figure 1. (a) Point set P in red, point set Q in blue and some of the Euclidean distances; (b) the case when PQ and most of the error is contributed by the second term in Equation (3).
Figure 1. (a) Point set P in red, point set Q in blue and some of the Euclidean distances; (b) the case when PQ and most of the error is contributed by the second term in Equation (3).
Sensors 08 08248f1
Figure 2. Nomad 200 mobile robot. The ring of ultrasonic sensors can be seen close to the top rim of the turret, and the structured-light laser system is seen pointing rightwards on top.
Figure 2. Nomad 200 mobile robot. The ring of ultrasonic sensors can be seen close to the top rim of the turret, and the structured-light laser system is seen pointing rightwards on top.
Sensors 08 08248f2
Figure 3. Laser data acquired from the three environments (top row, red) and the sensor configuration (top row, green); resulting maps of method 1 (2nd row, blue), method 2 (3rd row, magenta), and method 3 (4th row, green).
Figure 3. Laser data acquired from the three environments (top row, red) and the sensor configuration (top row, green); resulting maps of method 1 (2nd row, blue), method 2 (3rd row, magenta), and method 3 (4th row, green).
Sensors 08 08248f3
Figure 4. (a) An example curved surface and a meter stick; (b) a view of the indoor environment in Figure 3(i) looking towards the right, showing the top, right, and bottom walls.
Figure 4. (a) An example curved surface and a meter stick; (b) a view of the indoor environment in Figure 3(i) looking towards the right, showing the top, right, and bottom walls.
Sensors 08 08248f4
Table 1. Results of the three experiments based on the proposed criterion.
Table 1. Results of the three experiments based on the proposed criterion.
εmean (pixels)
experiment 1experiment 2experiment 3

method 11.196.644.37

method 21.404.044.26

method 32.242.132.49
Table 2. Results of the three experiments based on the Hausdorff metric. The Hausdorff distance is highlighted in boldface fonts.
Table 2. Results of the three experiments based on the Hausdorff metric. The Hausdorff distance is highlighted in boldface fonts.
Hausdorff distance (pixels)
experiment 1experiment 2experiment 3

h(P, Q)h(Q, P)h(P, Q)h(Q, P)h(P, Q)h(Q, P)

method 17.282.24184.96.08139.016.12

method 25.663.16103.88.54138.017.00

method 311.315.008.949.4324.0014.00
Table 3. Results of the three experiments based on the median criterion.
Table 3. Results of the three experiments based on the median criterion.
εmedian (pixels)
experiment 1experiment 2experiment 3

method 11.212.291.71

method 21.212.121.50

method 32.122.001.21

Share and Cite

MDPI and ACS Style

Barshan, B. Objective Error Criterion for Evaluation of Mapping Accuracy Based on Sensor Time-of-Flight Measurements. Sensors 2008, 8, 8248-8261. https://doi.org/10.3390/s8128248

AMA Style

Barshan B. Objective Error Criterion for Evaluation of Mapping Accuracy Based on Sensor Time-of-Flight Measurements. Sensors. 2008; 8(12):8248-8261. https://doi.org/10.3390/s8128248

Chicago/Turabian Style

Barshan, Billur. 2008. "Objective Error Criterion for Evaluation of Mapping Accuracy Based on Sensor Time-of-Flight Measurements" Sensors 8, no. 12: 8248-8261. https://doi.org/10.3390/s8128248

Article Metrics

Back to TopTop