Next Article in Journal
Porous Silicon Structures as Optical Gas Sensors
Next Article in Special Issue
Statistical Analysis of the Performance of MDL Enumeration for Multiple-Missed Detection in Array Processing
Previous Article in Journal
Underwater Acoustic Wavefront Visualization by Scanning Laser Doppler Vibrometer for the Characterization of Focused Ultrasonic Transducers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Fast Robot Identification and Mapping Algorithm Based on Kinect Sensor

1
National School of Software, Xidian University, Xi’an 710071, China
2
School of Computer Science and Engineering, Xi’an University of Technology,5 Jinhua S. Rd., Xi’an 710048, China
3
State Key Laboratory for Strength and Vibration of Mechanical Structures, Xi’an Jiaotong University, 28 Xianning W. Rd., Xi’an 710049, China
4
Department of Electrical and Computer Engineering, West Virginia University, 405 Fayette Pike, Montgomery, WV 25136, USA
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2015, 15(8), 19937-19967; https://doi.org/10.3390/s150819937
Submission received: 27 May 2015 / Revised: 23 June 2015 / Accepted: 5 August 2015 / Published: 14 August 2015
(This article belongs to the Special Issue Identification, Information & Knowledge in the Internet of Things)

Abstract

:
Internet of Things (IoT) is driving innovation in an ever-growing set of application domains such as intelligent processing for autonomous robots. For an autonomous robot, one grand challenge is how to sense its surrounding environment effectively. The Simultaneous Localization and Mapping with RGB-D Kinect camera sensor on robot, called RGB-D SLAM, has been developed for this purpose but some technical challenges must be addressed. Firstly, the efficiency of the algorithm cannot satisfy real-time requirements; secondly, the accuracy of the algorithm is unacceptable. In order to address these challenges, this paper proposes a set of novel improvement methods as follows. Firstly, the ORiented Brief (ORB) method is used in feature detection and descriptor extraction. Secondly, a bidirectional Fast Library for Approximate Nearest Neighbors (FLANN) k-Nearest Neighbor (KNN) algorithm is applied to feature match. Then, the improved RANdom SAmple Consensus (RANSAC) estimation method is adopted in the motion transformation. In the meantime, high precision General Iterative Closest Points (GICP) is utilized to register a point cloud in the motion transformation optimization. To improve the accuracy of SLAM, the reduced dynamic covariance scaling (DCS) algorithm is formulated as a global optimization problem under the G2O framework. The effectiveness of the improved algorithm has been verified by testing on standard data and comparing with the ground truth obtained on Freiburg University’s datasets. The Dr Robot X80 equipped with a Kinect camera is also applied in a building corridor to verify the correctness of the improved RGB-D SLAM algorithm. With the above experiments, it can be seen that the proposed algorithm achieves higher processing speed and better accuracy.

1. Introduction

Internet of Things (IoT) is driving innovation in an ever-growing set of application domains, such as intelligent processing for autonomous robots. For an autonomous robot one grand challenge, which triggers many requirements for the information processing algorithm, is how to sense its surrounding environment effectively. Therefore, the integration and management of the robot sensor data and the system model of the robot are very important.
In order to navigate [1] in unknown environments, mobile robots should build environment maps and estimate their own position in such maps. Solving these two problems simultaneously is called simultaneous localization and mapping (SLAM). RGB-D SLAM algorithm is divided into two parts: “front-end” and “back-end”. The “front-end” is to build a map, and the “back-end” is to optimize a map with a nonlinear error function.

1.1. Introduction to the “Front-End” Algorithm

In the “front-end” of the RGB-D SLAM algorithm, feature detection and descriptor extraction algorithms are used to process RGB-D images. Then, the feature matching is implemented by using the descriptor results. Finally, motion transformation is estimated and optimized from matching results.
The whole algorithm in the “front-end” is an iterative process, which is divided into feature detection, descriptor extraction, feature matching, motion estimation, and motion transformation optimization. The RGB-D SLAM algorithm in the front end is shown in Figure 1.
In iteration, two adjacent frames with RGB and depth images [2] are obtained, firstly. Then, the two RGB images are processed in the feature detection and descriptor extraction so that we can get a collection of pairs of 2D feature matching points. From each pair of 2D feature matching points, two feature points are, respectively, extracted, containing coordinate information in RGB images and depth information in depth images. After 3D coordinates in space of feature points are determined, pairs of 3D coordinate matching points are generated and then added to the collection of three-dimensional coordinates of matching point sets. According to the 3D coordinates matching point sets, the RANSAC method is used to estimate 6D motion transformation between two images. Then, 3D coordinates of matching points, ICP point cloud registration is operated by applying 6D motion transformation, after which the optimized 6D motion transformation can be obtained. The iterative process above must be repeated until there is no new inputing RGB and depth image. Therefore, the 6D motion transformation is the output of the front-end.
Figure 1. The sketch map for RGB-D SLAM algorithm in the front end.
Figure 1. The sketch map for RGB-D SLAM algorithm in the front end.
Sensors 15 19937 g001

1.1.1. Feature Detection and Descriptor Extraction

In the original RGB-D SLAM [3] descriptor algorithm, SIFT [4] and SURF algorithms are usually used to deal with the feature detection and descriptor extraction. The SIFT algorithm mainly detects features and is now widely used in the fields of computer vision and robot Visual SLAM. Whereas, the SURF algorithm has advantages on scale invariance, rotation invariance, illumination invariance, contrast invariant, and affine invariant of 64-dimensional feature descriptors. Additionally, the speed of the SURF algorithm is 3–7 times faster than that of the SIFT algorithm. Thus, the SURF algorithm is widely used the same as the SIFT algorithm in the original RGB-D SLAM.

1.1.2. Feature Matching

In the original RGB-D [5] SLAM algorithm, the BruteForce method is adapted to match the feature descriptor between two successive frames. In this way, the first frame is called the search frame, and the second frame is named the training frame. In this method, feature descriptors in the search frame and training frame are complied to match. That is to say, each feature descriptor is forcibly matched with the nearest feature descriptor. There are many ways to calculate the distance, such as Euclidean distance, Manhattan distance or Hamming distance, etc.

1.1.3. Motion Transformation Estimation

After determining the set of 3D coordinate matching points from the feature matching step, the motion transformation model of a high credibility can be estimated, according to the position relationship between the matching points. In the original RGB-D SLAM system, random sample consensus (RANSAC) is utilized. Fischler and Bolles put forward RANSAC in 1981. So far, RANSAC has been broadly used in the domain of computer vision.

1.1.4. Motion Transformation Optimization

The ICP method in the original RGB-D SLAM is used to optimize the motion transformation model. It iteratively solves the rigid body transformation between two point-sets by making the distance of two point-sets minimum.

1.2. Introduction to the “Back-End” Algorithm

In the back-end for RGB-D SLAM system, the graph-based pose optimization method is used, in which the initial pose graph in the front-end is used as the input data. Then, loop closure detection is performed, so false-positive loop closure constraints can be eliminated. Meanwhile, the nonlinear error function optimization method is applied to optimize the pose graph. Thus, the camera trajectory can be known after figuring out the globally-optimal camera position. The “back-end” optimization is very useful for the reconstruction of the 3D map environment. The RGB-D SLAM algorithm in the “back end” is shown in Figure 2.
Figure 2. The sketch map for RGB-D SLAM algorithm in the back end.
Figure 2. The sketch map for RGB-D SLAM algorithm in the back end.
Sensors 15 19937 g002
Firstly, the result for the pose graph is detected for the closed-loop and is added to the closed-loop constraints. After that, the pose graph can be optimized. Therefore, pose graph nodes and edges are obtained. Meanwhile, camera pose is received before the optimization. According to different camera pose results, the point cloud will be generated, and then it is transformed to the initial position. After that, the 3D map [6] environment is reconstructed.
The rest of the paper is organized as follows: in Section 2, the related work of the SLAM based on a RGB-D sensor is given. Section 3 presents our proposed algorithm and the performance evaluations are provided in Section 4. Section 5 concludes the paper.

2. Related Work

There have been numerous publications of VSLAM approaches in previous years. In recent years the emergence of graph-based VSLAM [7,8,9,10] approaches has been most notably shown.
The RGB-D SLAM system was first scientifically proposed by Henry et al. [11] who use visual features, in combination with GICP, to create and optimize a pose graph. Unfortunately, neither the software nor the data used for evaluation has been publicly available so that a direct comparison can be carried out.
The pioneering work done by Nister et al. had combined feature extraction and matching with RANSAC to obtain accurate pose estimation through stereo cameras. Izadi et al. tackled SLAM by splitting the tracking and mapping into separate problems. In this work the camera is tracked over a global model of the environment through an ICP algorithm. The environment model is made denser and refined over several observations such that tracking is improved. Unfortunately, pure ICP relies on very structured environments, thus, limiting this technique to only a room-type scenario and not a general indoor case. Furthermore, keeping track of a dense model can be computationally expensive and will not scale very well.
Gradient histogram-based feature descriptors have been frequently used in VSLAM [12]. They are said to deliver the most descriptive features, but whose computations are expensive. Tree-based [13] algorithms can be used to find corresponding features in different images.
Henry et al. propose a method to jointly optimize visual features based on RANSAC with ICP. This approach is simplistic in nature but can work very intuitively with RGB-D cameras. During optimization, bundle adjustment is applied to obtain the consistent maps. Audras et al. argued against a feature-based method due to their uncertain nature in matching and detection by introducing an appearance-based approach [14]. Their dense stereo tracking technique [15], converted to work with the RGB-D camera, predicted relative camera motion by minimizing a warping function between frames, given 3D point cloud information.
Newcombe et al. proposed to incrementally build a dense model of the scene and register each new measurement to this model. Whelan et al. extended the KinectFusion algorithm of Newcombe et al. to arbitrarily large scenes. As KinectFusion does not optimize previous camera poses, there is no possibility to correct accumulated errors in this model.
Kintinuous overcomes such a limitation by virtually moving the voxel grid with the current camera pose. The parts that are shifted out of the reconstruction volume are triangulated. However, so far, the system cannot deal with loop closures and, therefore, may drift indefinitely.
This paper takes RGB-D SLAM into consideration at present, the deficiency for the original RGB-D SLAM algorithm undergoes an in-depth analysis, and each process is optimized and verified for the RGB-D SLAM one by one.
The original RGB-D SLAM algorithm’s problems can be concluded as follows.
(1)
The original RGB-D SLAM algorithm runs the feature detection and descriptor extraction through the SIFT or SURF method. However, the two methods use algorithms of convolution, histogram statistics, image integration, and sliding window, which are all time-consuming strategies. Therefore, SIFT or SURF methods cannot satisfy the real-time requirement in the front-end.
(2)
The original RGB-D SLAM algorithm applies BruteForce method in feature matching; there is a large error with high time complexity. When the method is used to forcibly match the search frame with the training frame in RGB-D SLAM matching, each feature descriptor will look for the minimum distance of feature descriptors in the training frames. The problems for matching are as follows:
Firstly, the BruteForce method is matched violently, which has a high dimension of space-time complexity and easily leads to severely reducing the matching speed.
Secondly, not every feature descriptor in the search frames can have the right matching in the training frames.
Thirdly, when BruteForce is used, every search frame must find a fmatching point in the training frames through one direction, which leads to some error matching.
Finally, when the same or similar objects with little difference in background appeared in the environment, the BruteForce method will regard same or similar things as the correct matching.
(3)
The RANSAC method is applied to estimate the motion transformation in the original RGB-D SLAM. The evaluation data is not updated in this process and the previously evaluated data causes large errors in the assessment results.
(4)
The 3D coordinate matching point is used to register the ICP point cloud in the original RGB-D SLAM. However, there is some error matching for the 3D coordinates of matching points. Therefore, in the model for the motion transformation some problems arise as follows:
Firstly, when the point cloud is registered by the original ICP algorithm, it generates a lot of errors, such as overlap.
Secondly, point cloud is generated by coordinates of the 3D matching points. After calculating, coordinates of the 3D matching points have some errors themselves. Therefore the motion transformation optimization model will have many errors.
Finally, point cloud registration is utilized to optimize the motion transformation when RANSAC succeeds. However, if RANSAC fails, this procedure will be finished directly. Then the motion transformation sequence must be interrupted, which will cause large errors in building the 3D environment.
Due to the above reasons, the original RGB-D SLAM [16] algorithm appears slowly and inevitably, in the meantime, there is tremendous error in results, which cannot fulfill demands for real-time and high-precision. According to these problems, this paper proposes an improved method for the RGB-D SLAM algorithm.

3. Proposed Algorithms

This paper proposes to use the ORB for feature detector and descriptor extraction, FLANN KNN for feature matching enhancement, and an improved RANSAC for motion transformation estimation and the optimized GICP for motion transformation. The improved RGB-D SLAM [17] has great efficiency and high precision. The process is shown in Figure 3.
The improvement contents are listed as follows:
(1)
The ORB is applied to filter illegal depth information in the feature detection and descriptor extraction.
(2)
The bidirectional FLANN KNN is used to improve accuracy in feature matching.
(3)
The RANSAC is utilized to get more accurate inlier matching points in motion transformation estimation.
(4)
The high accuracy GICP is applied to improve the speed and accuracy for point cloud registration in motion transformation.
Figure 3. The process for improved RGB-D SLAM algorithm.
Figure 3. The process for improved RGB-D SLAM algorithm.
Sensors 15 19937 g003

3.1. The ORB Method for Feature Detection and Descriptor Extraction

The ORB is to detect features and, in the meantime, filter the illegal depth information [18]. The reason for filtering the illegal depth information is that such information can lead to large errors in results. However, when the illegal depth information was handled in feature detection, it can effectively decrease the number of descriptors, the computation [19] complexity, and also improve the speed.

3.2. Enhanced Feature Matching Method Based on FLANN

The process for the enhanced feature matching method based on FLANN is shown as follows:
Firstly, solving the slow speed of BruteForce problem, this paper runs the feature matching based on FLANN. The FLANN [20] is an implementation of the fast approximate nearest neighbor search library. For the nearest neighbor search in high dimensional space, FLANN has the best performance in the hierarchical search algorithm [21]. Compared with traditional nearest neighbor search algorithms, FLANN is improved by an order of magnitude.
Secondly, in order to find all the feature points of nearest distance in training frames, the search results are obtained by the FLANN KNN method and, then, are filtered by constraints. When a set P = { p 1 , p 2 , ... , p n } and a query point q M are defined in orientation M space, the N N ( q , p ) P is the nearest neighbor elements of q which is defined as N N ( q , P ) = arg min x p d ( x p ) d ( q , x ) , where d ( q , x ) expresses the distance of q and x. The searching method for KNN is to find nearest neighbors of K, which is defined as follows:
K N N ( q , P , K ) = A
where A is satisfied as Formula (2), it shows the constraints for P = { p 1 , p 2 , ... , p n } set.
| A | = K , A P , x A , y P A , d ( q , x ) d ( q , y )
In the KNN search method, let K = 2 , i.e., every feature point s i in search frames finds its two nearest neighbor such as t i 1 and t i 2 in training frames, the distance of s i between them is d i 1 and d i 2 . Compared the distance between d i 1 and d i 2 , we regard s i and t i as the correct matching points when the distance of d i 1 is less than d i 2 . Otherwise, when the distance for d i 1 and d i 2 is relatively close, we think that s i and d i 1 , s i , and d i 2 are not the correct matching points. Let r a t i o i = d i 1 d i 2 , when the ratio i is less than 0.6, we keep the corresponding matching points; otherwise, they must be removed.
Thirdly, in order to eliminate the error matching points for the single direction matching, the bidirectional FLANN KNN is used to search the correct matching points. For every feature point in search frames, the FLANN is used to find matching pairs in training frames and, meantime, it filters the nearest and the next nearest neighbor similar distance of the matching point. Then, the good matching points set of s1 can be recorded. Eventually, the above steps are executed in reverse direction and the good matching point set of s2 can be recorded.
Finally, in order to solve the same or similar objects in the environment, the homography matrix transformation is applied to optimize the matching results. The homography transformation is a projective plane to the reversible transformation of another plane. Two images in the same plane can be linked together through homography transformation. Then, the results are shown in a homography matrix. The homography matrix obtains matching point relationships between two images. In this paper, removing the error matching points is based on whether the object appeared in a scene. However, the background in frames has high similarity; therefore, we can eliminate these repeated object error matches without matching the background. In details, each p a i r ( i , j ) S extracts the coordinates ( x i , y i ) in query frames and the coordinates are added into query position sets in searching frame sets. The coordinates ( x i , y i ) extracted in training frames are added to training position sets. Then, the two coordinate sets are used to obtain the optimized homography transformation matrix through RANSAC. Meanwhile, the set s removes the matching pairs which cannot satisfy the homography matrix constraints. Finally, after all matching points are traversed by performing the above operations, the set S will have high confidence.
To sum up, the enhanced feature matching FLANN KNN method is used to match feature descriptors in the search and training frames, and the homography matrix transformation is utilized to optimize the matching results. The corresponding feature matching point set can be obtained by the improved feature matching method. After that, the depth information of two feature points is extracted from the depth information matrix, and the corresponding matching points are generated to be added into the 3D coordinates of matching point sets.

3.3. Improved RANSAC Method for Motion Estimation

RANSAC is an iterative process. The improved methods are as follows:
Firstly, the motion transformation model is obtained using inliers in an iterative process [22]. The entire 3D coordinates of matching points are evaluated to obtain the new pairs of inliers. The new inliers can reflect the true situation of the motion transformation model.
Secondly, the iteration process obtains the optimal motion transformation model, optimal inlier sets and some evaluated errors. If the number of local inliers is higher than the threshold value, and the errors are below the threshold, the motion transformation model should be calculated again by the optimal inliers set and recorded as a final motion transformation model. Then, the final motion transformation model is used to evaluate and calculate the final inliers set.

3.4. Motion Transformation Optimization Method Based On GICP

GICP is based on the standard ICP and the surface ICP, and solves the motion transformation through the probabilistic model. In order to simplify the description, we assume that two point clouds A = { a i } i = 1 , ... , N and B = { b i } i = 1 , ... , N are obtained after two matching points are calculated in circulation part on the standard ICP algorithm. Meanwhile, assuming the relation satisfying the condition of | | m i T b i | | > d max is deleted in A and B sets, the GICP algorithm is improved as follows:
Assuming A ^ = { a 1 ^ } and B ^ = { b 1 ^ } exist in the probabilistic model, A and B will be generated by a i ~ N ( a i ^ , C i A ) and b i ~ N ( b i ^ , C i B ) , and C i A , C i B are the covariance matrix to the measure point. Assuming that the elements in set A exactly corresponded that in set B, the motion transformation satisfies the equation as follows:
b i ^ = T * a i ^
For arbitrary rigid body transformation T, d i ( T ) = b i T a i is defined, and an equation is established as follows:
T = i T arg min d i ( T ) T ( C i B + T C i A T T ) 1 d i ( T )
Equation (3) is the improvement method for GICP. When C i B = I and C i A = 0 , Equation (4) degenerates the solving constrains as motion transformation; When C i B = P i 1 and C i A = 0 , Equation (4) degenerates T = { T arg min i | | P i d i | | 2 } .
If used directly for the motion transformation model, the 3D coordinates must have huge errors. Therefore, the final points satisfied with inlier matching are calculated in the improved RANSAC. Due to the error matching points being removed, this paper proposes that the final inliers matching point replaces the original 3D coordinate matching point to optimize the motion transformation.
When the number of matching inliers are too few (less than 20 pairs), the algorithm for GICP will fail. However, there are three methods that can be altered on GICP: first is the number threshold of the point cloud; second is that the 3D coordinates which can be used to run the matching point cloud method on GICP; third is that the GICP can be degenerated to ICP. In this paper, three methods are tested as shown in Table 1:
Table 1. Different results for Insufficient number of processing method in Freiburg dataset.
Table 1. Different results for Insufficient number of processing method in Freiburg dataset.
MethodReducing the Number of the GICP Threshold Matching Point Cloud for GICPInliers Point Cloud for ICP
deviation0.0833260.0020550.000684
Time (ms)32.335103.71619.530
From the Table 1, it can be seen that reducing the number of the GICP threshold leads to the poor results. Due to all the 3D coordinate points obtaining some errors for matching, the efficiency and the accuracy of the algorithm are reduced. Therefore the second method cannot be used. The third is that the method for degenerating GICP to ICP [23] is more efficient and accurate for point cloud matching.

3.5. Improved DCS Algorithm for Closed Loop Detection on G2O

In the original G2O [24], the DCS algorithm is used to detect the closed loop. In the DCS algorithm, Agarwal et al. analyzes the objective function in Equation (5). At the time, they pay special attention on how the switch variables influence the local minimum value of the objective function. They select an edge lmn as an example. The function can be divided into two parts; one contains all edges except mn, and the other only contains the edge mn. Then, the optimization problem can be represented as follows:
X * , S * = arg min X , S i | | d i o d o | | i 2 + i j m n | | d i j s l c | | Λ i j 2 + i j m n | | d i j s p | | Ξ i j 2 + | | S m n ( f ( x m , u m n ) x n ) | | Λ m n 2 + | | 1 S m n ) | | Ξ m n 2 = arg min X , S h ( X i j m n , S i j m n ) + s m n 2 χ l m n 2 + ( 1 s m n ) 2 ϕ = arg min X , S h ( X i j m n , S i j m n ) + g ( X i j = m n , S i j = m n )
where ϕ = Ξ i j 1 , the function h ( ) considers the error of odometry constraints, all the switch priors, and loop closure constraints, except the switch prior and constraint of edge mn. In turn, the function g ( ) only considers the edge mn. When the objective function converges, the partial derivative with respect to s m n must be zero. Since the term h ( ) doesn’t contain the variable s m n , taking the partial derivative of the Equation (5) with respect to s m n means seeking the partial derivative of the function g ( ) :
g ( ) = g ( ) s = 2 s χ l m n 2 2 ( 1 s ) Φ = 0
Solving the Equation (6), we can obtain:
s = Φ χ l 2 + Φ
We analyze the behavior of Equation (7), especially the range of the fraction value. It is obvious that the value of the right side of the equal sign in Equation (7) is greater than zero. However, in theory, s [ 0 , 1 ] . Therefore, through the derivation above, when the objective function obtains the minimum value, s is given by
s = min ( 1 , Φ χ l 2 + Φ )
By analyzing the property of the error function, we derive the analytic solution for calculating the weighting factor. The scaling factor s individually depends on the initial error of each loop closing constraint. Our solution to this problem provides a new smaller range of the variables s. While the change is relatively minor, the implication to optimization is profound. In sum, our algorithm not only has the advantages on computing complexity and convergence speed as the same as DCS, but also is superior to DCS [25]. It not only reduces the complexity and the running time, but also improves the convergence speed. In the following, we prove the good performance of our algorithm through many experiments.

4. Performance Evaluation

For the original RGB-D SLAM [26] algorithm, large errors and lack of real-time [27,28] and high precision existed. In this paper, improved methods of feature detection and descriptor extraction, feature matching, motion estimation and optimization of motion transformation [29] in the RGB-D SLAM algorithm are proposed. This section tests, compares, and evaluates these methods to verify their correctness.
In our experiments, the algorithm is running on the hardware environment for a Lenovo notebook computer; a Lenovo IdeaPad Y471A, Inter Core i5-2410M CPU @ 2.30 GHz dual core processor, three level caches and 4 GB of memory. The system environment is Ubuntu 12.04, of which the kernel version is 3.5.0-54-generic, and the compiler is GCC 4.6.3.
In order to prove the correctness [30], SIFT, SURF_default, SURF_0, ORB_default, and ORB_10000 are tested, respectively, compared, and assessed for the improved algorithm. The methods are described as follows in Table 2:
To test the feature detection and Descriptor, feature matching, motion transformation estimation, motion change optimization and the whole front-end algorithm, we use the rgbd_dataset_freiburg1_360 dataset, which contains 754 frames for RGB and Depth images.
Table 2. Method Description.
Table 2. Method Description.
MethodsDescription
SIFTThe original method
SURF_defaultDepthMin~DepthMax is 0~5000
SURF_0DepthMin~DepthMax is 1000~3000, which is more rigid constraints.
ORB_defaultDepthMin~DepthMax is 0~5000
ORB_10000DepthMin~DepthMax is 1000~3000, which is more rigid constraints.

4.1. Feature Detection and Descriptor Extraction

Feature detection and descriptor extraction can be respectively validated by methods of SIFT, SURF_default, SURF_0, ORG_default, and ORB_10000.

Comparison for Testing Results of Single Image

In order to test feature detection and descriptor extraction, the results for 337th RGB image in datasets are shown in Figure 4.
Figure 4. Results of 337th RGB image in datasets for feature detection and descriptor extraction (a) SIFT; (b) SURF_default; (c) SURF_0; (d) ORB_default. (e) ORB_10000.
Figure 4. Results of 337th RGB image in datasets for feature detection and descriptor extraction (a) SIFT; (b) SURF_default; (c) SURF_0; (d) ORB_default. (e) ORB_10000.
Sensors 15 19937 g004
The cost times for five methods of feature detection and feature extraction above are listed in Table 3.
From Table 3, it can be seen that cost times for ORB_default and ORB_10000 are significantly less than SIFT, SURF_default and SURF_0. However, due to the uncertainty for one image, all images in datasets should be tested to verify the correctness.
Table 3. Comparison cost times for five different methods.
Table 3. Comparison cost times for five different methods.
Method for Feature Detection and Descriptor ExtractionTime for Feature Detection (ms)Time for Descriptor Extraction (ms)The Number for Feature Points
SIFT34.91049.082350
SURF_default100.008169.2841124
SURF_0147.597264.0022586
ORB_default6.6228.302500
ORB_100008.42215.0942462

4.2. Feature Matching

In original RGB-D SLAM algorithm, the matching method based on FLANN is proposed. In this section, the results for SIFT, SURF_default, SURF_0, ORB_default and ORB_10000 methods are used to validate the correctness of the improved feature matching method.

4.2.1. Comparison for Two Adjacent Frames

(a) Comparison FLANN KNN with BruteForce method
When BruthForce is used to feature matching, the results obtain a lot of false matching. For example, when illegal depth information feature points for 337th and 338th frames are filtered by the BruthForce method in SURF_default, results are shown in Figure 5a,b.
Figure 5. Results for SURF_default feature detection on frame 337th and 338th in datasets. (a) 337th Frame SURF_default; (b) 338th Frame SURF_default.
Figure 5. Results for SURF_default feature detection on frame 337th and 338th in datasets. (a) 337th Frame SURF_default; (b) 338th Frame SURF_default.
Sensors 15 19937 g005
Assuming that features are matched by using the BruteForce method, 1124 pairs of matching points are obtained, which costs 54.725 ms. The maximum distance between the matching points is 0.780452 cm and the minimum distance is 0.025042 cm. As we can see, the maximum distance is 31 times that of the minimum. Meanwhile, there are many false matches in Figure 5a.
(b) Comparison Bidirectional KNN with KNN Results
As shown in Figure 6a, the FLANN KNN method still obtains some errors. Therefore, with the purpose of greater accuracy for subsequent algorithms, this paper proposes Bidirectional KNN to eliminate false matching from the FLANN KNN.
When these forward KNN matching are finished, the reverse FLANN KNN started to match. 1191 pairs of matching points are obtained, initially. The system takes 32.707 ms, the maximum distance is 0.839925 cm, and the minimum is 0.025042 cm. Then, using the nearest neighbor and next nearest neighbor ratio of 0.6 to filter, matching points are gained for 505 pairs. Wherein the maximum distance between the matching points is 0.470719 cm, the minimum distance is 0.025042 cm. The maximum distance is 18 times the minimum distance. As we can see in Figure 6b, there is less error for matching points.
Figure 6. Results for BruteForce feature matching. (a) No threshold filtration; (b) 10 times the minimum distance; (c) Five times the minimum distance; (d) Two times the minimum distance.
Figure 6. Results for BruteForce feature matching. (a) No threshold filtration; (b) 10 times the minimum distance; (c) Five times the minimum distance; (d) Two times the minimum distance.
Sensors 15 19937 g006aSensors 15 19937 g006b
From the above results, the forward and reverse matching results are not the same and these two matches both have some false matches. Therefore, the intersection between two matches acquires the bidirectional matching, which can eliminate false matches. The results for bidirectional matching contain 412 pairs of matching points. The maximum distance between matching pairs is 0.386766 cm, the minimum distance is 0.025042 cm, and the maximum distance is 15 times the minimum to further eliminate some false matching points, as shown in Figure 7c. Comparing Figure 7a with Figure 7b, AB and EF mismatches have been eliminated by bidirectional matching methods.
Figure 7. Feature matching results for Bidirectional FLANN KNN method and Homography transformation. (a) Forward FLANN KNN Filtration, ratio of 0.6; (b) Reverse FLANN KNN Filtration, ratio of 0.6; (c) Bidirectional FLANN KNN Filtration, ratio of 0.6.
Figure 7. Feature matching results for Bidirectional FLANN KNN method and Homography transformation. (a) Forward FLANN KNN Filtration, ratio of 0.6; (b) Reverse FLANN KNN Filtration, ratio of 0.6; (c) Bidirectional FLANN KNN Filtration, ratio of 0.6.
Sensors 15 19937 g007aSensors 15 19937 g007b

4.2.2. Comprehensive Comparison for Feature Detection, Feature Extraction and Feature Matching

The total time for feature detection, descriptor extraction and feature matching is analyzed in Table 4:
Table 4. Comprehensive Comparison for Feature Detection, Feature Extraction and Feature Matching.
Table 4. Comprehensive Comparison for Feature Detection, Feature Extraction and Feature Matching.
Feature Detection and Descriptor Extraction MethodNumber for Feature PointsTime for Feature Detection and Descriptor Extraction (ms)Number for Matching PairsTime for Feature Matching (ms)Total Time (ms)
SIFT168658310140
SURF_default60416618235367
SURF_02360375255134884
ORB_default388121131337
ORB_100001112163273971
From Table 4 it can be seen that the methods which are the improved feature detection, descriptor extraction, and bidirectional FLANN KNN and homography transformation are faster than before, just 37 ms. In this case, the number of matching points is 113. The second one is ORB_10000, which took 71 ms and the number of matching pairs is 327. The third one is SIFT, which took 140 ms and the number of matching pairs is 327. The forth one is SURF_default, which took 367 ms and the number of matching pairs is 182. The last one is SURF_0, which took 884 ms and the number of matching pairs is 255. In a word, the ORB method for feature detection, descriptor extraction, and improved bidirectional FLANN KNN and homography transformation cannot only get more correct matching points, but also greatly reduces the time required for the algorithm.

4.3. Motion Transformation Estimation

4.3.1. Comparison of Results to Two Adjacent Frames on Motion Transformation Estimation

Feature detection and descriptor extraction are also handled by SIFT, SURF_default, SURF_0, ORB_default, and ORB_0 on the dataset of the 337th and 338th frames. Then, for feature matching results, the number of inliers is obtained by the original and improved motion transformation estimation methods. Additionally, the proportion of inliers in pairs of matching features, the error value, and time consumption are calculated. The results are shown in Table 5.
Table 5. Results for motion transformation estimation on 337th and 338th frames.
Table 5. Results for motion transformation estimation on 337th and 338th frames.
MethodsThe Original Motion Transformation Estimation MethodThe Improved Motion Transformation Estimation Method
The Number of InliersThe Proportion of Inliers (%)DeviationThe Number of InliersThe Proportion of Inliers (%)Deviation
SIFT187770.0217976134880.0210702
SURF_default622640.0235837311910.022127
SURF_0FailureFailureFailure377910.0218584
ORB_default360790.0250409150810.0261754
ORB_100001621750.0221271731880.0200384
From Table 5, the inliers ratio, the error, and time consumption for improved motion transformation estimation is significantly lower than the original method. Therefore, it is obvious that improved motion transformation estimation is better than the original.

4.3.2. Comparison of Motion Transformation Estimation Results for All Images in Datasets

For all Images in datasets, the error value and consuming time are shown in Figure 8.
Figure 8. Results of motion transformation estimation for original and improved method on all images. (a) The number of original inliers. (b) The number of improved inliers. (c) The proportion of original inliers. (d) The proportion of improved inliers. (e) Deviation for original method. (f) Deviation for improved method. (g) Time for original method. (h) Time for improved method.
Figure 8. Results of motion transformation estimation for original and improved method on all images. (a) The number of original inliers. (b) The number of improved inliers. (c) The proportion of original inliers. (d) The proportion of improved inliers. (e) Deviation for original method. (f) Deviation for improved method. (g) Time for original method. (h) Time for improved method.
Sensors 15 19937 g008aSensors 15 19937 g008b
In Figure 8a,c,e,g, the blue line of SURF_0 is very short. The reason is that, in the original algorithm, the SURF_0 has a large number of failures in the motion transformation estimation stage. Moreover, when failure happens, the number of the corresponding inliers, the proportion, the error value, and the time for inliers on the original method are not statistical. It indicates that the algorithm is not stable. However, in Figure 8b,d,f,h, the blue line of SURF_0 is normal. It shows that the failure can be solved by the improved method.
Comparing Figure 8a with Figure 8b, whatever the feature detection and descriptor extraction method is used, the number of inliers is obviously less than the original method. Other comparisons have reached the same conclusion.

4.4. Motion Change Optimization

4.4.1. Comparison Results of Motion Change Optimization for Adjective Frames

The original and the improved method deal with motion transformation optimization by 337th and 338th frame images in datasets. The error value and time consumption are shown in Table 6:
Table 6. Results for motion transformation optimization method.
Table 6. Results for motion transformation optimization method.
MethodThe Original Motion Transformation Optimization MethodThe Improved Motion Transformation Optimization Method
DeviationTime (ms)DeviationTime (ms)
SIFT0.00444218190.00096460159
SURF_default0.00397572510.00080667787
SURF_0failurefailure0.00077632297
ORB_default0.0113304360.00066115849
ORB_100000.002543921010.00046986354
From Table 6, even though the improved motion transformation optimization spends 1.5–3 times for the original method, the error is smaller by one order of magnitude than the original method, and it can deal with failure of the original method. Therefore, the improved motion transformation optimization method is correct.

4.4.2. Comparison Results of Motion Transformation Optimization for All Images in Datasets

The original and improved motion transformation optimization method are tested by 753 adjacent frames in Freiburg 360 datasets, and obtained the error value and time, as shown in Figure 9.
After comparison Figure 9a with Figure 9b, it can be found that the error of the improved method is one order of magnitude smaller than the original motion transformation optimization. From Figure 9c,d, the time for the improved method is significantly less than the original motion transformation optimization. Even though the time is longer than ICP in a certain extent, the accuracy of the GICP algorithm is much higher than that of ICP. In order to improve the accuracy of the algorithm, a little time can be acceptable.
In Figure 9c the blue line of SURF_0 is very short. The reason is that in the original algorithm, SURF_0, has a large number of failures in the motion transformation estimation; the line is shorter, the failures are more numerous. This indicates that the unstable algorithm needs to be improved. However, in Figure 9d, the blue line of SURF_0 is normal. This shows that the failure can be solved by the improved method.
Figure 9. Comparison results of motion transformation optimization for all images in datasets. (a) The original error; (b) The improved error; (c) The original time consumption; (d) The improved time consumption.
Figure 9. Comparison results of motion transformation optimization for all images in datasets. (a) The original error; (b) The improved error; (c) The original time consumption; (d) The improved time consumption.
Sensors 15 19937 g009

4.5. Comparison of the Whole “Front-End” Algorithm

Firstly, the original SIFT, SURF_default, SURF_0, ORB_default, and ORB_10000 methods are used in the original front-end algorithm, which include feature detection, feature descriptor, and original feature matching. After that, using the original motion transformation estimation method obtains the motion transformation.
Then, the original and the improved five motion transformation are compared with each other. The results are shown below.

Comparison of the “Front-End” Optimization Results with Ground Truth in Datasets

In order to prove the correctness and effectiveness, the whole algorithm is compared and evaluated by the translational error and rotational error.
The results are shown in Figure 10a,c. Then, comparing the improved five groups of the motion transformation sequence with ground truth in datasets, the translational error and rotational error of the five methods are calculated, respectively. The results are shown in Figure 10b,d.
Figure 10a shows that the five translational errors are obtained from the original front-end algorithm. As we can see, the translational errors of the five methods are similar, which is about 0.05 m. Figure 10b shows that the five translational errors are obtained from the improved front-end algorithm. And the translational errors of the five methods are similar, which is about 0.05 m. Figure 10c shows that the five rotational errors are obtained from the original front-end algorithm. As we can see, the rotational errors of the five methods are similar, which is about 0.05°. Figure 10d shows that the five rotational errors are obtained from the improved front end algorithm. As we can see, the rotational errors of the five methods are similar, which is about 0.05°.
From Figure 10a,b, compared with the original method, the improved method decreases the translational error in SURF_0 and ORB_10000. From Figure 10c,d, the improved method decreases the rotational error in SURF_0 and ORB_10000, compared with the original method. It is important to note that the translational error staying around 0.05 m and rotational error being 0.05° depend on the inherent error of Kinect camera.
(1) Comparison of Camera Pose In “Front-End”
Camera pose can be used to prove the correctness and effectiveness in five methods from the original and improved “front-end” algorithm.
Firstly, camera poses are obtained by the original five groups of the motion transformation sequence, and the real camera poses are received by searching ground truth. The results are shown in Figure 11a,c. Then, camera poses are obtained by the improved five groups of the motion transformation sequence, and the real camera poses are received by searching ground truth. The results are shown in Figure 11b,d.
Figure 10. Comparison of the original and the improved algorithm for transformational error and rotational error. (a) The transformational error in original; (b) The transformational error in improved; (c) The rotational error in original; (d) The rotational error in improved.
Figure 10. Comparison of the original and the improved algorithm for transformational error and rotational error. (a) The transformational error in original; (b) The transformational error in improved; (c) The rotational error in original; (d) The rotational error in improved.
Sensors 15 19937 g010aSensors 15 19937 g010b
Figure 11. Comparison the original and the improved for Camera pose. (a) Horizontal poses for original algorithm; (b) Horizontal poses for improved algorithm; (c) Vertical poses for original algorithm; (d) Vertical poses for improved algorithm.
Figure 11. Comparison the original and the improved for Camera pose. (a) Horizontal poses for original algorithm; (b) Horizontal poses for improved algorithm; (c) Vertical poses for original algorithm; (d) Vertical poses for improved algorithm.
Sensors 15 19937 g011
Figure 11a shows that the original front-end algorithm obtains the camera pose in horizontal, which is the X axis and Z axis of the pose. No matter which method in the original front-end algorithm is used, it must have errors compared with the ground truth. We can conclude that the minimum error is the method of SURF_0. The results of SIFT and SURF_default are slightly larger, and that of ORB_default and ORB_10000 is much larger.
Figure 11b shows that the improved front-end algorithm acquires the camera pose in horizontal, which is the X axis and Z axis of the pose. No matter which method the original front-end algorithm of method is used, it has errors compared with the ground truth. We can conclude that the minimum error is the method of SURF_0. The results for SIFT and SURF_default are slightly larger, and the ORB_default and ORB_10000 is much larger. However, it has obvious improvements compared to the original algorithm.
Figure 11c shows that the original front-end algorithm obtains vertical posture which is the Y axis of the pose. No matter which method the original front-end algorithm of method is used, it must have errors compared with the ground truth. Smaller errors are seen in SIFT, SURF_default and SURF_0. The results for ORB_default and ORB_10000 are much larger.
Figure 11d shows that the improved front-end algorithm acquires vertical posture which is the Y axis of the pose. No matter which method of the improved front-end algorithm is used, it must have errors compared with the ground truth. However, compared with the original front-end algorithm, it has been greatly improved.
From the comparison for four pictures, the improved algorithm reduces the error for camera pose, which proved the correctness.
(2) The Results for Stitching On “Front-End” Map
Five methods for SIFT, SURF_default, SURF_0, ORB_default, and ORB_10000 are, respectively, compared by the original and improved algorithm.
First, the original front-end algorithm is implemented by SIFT, SURF_default, SURF_0, ORB_default, and ORB_100. Then, according to the front-end algorithm, the maps that the camera poses, joined, are shown in Figure 12a,c,e,g,i. Meanwhile, the improved front-end algorithm is implemented by SIFT, SURF_default, SURF_0, ORB_default, and ORB_100, after that, the map of the camera poses, joined, are shown in Figure 12b,d,f,h,j.
Figure 12. The results for stitching on the original and the improved front-end map. (a) Original SIFT; (b) Improved SIFT front end; (c) Original SURF_default; (d) Improved SURF_default; (e) Original SURF_0; (f) Improved SURF_0; (g) Original ORB _default; (h) Improved ORB _default; (i) Original ORB _10000; (j) Improved ORB _10000.
Figure 12. The results for stitching on the original and the improved front-end map. (a) Original SIFT; (b) Improved SIFT front end; (c) Original SURF_default; (d) Improved SURF_default; (e) Original SURF_0; (f) Improved SURF_0; (g) Original ORB _default; (h) Improved ORB _default; (i) Original ORB _10000; (j) Improved ORB _10000.
Sensors 15 19937 g012aSensors 15 19937 g012b
Comparing (a), (c), (e), (g), and (i) in Figure 12, the horizontal and vertical error for SIFT and SURF_default in the front-end algorithm is smaller. However, the error from SURF_0, ORB_default, and ORB_10000 is very large, with serious bending directly occurring in stitching results.
Comparing (b), (d), (f), (h), and (j) in Figure 12, splicing errors from the five methods is small. The splicing results for SURF_0, ORB_default, and ORB_10000 method improves, obviously, which proved the correctness of the improved front-end algorithm.

4.6. The Actual Environment Test Based on Dr Robot X80

From the above experiments, we use the algorithm in the real Dr Robot X80 environment to verify its effectiveness.
RGB-D SLAM system architecture based on Dr Robot X80 is shown in Figure 13:
Figure 13. RGB-D SLAM system architecture based on Dr Robot X80.
Figure 13. RGB-D SLAM system architecture based on Dr Robot X80.
Sensors 15 19937 g013
The RGB-D SLAM system for Dr X80 Robot is built in the environment above, and collects 475 RGB-D images. Then, SIFT, SURF, and ORB are used to test the original and improved RGB-D SLAM algorithm by the 475 frames of RGB-D images.
First, the original RGB-D SLAM algorithm uses SIFT, SURF, and ORB are run. The splicing results of a few feature points for the original front-end algorithm are poor, and there is no effective optimization for the front-end results.
Then, SIFT, SURF, and ORB are used in the improved RGB-D SLAM algorithm. The front-end results are shown in Figure 14a,c,e. The improved method has very good results and a small error. The back-end results are shown in Figure 14b,d,f. The improved RGB-D SLAM back-end algorithm has better splicing results in the case of less feature points.
Comparing Figure 14 with Figure 15, the improved RGB-D SLAM algorithm not only has better spliced the 3D map, but also obtains the accurate trajectory of the robot. The trajectory for the robot is shown in Figure 16.
Figure 14. The splicing results for the original RGB-D SLAM based on Dr Robot X80. (a) SIFT for the front-end; (b) SIFT for the back-end; (c) SURF for the front-end; (d) SURF for the back-end; (e) ORB for the front-end; (f) ORB for the back-end.
Figure 14. The splicing results for the original RGB-D SLAM based on Dr Robot X80. (a) SIFT for the front-end; (b) SIFT for the back-end; (c) SURF for the front-end; (d) SURF for the back-end; (e) ORB for the front-end; (f) ORB for the back-end.
Sensors 15 19937 g014
Figure 15. The splicing results for the improved RGB-D SLAM based on Dr Robot X80. (a) SIFT for the front-end; (b) SIFT for the back-end; (c) SURF for the front-end; (d) SURF for the back-end; (e) ORB for the front-end; (f) ORB for the back-end.
Figure 15. The splicing results for the improved RGB-D SLAM based on Dr Robot X80. (a) SIFT for the front-end; (b) SIFT for the back-end; (c) SURF for the front-end; (d) SURF for the back-end; (e) ORB for the front-end; (f) ORB for the back-end.
Sensors 15 19937 g015aSensors 15 19937 g015b
Figure 16. The optimization of Robot trajectory by the improved RGB-D SLAM based on Dr Robot X80.
Figure 16. The optimization of Robot trajectory by the improved RGB-D SLAM based on Dr Robot X80.
Sensors 15 19937 g016
Thus, the improved RGB-D SLAM algorithm can successfully solve the splicing problem for the few feature points.
In summary, through a lot of experiments, this chapter evaluates the efficiency and accuracy for the improved RGB-D SLAM algorithm and verifies the correctness for each step.

5. Conclusions and Outlook

In order to clearly illustrate the results for above sections, the names of different methods are defined in Table 7.
Table 7. The name for method definition table.
Table 7. The name for method definition table.
Method NameThe Special Content
Original SIFT_RGB-D_SLAMOriginal SIFT+ Original matching + Original RANSAC+ Original ICP+ Original back-end
Original SURF_RGB-D_SLAMOriginal SURF+ Original matching + Original RANSAC+ Original ICP+ Original back-end
Original ORB_RGB-D_SLAMOriginal ORB+ Original matching + Original RANSAC+ Original ICP+ Original back-end
Improved SIFT_RGB-D_SLAMImproved SIFT+ Improved matching + Improved RANSAC+ Improved ICP+ Original back-end
Improved SURF_RGB-D_SLAMImproved SURF+ Improved matching + Improved RANSAC+ Improved ICP+ Original back-end
Improved ORB_RGB-D_SLAMImproved ORB+ Improved matching + Improved RANSAC+ Improved ICP+ Original back-end
The implementation of the results before can be easily seen as Table 8.
Table 8. Schematic diagram for a comprehensive analysis of the experimental results.
Table 8. Schematic diagram for a comprehensive analysis of the experimental results.
Speed
LowHigh
PrecisionLowOriginal SIFT_RGB-D_SLAM, Original SURF_RGB-D_SLAMOriginal ORB_RGB-D_SLAM
HighImproved SIFT_RGB-D_SLAM, Improved SURF_RGB-D_SLAMImproved ORB_RGB-D_SLAM
As shown in Table 8, the red part is the RGB-D SLAM algorithm with low speed and low precision; the green part is the RGB-D SLAM algorithm with high speed and high precision; the orange part for the upper left corner is the RGB-D SLAM algorithm with low speed and high precision; the orange part for the lower left corner is the RGB-D SLAM algorithm with high speed and low precision.
Both the original SIFT_RGB-D_SLAM algorithm and the original SURF_RGB-D_SLAM algorithm are slow in speed and low in precision. The speed and accuracy of the original SIFT_RGB-D_SLAM algorithm are slightly higher than the original SURF_RGB-D_SLAM algorithm, and the original ORB_RGB-D_SLAM algorithm is high in speed and low in accuracy. The improved SIFT_RGB-D_SLAM algorithm and improved SURF_RGB-D_SLAM algorithm is a low speed and high precision algorithm, the improved SIFT_RGB-D_SLAM algorithm is slightly higher than the improved SURF_RGB-D_SLAM algorithm. Compared with the original SIFT_RGB-D_SLAM algorithm, the SIFT_RGB-D_SLAM algorithm can improve the accuracy of velocity, but is slightly decreased; compared with the original SURF_RGB-D_SLAM algorithm, accuracy of the improved SURF_RGB-D_SLAM algorithm is greatly improved, and its speed is also increased slightly. The improved ORB_RGB-D_SLAM algorithm is a high speed and high precision algorithm; compared with the original ORB_RGB-D_SLAM algorithm, the improved ORB_RGB-D_SLAM algorithm’s accuracy has been greatly increased, and the velocity is decreased slightly. On the other hand, accuracy of the improved ORB_RGB-D_SLAM algorithm and that of improved SIFT_RGB-D_SLAM algorithm are basically the same, but the speed of the former is much higher than that of the latter.
To sum up, the improved algorithm can solve the low speed and low accuracy problem of the RGB-D SLAM algorithm, and it satisfies the requirements of high speed and high precision. Therefore, we can conveniently apply this improved robot identification and mapping algorithm to the IoT system constructed by a robot, it can improve the robot’s automation ability, and enhance the IoT system’s robustness in real-time.

Acknowledgments

This project is supported by the National Natural Science Foundation of China (NSFC) Grant (No. 61305109, No. 61401324, No. 61072105), by 863 Program (2013AA014601), by Shaanxi Scientific research plan (2013K06-09, 2014k07-11) and by West Virginia Higher Education Policy Commission Research Trust Fund (No. 2W762). This job is also supported by Beilin District 2012 High-tech Plan, Xi’an, China (No. GX1504) and supported by Xi’an science and technology project (CXY1440(6)) and Supported by China Postdoctoral Science Foundation (No. 2013M542370) and supported by the Specialized Research Fund for the Doctoral Program of Higher Education of China (Grant No. 20136118120010).

Author Contributions

Liang Zhang and Peiyi Shen contributed to the improvement of the front-end algorithms and verification of the experiments, Guangming Zhu, Wei Wei, and Houbing Song contributed to the back end algorithms and verification of the experiments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wei, W.; Qi, Y. Information potential fields navigation in wireless Ad-Hoc sensor networks. Sensors 2011, 11, 4794–4807. [Google Scholar] [CrossRef] [PubMed]
  2. Hogman, V. Building a 3D Map from RGB-D Sensors; Computer Vision and Active Perception Laboratory, Royal Institute of Technology (KTH): Stockholm, Sweden, 2013. [Google Scholar]
  3. Durrant-Whyte, H.; Bailey, T. Simultaneous Localization and Mapping: Part I. IEEE Robot. Autom. Mag. 2006, 13, 99–110. [Google Scholar] [CrossRef]
  4. Rob Hess. An Open-Source SIFT Library. Available online: http: //blogs.oregonstate.edu/hess/code/sift/ (access on 27 July 2015).
  5. Kerl, C.; Sturm, J.; Cremers, D. Robust Odometry Estimation for RGB-D Cameras. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013.
  6. Konolige, K.; Agrawal, M. FrameSLAM: From Bundle Adjustment to Real-Time Visual Mapping. IEEE Trans. Robot. 2008, 24, 1066–1077. [Google Scholar] [CrossRef]
  7. Thrun, S.; Montemerlo, M. The GraphSLAM Algorithm with Applications to Large-Scale Mapping of Urban Structures. Int. J. Robot. Res. 2005, 25, 403–430. [Google Scholar] [CrossRef]
  8. Wei, W.; Qiang, Y.; Zhang, J. A Bijection between Lattice-Valued Filters and Lattice-Valued Congruences in Residuated Lattices. Math. Probl. Eng. 2013, 2013. [Google Scholar] [CrossRef]
  9. Endres, F.; Hess, J.; Engelhard, N.; Sturm, J.; Burgard, W. 6D visual SLAM for RGB-D sensors. Automatisierungstechnik 2012, 60, 270–278. [Google Scholar] [CrossRef]
  10. Fioraio, N.; Konolige, K. Realtime Visual and Point Cloud SLAM; Willow Garage: Menlo Park, CA, USA, 2012. [Google Scholar]
  11. Henry, P.; Krainin, M.; Herbst, E.; Ren, X.; Fox, D. RGB-D mapping: Using Kinect-style depth cameras for dense 3D modeling of indoor environments. Int. J. Robot. Res. 2012, 31, 647–663. [Google Scholar] [CrossRef]
  12. Whelan, T.; Johannsson, H.; Kaess, M.; Leonard, J.J.; McDonald, J. Robust real-time visual odometry for dense RGB-D mapping. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013.
  13. Zeng, M.; Zhao, F.; Zheng, J.; Liu, X. Octree-based fusion for realtime 3D reconstruction. Graph. Models 2013, 75, 126–136. [Google Scholar] [CrossRef]
  14. Wei, W.; Yang, X.L.; Shen, P.Y.; Zhou, B. Holes detection in anisotropic sensornets: Topological methods. Int. J. Distrib. Sens. Netw. 2012, 2012. [Google Scholar] [CrossRef]
  15. Newcombe, R.A.; Izadi, S.; Hilliges, O.; Molyneaux, D.; Kim, D.; Davison, A.J.; Kohli, P.; Shotton, J.; Hodges, S.; Fitzgibbon, A. KinectFusion: Real-Time Dense Surface Mapping and Tracking. In Proceedings of the 2011 10th IEEE International Symposium on Mixed and Augmented Reality (ISMAR), Basel, Switzerland, 26–29 October 2011.
  16. Endres, F.; Hess, J.; Sturm, J.; Cremers, D.; Burgard, W. 3D Mapping with an RGB-D Camera. IEEE Trans. Robot. 2013, 30, 177–187. [Google Scholar] [CrossRef]
  17. Newcombe, R.A.; Lovegrove, S.J.; Davison, A.J. DTAM: Dense tracking and mapping in real-time. In Proceedings of the 13th International Conference on Computer Vision (ICCV), Barcelona, Spain, 6–13 November 2011.
  18. Henry, P.; Krainin, M.; Herbst, E.; Ren, X.; Fox, D. RGB-D mapping: Using depth cameras for dense 3D modeling of indoor environments. In Proceedings of the 12th International Symposium on Experimental Robotics (ISER), Delhi, India, 18–21 December 2010.
  19. Wei, W.; Xu, Q.; Wang, L.; Hei, X.H.; Shen, P.; Shi, W.; Shan, L. GI/Geom/1 queue based on communication model for mesh networks. Int. J. Commun. Syst. 2014, 27, 3013–3029. [Google Scholar] [CrossRef]
  20. Muja, M.; Lowe, D.G. Scalable Nearest Neighbor Algorithms for High Dimensional Data. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 2227–2240. [Google Scholar] [CrossRef]
  21. Grisetti, G.; Kümmerle, R.; Stachniss, C.; Frese, U.; Hertzberg, C. Hierarchical optimization on manifolds for online 2D and 3D mapping. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Anchorage, AK, USA, 3–8 May 2010.
  22. Wei, W.; Srivastava, H.M.; Zhang, Y.; Wang, L.; Shen, P.; Zhang, J. A local fractional integral inequality on fractal space analogous to Anderson’s inequality. Abstr. Appl. Anal. 2014, 2014. [Google Scholar] [CrossRef]
  23. Wei, W.; Yang, X.L.; Zhou, B.; Feng, J.; Shen, P.Y. Combined energy minimization for image reconstruction from few views. Math. Probl. Eng. 2012, 2012. [Google Scholar] [CrossRef]
  24. Kuemmerle, R.; Grisetti, G.; Strasdat, H.; Konolige, K.; Burgard, W. G2O: A General Framework for Graph Optimization. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation (ICRA), Shanghai, China, 9–13 May 2011.
  25. Engelhard, N.; Endres, F.; Hess, J.; Sturm, J.; Burgard, W. Real-time 3D visual SLAM with a hand-held RGB-D camera. In Proceedings of the RGB-D Workshop on 3D Perception in Robotics at the European Robotics Forum, Vasteras, Sweden, 8 April 2011.
  26. Kerl, C.; Sturm, J.; Cremers, D. Dense Visual SLAM for RGB-D Cameras. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Tokyo, Japan, 3–7 November 2013.
  27. Steinbruecker, F.; Sturm, J.; Cremers, D. Real-Time Visual Odometry from Dense RGB-D Images. In Proceedings of the 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), Barcelona, Spain, 6–13 November 2011.
  28. Klein, G.; Murray, D. Parallel tracking and mapping for small AR workspaces. In Proceedings of the 6th IEEE and ACM International Symposium on Mixed and Augmented Reality (ISMAR), Nara, Japan, 13–16 November 2007.
  29. Whelan, T.; Kaess, M.; Fallon, M.; Johannsson, H.; Leonard, J.; McDonald, J. Kintinuous: Spatially extended KinectFusion. In Proceedings of the RSS Workshop on RGB-D: Advanced Reasoning with Depth Cameras, Sydney, Australia, 9–10 July 2012.
  30. Agarwal, P.; Tipaldi, G.; Spinello, L.; Stachniss, C.; Burgard, W. Robust map optimization using dynamic covariance scaling. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation (ICRA), Karlsruhe, Germany, 6–10 May 2013.

Share and Cite

MDPI and ACS Style

Zhang, L.; Shen, P.; Zhu, G.; Wei, W.; Song, H. A Fast Robot Identification and Mapping Algorithm Based on Kinect Sensor. Sensors 2015, 15, 19937-19967. https://doi.org/10.3390/s150819937

AMA Style

Zhang L, Shen P, Zhu G, Wei W, Song H. A Fast Robot Identification and Mapping Algorithm Based on Kinect Sensor. Sensors. 2015; 15(8):19937-19967. https://doi.org/10.3390/s150819937

Chicago/Turabian Style

Zhang, Liang, Peiyi Shen, Guangming Zhu, Wei Wei, and Houbing Song. 2015. "A Fast Robot Identification and Mapping Algorithm Based on Kinect Sensor" Sensors 15, no. 8: 19937-19967. https://doi.org/10.3390/s150819937

APA Style

Zhang, L., Shen, P., Zhu, G., Wei, W., & Song, H. (2015). A Fast Robot Identification and Mapping Algorithm Based on Kinect Sensor. Sensors, 15(8), 19937-19967. https://doi.org/10.3390/s150819937

Article Metrics

Back to TopTop