Next Article in Journal
Cooling Cyclic Air of Marine Engine with Water-Fuel Emulsion Combustion by Exhaust Heat Recovery Chiller
Next Article in Special Issue
A Study on the Modeling Method of Cage Slip and Its Effects on the Vibration Response of Rolling-Element Bearing
Previous Article in Journal
Numerical Modeling of the Effects of Pore Characteristics on the Electric Breakdown of Rock for Plasma Pulse Geo Drilling
Previous Article in Special Issue
Intelligent Algorithm for Variable Scale Adaptive Feature Separation of Mechanical Composite Fault Signals
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reinforcement Learning Path Planning Method with Error Estimation

School of Marine Science and Technology, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Energies 2022, 15(1), 247; https://doi.org/10.3390/en15010247
Submission received: 31 October 2021 / Revised: 11 December 2021 / Accepted: 21 December 2021 / Published: 30 December 2021

Abstract

:
Path planning is often considered as an important task in autonomous driving applications. Current planning method only concerns the knowledge of robot kinematics, however, in GPS denied environments, the robot odometry sensor often causes accumulated error. To address this problem, an improved path planning algorithm is proposed based on reinforcement learning method, which also calculates the characteristics of the cumulated error during the planning procedure. The cumulative error path is calculated by the map with convex target processing, while modifying the algorithm reward and punishment parameters based on the error estimation strategy. To verify the proposed approach, simulation experiments exhibited that the algorithm effectively avoid the error drift in path planning.

1. Introduction

Path planning is often utilized to design the optimal trajectory from the start point to the destination. To achieve this goal, various information is required such as the geometric and dynamic information [1], environment map [2], initial state and target state [3]. Furthermore, due to different requirements [4], the collision-free paths are also calculated through multidimensional based on structured scenarios [5].
So far, path planning methods consist of heuristic searching, sampling planning, and model-dependent methods [6]. However, none of the aforementioned methods considers the localization uncertainty issue. In scenarios of GPS denied environments, noisy odometer measurements often lead to unbounded positioning errors. Hence, optimization based on error statistical characteristics is considered, by using the machine learning method. Ref. [7] assumed the background knowledge of the distances is known, and the planning process only needs to update the entries of Q-Table once with four derived attributes. Ref. [8] formulated the behavioral rules of the agent according to the prior environment characteristics. Based on Dubin’s vehicle model, ref. [9] used the dynamic MPNet to generate a suboptimal path. Ref. [10] proposed the integral reinforcement learning path planning solution in continuous state space. Ref. [11] combined soft mobile robot modeling with iterative learning control, and considered adaptive motion planning and trajectory tracking algorithms as its kinematic and dynamic states. Ref. [12] proposed a path planning method called CNPP-Convolutional Network. The method allows multipath planning at one time, that is, a multipath search can be realized through a single prediction iteration of the network. Although the machine learning-based methods were well studied, the method which considers the positioning error is still missing.
Furthermore, traditional planning methods rarely perform fault-tolerant on the control strategy [13]. Therefore, it is quite challenging to plan the path without considering the inevitable errors. Ref. [14] demonstrated a fully convolutional network method to learn human perception, with a real-time RRT* planner to enhance the planner’s ability. By the deep Q-networks calculated from the dense network framework, ref. [15] solved the robot drift issue. Ref. [16] improved the hybrid reciprocating speed obstacle method with dynamic constraints. However, the current approach does not consider the mechanism of initial cumulative error to reduce the positioning uncertainty, which is unacceptable for robots with fine operations.
The proposed approach bridges gaps between the localization process with theoretical qualitative analysis, and the path planning process with the reinforcement learning algorithm, to address cumulative error issues during the navigation procedure. In this paper, an improved path model based on error strategy is established, and the optimal path is calculated through offline learning. The core idea of this paper is that the robot tracking forward will produce the cumulative error drift relative to set path, and the change of set path will also change the drift error. Based on this, the error drift is considered and processed in process of path planning, which can effectively reduce the tracking error.
This paper is organized as follows: Section 2 introduces the statistical characteristics of localization error. Section 3 presents the reinforcement learning-based path planning framework. Section 4 analyzes the effect of the planning algorithm and Section 5 concludes the paper and discusses possible directions for future work.

2. Background

Normally, the robot uses GPS combined with attitude and inertial sensors (gyroscopes and accelerometers) for localization. In the absence of GPS scenarios, the position is calculated by the dead-reckoning approach. It is assumed that the errors in IMU come from measuring white noise only, ignoring Shura period oscillations, Foucault period oscillations, and Earth period oscillations. Assuming the noise is distributed with independent and identical distribution (IID) in polar coordinates, the robot needs to regularly correct errors to achieve high-precision navigation.
Hence, the main challenge is the drift caused by noisy measurement, whereas the localization error accumulates nonlinearly through distance. Based on our original work [17], a statistical formula concerning the noisy measurement is proposed, to obtain the cumulative error growth rate. The solution is as follows:
In polar coordinates, the robot position is localized by measuring the relative angle and distance step by step, which is shown in Figure 1. The corresponding metric is expressed as:
θ n m = θ ¯ n + θ ˜ n ; d n m = d ¯ n + d ˜ n
where n is the time index, d , θ represents relative distance and direction between consecutive frames. The pose measurement ( θ n m , d n m ) is then consisted of ground truth ( θ ¯ n , d ¯ n ) and error ( θ ˜ n , d ˜ n ) with standard deviation δ θ and δ d .
The principle of robot dead reckoning in the Cartesian coordinate system is as follows:
x n m = i = 1 n d i m sin j = 1 i θ j m
y n m = i = 1 n d i m cos j = 1 i θ j m
Since the localization drift based on the noisy measurement is unbounded, it is challenging to estimate the cumulative error; however, the error could be presented by its statistical characteristics.
Assuming the ground truth is known, the trajectory is expressed as:
x n m = x n ¯ + x n ˜ = i = 1 n d i m sin j = 1 i θ j m = i = 1 n d i ¯ + d ˜ i sin j = 1 i θ j ¯ + θ ˜ j = i = 1 n d i ¯ + i = 1 n d ˜ i · sin j = 1 i θ j ¯ cos j = 1 i θ ˜ j + cos j = 1 i θ ¯ j sin j = 1 i θ ˜ j
Rearranging the above formula, the mathematical expression of cumulative error in x direction is acquired:
x n ˜ = i = 1 n d i ¯ sin j = 1 i θ j ¯ cos j = 1 i θ ˜ j 1 + cos j = 1 i θ j ¯ sin j = 1 i θ ˜ j + i = 1 n d ˜ i sin j = 1 i θ j ¯ cos j = 1 i θ ˜ j + cos j = 1 i θ j ¯ sin j = 1 i θ ˜ j
It is observed that the cumulative errors strongly depend on the ground truth. Furthermore, the first- and second-order moments of the cumulative error are expressed as:
E x ˜ | θ ¯ , d ¯ = i = 1 n d i ¯ sin j = 1 i θ j ¯ e i δ θ 2 2 1
var x ˜ | θ ¯ , d ¯ = E x ˜ | θ ¯ , d ¯ E 2 x ˜ | θ ¯ , d ¯ = A + B + C E 2 x ˜ | θ ¯ , d ¯
where:
A = i = 1 n d i 2 ¯ sin 2 j = 1 i θ j ¯ 0.5 e 2 i δ θ 2 + 1.5 2 e i δ θ 2 2 + 0.5 cos 2 j = 1 i θ j ¯ e 2 i δ θ 2 + 1
B = 2 i = 1 n 1 p = 1 + i n d i d p ¯ sin 2 j = 1 i θ j ¯ cos Δ θ ¯ 1 + 0.5 1 + e 2 i δ θ 2 e 0.5 p i δ θ 2 e 0.5 i δ θ 2 e 0.5 i δ θ 2 e 0.5 p i δ θ 2 + sin j = 1 i θ j ¯ sin Δ θ ¯ cos j = 1 i θ j ¯ 1 + 0.5 1 + e 2 i δ θ 2 e 0.5 p i δ θ 2 + 1 e 0.5 i δ θ 2 e 0.5 i δ θ 2 e 0.5 p i δ θ 2 0.5 1 e 2 i δ θ 2 e 0.5 p i δ θ 2 + cos 2 j = 1 i θ j ¯ cos Δ θ ¯ · 0.5 1 e 2 i δ θ 2 e 0.5 p i δ θ 2
C = i = 1 n 0.5 sin 2 j = 1 i θ j ¯ e 2 i δ θ 2 + 1 + 0.5 cos 2 j = 1 i θ j ¯ 1 e 2 i δ θ 2
However, the ground truth is almost nonexistent in practice. To address this problem, the expected values of true moments are evaluated conditional on the noisy relative measurements:
E x n m ˜ = i = 1 n d i m e i δ θ 2 e 0.5 i δ θ 2 sin j = 1 i θ j m
var x n m ˜ = A 1 + B 1 + C 1 E 2 x n m ˜
where:
A 1 = i = 1 n d i m 2 0.5 e 2 i δ θ 2 + 1.5 2 e 0.5 i δ θ 2 0.5 1 + e 2 i δ θ 2 sin 2 j = 1 i θ j m + 0.5 1 e 2 i δ θ 2 cos 2 j = 1 i θ j m + 0.5 1 + e 2 i δ θ 2 0.5 1 + e 2 i δ θ 2 cos 2 j = 1 i θ j m + 0.5 1 + e 2 i δ θ 2 sin 2 j = 1 i θ j m
B 1 = 2 i = 1 n 1 p = 1 + i n d i m d p m 0.5 1 + e 2 i δ θ 2 · sin 2 j = 1 i θ j m + 0.5 1 e 2 i δ θ 2 · cos 2 j = 1 i θ j m · cos Δ θ m e 0.5 p i δ θ 2 + sin j = 1 i θ j m · sin Δ θ m · cos j = 1 i θ j m · e 2 i δ θ 2 e 0.5 p i δ θ 2 + 0.5 1 + e 2 i δ θ 2 · cos 2 j = 1 i θ j m + 0.5 1 e 2 i δ θ 2 · sin 2 j = 1 i θ j m cos Δ θ m e 0.5 p i δ θ 2
C 1 = i = 1 n 0.25 e 2 i δ θ 2 + 1 e 2 i δ θ 2 + 1 · sin 2 j = 1 i θ j m + 1 e 2 i δ θ 2 cos 2 j = 1 i θ j m + 0.25 1 e 2 i δ θ 2 1 + e 2 i δ θ 2 · cos 2 j = 1 i θ j m + 1 e 2 i δ θ 2 · sin 2 j = 1 i θ j m
More details could be found in [18] in the same manner, the complete cumulative errors are therefore calculated. In the course of dead reckoning navigation based on inertial navigation, the existence of error follows the above calculation results. Meanwhile, the propagation of errors accumulates with the growth of time. Through the statistical analysis of the cumulative error, the expected representation of the error can be obtained when the path is determined, which is helpful for us to carry out the route planning task. In the next section, RL-based approach is then utilized for evaluating rewards and punishments to reduce path drift.

3. Planning Strategy

The mathematical expression of path error is obtained, which is the basis of path planning. In the path planning algorithm of Q-learning, the robot explores the environment through discrete actions. This will produce different paths in each iteration, which means different errors, which is the part of the algorithm that needs to be optimized.

3.1. Principle of Q-Learning

Q-Learning is considered as a value-based learning algorithm [19], which makes constant attempts in the environment and adjusts the strategy according to the feedback information obtained from the attempts. By Q-Learning, the robot finally generates a better strategy π to perform optimal actions in different states. Let S be the state set of the agent, A ( s ) is the set of actions that can be performed in the state s S . The discount reinforcement of a given state-action pair ( s , a ) is defined as follows:
V λ ( s , a ) = E t = 0 λ t · r t ( s t , a t ) | s 0 = s , a 0 = a
where λ is the discount factor and r t ( s t , a t ) is the reinforcement given at time t after performing action a t in state s t .
Q-learning assigns a Q-value to each state-action pair ( s , a ) , which is to perform action a in state s. Then, in each subsequent state s, the approximate value of the discount reinforcement obtained by the action a that maximizes Q ( s , a ) is selected.
The formulae used by Q-learning to update the Q-values are:
Q s t , a t Q s t , a t + 1 N r s + λ V ( s ) Q s t , a t
V ( s ) max a A s Q ( s , a )
where N is the cumulative quantity, a , s represent the next selected actions and state.
The Q-Learning-based path planning method is widely used in different platforms and scenarios, such as on smart ships [20], modular platforms [21], and urban autonomous driving scenarios [22]. This paper realizes the optimization of the path by adding the restriction conditions for judging the system error in the Q-Learning method.

3.2. Proposed Strategy

Tradition Q-Learning only selects the local optimum and does not follow the interaction sequence. In path planning, the algorithm has to select actions on each discrete grid [23]. The algorithm first establishes and initializes the action space and state space according to the planning task. Once a boundary or obstacle is encountered, the Q-Table would be updated according to the corresponding reward. After multiple iterations, a relatively fixed Q-Table and action selection sequence would be selected. During the planning procedure, each reward and punishment function contributes the robot to reaching the destination quickly and smoothly. Meanwhile, it is also required to reduce accumulated errors during the planning process.
Once the odometry sensor contains noises, the statistics of entire path are required to be re-calculated step by step. As the conventional planning method cannot work effectively in the beginning, a larger overall deviation may exist in the end. In comparison, the Q-Learning algorithm generates multiple paths through iteration, which is more conducive to effectively reducing the cumulative error.
According to the error estimation models, fewer turning actions result in a smaller cumulative error. Therefore, the reward strategy is adapted to the direction of error reduction. The complete pseudo-code of path planning strategy based on cumulative error statistics is shown in Algorithm 1.
Algorithm 1: Path planning algorithm considering cumulative error statistics
Require: Original map data, reward and punishment function rules, algorithm termination conditions
1:  initial Q ( s , a ) according to map size;
2:  repeat (for each episode):
3:        Initialize s
4:        repeat (for each step of episode):
5:              Choose a from s using ε g r e e d y from Q;
6:              if a and _ a is not consistent then
7:                    Give extra punishment;
8:              else
9:                    Normal rewards and punishments;
10:             end if
11:             Take action a, observe r, s ;
12:       until s is terminal
13:  until the path expectation and variance meet the set value or  e p i s o d e s e t v a l u e
After solving the Q value update, the algorithm will find the corresponding action with a strategy similar to greedy search. Based on considering the cumulative path error, the algorithm updates each iteration to estimate the advantages and disadvantages of action selection by calculating the path error. By avoiding local optimality and other problems, the algorithm can converge to the ideal path.
The planning strategy for error reduction needs the support of a priori map, and the algorithm learning based on Q-Learning is carried out based on determining the error distribution. This will be expanded in detail in the simulation section.

4. Simulation and Discussion

In planning process, paths candidates are crossing through the edges of obstacles, and there are many turns to significantly drop dead-reckoning performance. The robot is drifted from the desired path while the cumulative error increases. As the proposed algorithm may also fall into a suboptimal solution, the mature ε g r e e d y method [24] is thus used to effectively execute the interaction.
For Q-Learning algorithm, the action is discontinuous, and the map used in path planning is also discrete. The size of the initial map used in this paper is 100 m × 100 m, and the minimum resolution is 0.02 m (see Figure 2). In the algorithm learning process, a map that is too fine will cause a huge amount of calculation and a long time-consuming. And for robot motion control, the control strategy does not require very precise route guidance. Therefore, the initial map needs to be smoothed and zoomed to match the algorithm, especially for path planning of nonholonomically constrained robots, e.g., Autonomous Underwater Vehicles.
Given a configuration space (C-space) Ω , this paper assume that the complete set of constraints is described in a cost function τ [25]. The cost function depends only on the configuration x in isotropic case:
τ : Ω R + , x τ ( x ) , τ ( x ) > 0
The lower bound of the radius R of curvature of the obstacle boundary be expressed as:
R m i n inf Ω τ sup Ω τ
Smoothing is performed by including kinematics and curvature constraints on map obstacles. The final grid map is set to the resolution of 5 m/cell, as shown in Figure 3. The initial position is set at ( 0.5 , 0.5 ) , and the initial pose is defined as facing to the right, as shown in the red dot (lower left corner) in the figure. The target position is set at ( 97.5 , 97.5 ) , as shown in the red circle (upper right corner). The learning of the algorithm is mainly related to the a priori map and the starting point. The changing endpoint is always suitable for the cumulative error map to complete the learning and obtain the ideal path.
In general, the robot can accurately reach the end-point through the tracking process, with the help of high-precision GPS. In GPS-denied scenarios, considering the cumulative error originated from the noisy odometry sensors, with the distribution of N ( 0 , 1 ) and N ( 0 , 0.1 ) in both range and angular, respectively. As a result of the traditional reinforcement learning plan, the path can reach the end perfectly. At the moment, the reward/penalty is set to 1 / 1 , and ε is set to 0.9 . However, considering the influence of the above-mentioned sensor noise, the actual path of the robot will deviate from original path and end-point to a large extent, as shown in Figure 4. The solid red path is originally planned and the red dotted path is one affected by the accumulated error. The cumulative error in paths with many turns is relatively large. In a scenario where the robot only relies on odometry, it is easy to deviate from the predetermined path.
Meanwhile, the proposed algorithm with reset parameters of rewards and punishments is adopted to obtain the ideal planning results. The specific parameter settings used in this paper are shown in Table 1, which does not imply optimal parameters. This paper has passed multiple tests and achieved a good simulation effect when ε = 0.95 , and it is easier to get a convergence solution. Reinforcement learning algorithms based on probability selection have inconsistencies in results. This paper selects typical results as shown in Figure 5. The number of turns on planned path reduced, in which the cumulative error is significantly eliminated. The cumulative error is inevitable, whereas the localization error of the planning algorithm could be calculated effectively.
In the scene of constant robot speed, the cumulative error increases with time. Moreover, the cumulative error seems to increase significantly as the path turns increase. Compared with the traditional method, the proposed algorithm has a significant effect in reducing the cumulative error in the x and y directions (as shown in Figure 6), and the end-point of the tracking process is closer to the target point at the Euclidean distance. It is also verified from the application scenario that in the course of dead reckoning based on odometer, the positioning error caused by gyroscope error is greater than that caused by accelerometer error [26]. The reduction of accumulated error verifies the effectiveness of the algorithm.
In the process of multiple iterations of planning, the expected value of each path in the x and y directions will fluctuate greatly, especially at the beginning of the algorithm. This is caused by the complicated path caused by probability in the early iteration process. After the algorithm converges, the expected values in the x and y directions remain around 4750 and 4750, as shown in Figure 7, which verifies the reliability of the algorithm.
As the action selection is unknown, reinforcement learning algorithms will converge to different results, or even fail to converge to the best solution. In this paper, a heat map is made to visualize the iterative process of the algorithm, as shown in Figure 8. The agent reaches the area outside the optimal path many times, and finally, it converges to the vicinity of the optimal solution.

5. Conclusions

In GPS-denied scenarios, a path planning algorithm based on error estimation is proposed. Firstly, the cumulative error of the sensor noise is analyzed and the expectation and variance are described. Secondly, the reinforcement learning algorithm is utilized by considering the cumulative error during the planning process. Finally, the learning strategy is conducted to acquire an optimal path with noisy relative measurements. Simulation verifies that the proposed algorithm can effectively reduce the localization error of the robot in complex environments. The convergence results can be obtained through multiple learning processes, and the error expectation conforms to the theory, which shows the effectiveness and stability of the algorithm.

Author Contributions

F.Z. provided the initial motivation and ideas. C.W. conducted manuscript writing and data analysis. C.C. and D.Y. realized the design of the simulation experiment. G.P. provided financial support. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Natural Science Foundation of China (52171322), the National Key Research and Development Program (2020YFB1313200), and the Fundamental Research Funds for the Central Universities (D5000210944).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors appreciated the participation of all the subjects in the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
GPSGlobal Positioning System
IIDIndependent Identically Distribution
R.LReinforcement Learning

References

  1. Bidot, J.; Karlsson, L.; Lagriffoul, F.; Saffiotti, A. Geometric backtracking for combined task and motion planning in robotic systems. Artif. Intell. 2013, 247, 229–265. [Google Scholar] [CrossRef]
  2. Peng, Y.; Green, P.N. Environment mapping, map constructing, and path planning for underwater navigation of a low-cost μAUV in a cluttered nuclear storage pond. IAES Int. J. Robot. Autom. 2019, 8, 277–292. [Google Scholar] [CrossRef]
  3. Choset, H.; Lynch, K.; Hutchinson, S.; Kantor, G.; Burgard, W.; Kavraki, L.; Thrun, S. Principles of Robot Motion: Theory, Algorithms, and Implementation. 2005. Available online: https://ieeexplore.ieee.org/servlet/opac?bknumber=6267238 (accessed on 9 November 2021).
  4. Ibraheem, I.K.; Hassan, F. Path Planning of an Autonomous Mobile Robot in a Dynamic Environment using Modified Bat Swarm Optimization. arXiv 2018, arXiv:1807.05352. [Google Scholar]
  5. Zeng, J.; Qin, L.; Hu, Y.; Yin, Q.; Hu, C. Integrating a Path Planner and an Adaptive Motion Controller for Navigation in Dynamic Environments. Appl. Sci. 2019, 9, 1384. [Google Scholar] [CrossRef] [Green Version]
  6. Yilmaz, N.K.; Evangelinos, C.; Lermusiaux, P.; Patrikalakis, N.M. Path Planning of Autonomous Underwater Vehicles for Adaptive Sampling Using Mixed Integer Linear Programming. IEEE J. Ocean. Eng. 2008, 33, 522–537. [Google Scholar] [CrossRef] [Green Version]
  7. Konar, A.; Chakraborty, I.; Singh, S.; Jain, L.; Nagar, A. A Deterministic Improved Q-Learning for Path Planning of a Mobile Robot. Syst. Man Cybern. Syst. IEEE Trans. 2013, 43, 1141–1153. [Google Scholar] [CrossRef] [Green Version]
  8. Kaiqiang, T.; Fu, H.; Jiangt, H.; Liu, C.; Wang, L. Reinforcement Learning for Robots Path Planning with Rule-based Shallow-trial. In Proceedings of the 2019 IEEE 16th International Conference on Networking, Sensing and Control (ICNSC), Banff, AB, Canada, 9–11 May 2019; pp. 340–345. [Google Scholar] [CrossRef]
  9. Johnson, J.J.; Li, L.; Liu, F.; Qureshi, A.H.; Yip, M.C. Dynamically Constrained Motion Planning Networks for Non-Holonomic Robots. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2021. [Google Scholar] [CrossRef]
  10. He, C.; Wan, Y.; Gu, Y.; Lewis, F. Integral reinforcement learning-based approximate minimum time-energy path planning in an unknown environment. Int. J. Robust Nonlinear Control 2020, 31, 1905–1922. [Google Scholar] [CrossRef]
  11. Luo, M.; Wan, Z.; Sun, Y.; Skorina, E.H.; Tao, W.; Chen, F.; Gopalka, L.; Yang, H.; Onal, C.D. Motion Planning and Iterative Learning Control of a Modular Soft Robotic Snake. Front. Robot. AI 2020, 7, 191. [Google Scholar] [CrossRef] [PubMed]
  12. Kulvicius, T.; Herzog, S.; Lüddecke, T.; Tamosiunaite, M.; Wörgötter, F. One-Shot Multi-Path Planning Using Fully Convolutional Networks in a Comparison to Other Algorithms. Front. Neurorobotics 2021, 14, 600984. [Google Scholar] [CrossRef] [PubMed]
  13. Rolland, L. Path Planning Kinematics Simulation of CNC Machine Tools Based on Parallel Manipulators. Mech. Mach. Sci. 2015, 29, 147–192. [Google Scholar] [CrossRef] [Green Version]
  14. Pérez Higueras, N.; Caballero, F.; Merino, L. Learning Human-Aware Path Planning with Fully Convolutional Networks. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 5897–5902. [Google Scholar] [CrossRef] [Green Version]
  15. Lv, L.; Zhang, S.; Ding, D.; Wang, Y. Path Planning via an Improved DQN-based Learning Policy. IEEE Access 2019, 7, 67319–67330. [Google Scholar] [CrossRef]
  16. Sainte Catherine, M.; Lucet, E. A modified Hybrid Reciprocal Velocity Obstacles approach for multirobot motion planning without communication. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October–24 January 2021; pp. 5708–5714. [Google Scholar] [CrossRef]
  17. Zhang, F.; Simon, C.; Chen, G.; Buckl, C.; Knoll, A. Cumulative error estimation from noisy relative measurements. In Proceedings of the 16th International IEEE Conference on Intelligent Transportation Systems (ITSC 2013), The Hague, Netherlands, 6–9 October 2013; pp. 1422–1429. [Google Scholar] [CrossRef] [Green Version]
  18. Zhang, F.; Knoll, A. Systematic Error Modeling and Bias Estimation. Sensors 2016, 16, 729. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Cjc, H.W. Learning with Delayed Rewards. Ph.D. Thesis, University of Cambridge, Cambridge, MA, USA, 1989; pp. 233–235. [Google Scholar] [CrossRef]
  20. Chen, C.; Chen, X.Q.; Ma, F.; Zeng, X.J.; Wang, J. A knowledge-free path planning approach for smart ships based on reinforcement learning. Ocean Eng. 2019, 189, 106299. [Google Scholar] [CrossRef]
  21. Haghzad Klidbary, S.; Bagheri Shouraki, S.; Sheikhpour, S. Path planning of modular robots on various terrains using Q-learning versus optimization algorithms. Intell. Serv. Robot. 2017, 10, 121–136. [Google Scholar] [CrossRef]
  22. Zhang, X.; Li, H.; Peng, J.; Liu, W. A Cooperative Q-Learning Path Planning Algorithm for Origin-Destination Pairs in Urban Road Networks. Math. Probl. Eng. 2015, 2015, 146070. [Google Scholar] [CrossRef] [Green Version]
  23. Su, M.C.; Huang, D.Y.; Chou, C.H.; Hsieh, C.C. A reinforcement-learning approach to robot navigation. In Proceedings of the IEEE International Conference on Networking, Sensing and Control, Taipei, Taiwan, 21–23 March 2004; Volume 1, pp. 665–669. [Google Scholar] [CrossRef]
  24. Sutton, R.; Barto, A. Reinforcement Learning: An Introduction. IEEE Trans. Neural Networks 1998, 9, 1054. [Google Scholar] [CrossRef]
  25. Petres, C.; Pailhas, Y.; Patron, P.; Petillot, Y.; Evans, J.; Lane, D. Path Planning for Autonomous Underwater Vehicles. Robot. IEEE Trans. 2007, 23, 331–341. [Google Scholar] [CrossRef]
  26. Poddar, S.; Kumar, V.; Kumar, A. A Comprehensive Overview of Inertial Sensor Calibration Techniques. J. Dyn. Syst. Meas. Control 2017, 139, 011006. [Google Scholar] [CrossRef]
Figure 1. Expression of relative measurement and position in polar coordinate.
Figure 1. Expression of relative measurement and position in polar coordinate.
Energies 15 00247 g001
Figure 2. Initial global map.
Figure 2. Initial global map.
Energies 15 00247 g002
Figure 3. Rasterized map after convex target processing.
Figure 3. Rasterized map after convex target processing.
Energies 15 00247 g003
Figure 4. Results of traditional path planning methods and paths with errors.
Figure 4. Results of traditional path planning methods and paths with errors.
Energies 15 00247 g004
Figure 5. Comparison of planning effects of two algorithms.
Figure 5. Comparison of planning effects of two algorithms.
Energies 15 00247 g005
Figure 6. Cumulative error of two algorithms in x and y directions.
Figure 6. Cumulative error of two algorithms in x and y directions.
Energies 15 00247 g006
Figure 7. Statistics of expected values in the x-axis and y-axis directions.
Figure 7. Statistics of expected values in the x-axis and y-axis directions.
Energies 15 00247 g007
Figure 8. Heat map of agent selection location during reinforcement learning.
Figure 8. Heat map of agent selection location during reinforcement learning.
Energies 15 00247 g008
Table 1. Algorithm parameter settings.
Table 1. Algorithm parameter settings.
StatesReward/PenaltyValue
Reach the boundary or obstaclepenalty−1
Reach the end pointreward100
Drive straightpenalty−0.0001
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhang, F.; Wang, C.; Cheng, C.; Yang, D.; Pan, G. Reinforcement Learning Path Planning Method with Error Estimation. Energies 2022, 15, 247. https://doi.org/10.3390/en15010247

AMA Style

Zhang F, Wang C, Cheng C, Yang D, Pan G. Reinforcement Learning Path Planning Method with Error Estimation. Energies. 2022; 15(1):247. https://doi.org/10.3390/en15010247

Chicago/Turabian Style

Zhang, Feihu, Can Wang, Chensheng Cheng, Dianyu Yang, and Guang Pan. 2022. "Reinforcement Learning Path Planning Method with Error Estimation" Energies 15, no. 1: 247. https://doi.org/10.3390/en15010247

APA Style

Zhang, F., Wang, C., Cheng, C., Yang, D., & Pan, G. (2022). Reinforcement Learning Path Planning Method with Error Estimation. Energies, 15(1), 247. https://doi.org/10.3390/en15010247

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop