Asynchronous reinforcement learning algorithms for solving discrete space path planning problems
作者:Xingyu Zhao, Shifei Ding, Yuexuan An, Weikuan Jia
摘要
Reinforcement learning has great potential in solving practical problems, but when combining it with neural networks to solve small scale discrete space problems, it may easily trap in a local minimum value. Traditional reinforcement learning utilizes continuous updating of a single agent to learn policies, which easily leads to a slow convergence speed. In order to solve the above problems, we combine asynchronous methods with existing tabular reinforcement learning algorithms, propose a parallel architecture to solve the discrete space path planning problem, and present some new variants of asynchronous reinforcement learning algorithms. We apply these algorithms on the standard reinforcement learning environment problems, and the experimental results show that these methods can solve discrete space path planning problems efficiently. One of these algorithms, Asynchronous Phased Dyna-Q, which surpasses existing asynchronous reinforcement learning algorithms, can well balance exploration and exploitation.
论文关键词:Reinforcement learning, Path planning, Dyna architecture, Asynchronous methods, Discrete space
论文评审过程:
论文官网地址:https://doi.org/10.1007/s10489-018-1241-z