• Electronics Optics & Control
  • Vol. 26, Issue 9, 29 (2019)
DUAN Jianmin and CHEN Qianglong
Author Affiliations
  • [in Chinese]
  • show less
    DOI: 10.3969/j.issn.1671-637x.2019.09.007 Cite this Article
    DUAN Jianmin, CHEN Qianglong. Prior Knowledge Based Q-Learning Path Planning Algorithm[J]. Electronics Optics & Control, 2019, 26(9): 29 Copy Citation Text show less

    Abstract

    The standard Q-Learning algorithm based on Markov decision process in reinforcement learning can obtain an optimum path, but the method has the shortcomings of slow convergence rate and low planning efficiency, and thus can not be directly applied to the real environment. This paper proposes a Q-Learning path planning algorithm for mobile robots based on the potential energy field knowledge.By introducing the potential energy value into the environment as the search heuristic information to initialize the Q value, the rapid convergence of the mobile robot can be guided in the early stage of learning, and the blindness of the traditional reinforcement learning process is avoided, which makes it suitable for direct learning in a real environment. The simulation result shows that: Compared with existing algorithms, the proposed algorithm not only improves the convergence speed, but also shortens the learning time, which can make the mobile robot find a better collision-free path quickly.
    DUAN Jianmin, CHEN Qianglong. Prior Knowledge Based Q-Learning Path Planning Algorithm[J]. Electronics Optics & Control, 2019, 26(9): 29
    Download Citation