Modified Q-learning with distance metric and virtual target on path planning of mobile robot

Path planning is an essential element in mobile robot navigation. One of the popular path planners is Q-learning – a type of reinforcement learning that learns with little or no prior knowledge of the environment. Despite the successful implementation of Q-learning reported in numerous studies,...

全面介绍

Saved in:
书目详细资料
Main Authors: Ee, Soong Low, Ong, Pauline, Cheng, Yee Low, Omar, Rosli
格式: Article
语言:English
出版: Elsevier 2020
主题:
在线阅读:http://eprints.uthm.edu.my/7147/1/J14146_d4792edcf485886c4b01ef6a4fbc4dca.pdf
http://eprints.uthm.edu.my/7147/
https://doi.org/10.1016/j.eswa.2022.117191
标签: 添加标签
没有标签, 成为第一个标记此记录!
实物特征
总结:Path planning is an essential element in mobile robot navigation. One of the popular path planners is Q-learning – a type of reinforcement learning that learns with little or no prior knowledge of the environment. Despite the successful implementation of Q-learning reported in numerous studies, its slow convergence associated with the curse of dimensionality may limit the performance in practice. To solve this problem, an Improved Q-learning (IQL) with three modifications is introduced in this study. First, a distance metric is added to Q-learning to guide the agent moves towards the target. Second, the Q function of Q-learning is modified to overcome dead-ends more effectively. Lastly, the virtual target concept is introduced in Q-learning to bypass dead-ends. Experi�mental results across twenty types of navigation maps show that the proposed strategies accelerate the learning speed of IQL in comparison with the Q-learning. Besides, performance comparison with seven well-known path planners indicates its efficiency in terms of the path smoothness, time taken, shortest distance and total distance used.