透過您的圖書館登入
IP:216.73.216.60
  • 期刊

Robot Path Planning Method Based on Reinforcement Learning in Unknown Environment

摘要


Aiming at the problem that traditional path planning methods depend on environment model, the path planning of mobile robots in unknown environment is studied by using Sarsa and Q-learning algorithms based on reinforcement learning. Firstly, the simulation environment is established by using grid map, and the algorithm and Q-value table are designed. Secondly, the reward value is obtained by the interaction between the robot and the environment, and the action strategy set is changed by modifying the Q-value table iteratively. Finally, the action of the robot tends to the optimal action set. The experimental results show that under the same iteration steps, the Sarsa algorithm tends to be conservative and converges slowly, while the Q-learning algorithm is easier to converge, can quickly plan the path, and has better exploration ability in unknown environment.

參考文獻


Koubaa A, Bennaceur H, Chaari I, et al. Introduction to Mobile Robot Path Planning [M]// Robot Path Planning and Cooperation: Foundations, Algorithms and Experimentations. 2018.
Koren Y, Borenstein J. Potential field methods and their inherent limitations for mobile robot navigation[C]// IEEE International Conference on Robotics and Automation. IEEE, 1991:1398-1404 vol.2.
Hee Rak Beom, Kyung Suck Cho. A sensor-based navigation for a mobile robot using fuzzy logic and reinforcement learning [J]. Systems Man & Cybernetics IEEE Transactions on, 1995, 25(3):464-477.
Castillo O, Leonardo Trujillo, Patricia Melin. Multiple Objective Genetic Algorithms for Path-planning Optimization in Autonomous Mobile Robots [J]. Soft Computing, 2007, 11(3):269-279.
Clerc M, Kennedy, J. The particle swarm - explosion, stability, and convergence in a multidimensional complex space [J]. IEEE Transactions on Evolutionary Computation, 2002, 6(1):58-73.

延伸閱讀