首页 | 本学科首页   官方微博 | 高级检索  
     

基于区域扩张策略的势场强化学习算法路径规划研究
引用本文:阳杰,张凯.基于区域扩张策略的势场强化学习算法路径规划研究[J].微处理机,2021(1):47-51.
作者姓名:阳杰  张凯
作者单位:西安工程大学电子信息学院
摘    要:未知连续环境状态下的Q学习路径规划算法在执行对环境的试错时收敛速度慢,容易陷入局部,不利于对真实未知环境的探索,为解决此问题,针对Q学习路径规划问题提出一种基于Metropolis准则的区域扩张策略的势场强化学习算法。算法为环境提供势场先验知识初始化状态信息,消除初始时刻的盲目性,提高学习效率,同时引入基于Metropolis准则的区域扩张陷阱区域剔除探索,剔除陷阱障碍物环境的凹形区域。通过MATLAB对多种环境的仿真实验,验证了算法有效性。

关 键 词:Q学习  METROPOLIS准则  势场强化学习  探索区域扩张  MATLAB仿真

Study on Path Planning of Potential Field Reinforcement Learning Algorithm Based on Regional Expansion Strategy
YANG Jie,ZHANG Kai.Study on Path Planning of Potential Field Reinforcement Learning Algorithm Based on Regional Expansion Strategy[J].Microprocessors,2021(1):47-51.
Authors:YANG Jie  ZHANG Kai
Affiliation:(School of Electronics and Information,Xi'an Polytechnic University,Xi'an 710600,China)
Abstract:Q-learning path planning algorithm in unknown continuous environment is slow in convergence and easy to fall into local area when executing trial and error on the environment,which is not conducive to the exploration of real unknown environment.To solve the problem,a potential field reinforcement learning algorithm based on Metropolis criterion is proposed for Q-learning path planning.The algorithm provides the environment with the prior knowledge of potential field to initialize the state information,eliminates the blindness at the initial moment and improves the learning efficiency.At the same time,it introduces the exploration of eliminating the trap area of regional expansion based on Metropolis criterion to eliminate the concave area of the trap obstacle environment.The effectiveness of the algorithm is verified by MATLAB simulation experiments in various environments.
Keywords:Q-learning  Metropolis criterion  Potential field reinforcement learning  Explore regional expansion  MATLAB simulation
本文献已被 维普 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号