首页 | 本学科首页   官方微博 | 高级检索  
     


An efficient initialization approach of Q-learning for mobile robots
Authors:Yong Song  Yi-bin Li  Cai-hong Li  Gui-fang Zhang
Affiliation:1.School of Control Science and Engineering,Shandong University,Jinan,China;2.School of Mechanical, Electrical & Information Engineering,Shandong University at Weihai,Weihai,China;3.School of Control Science and Engineering,Shandong University,Jinan,China;4.School of Computer Science and Technology,Shandong University of Technology,Zibo,China
Abstract:This article demonstrates that Q-learning can be accelerated by appropriately specifying initial Q-values using dynamic wave expansion neural network. In our method, the neural network has the same topography as robot work space. Each neuron corresponds to a certain discrete state. Every neuron of the network will reach an equilibrium state according to the initial environment information. The activity of the special neuron denotes the maximum cumulative reward by following the optimal policy from the corresponding state when the network is stable. Then the initial Q-values are defined as the immediate reward plus the maximum cumulative reward by following the optimal policy beginning at the succeeding state. In this way, we create a mapping between the known environment information and the initial values of Q-table based on neural network. The prior knowledge can be incorporated into the learning system, and give robots a better learning foundation. Results of experiments in a grid world problem show that neural network-based Q-learning enables a robot to acquire an optimal policy with better learning performance compared to conventional Q-learning and potential field-based Qlearning.
Keywords:
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号