首页 | 本学科首页   官方微博 | 高级检索  
     


On the Convergence of Temporal-Difference Learning with Linear Function Approximation
Authors:Vladislav Tadi?
Affiliation:(1) Department of Electrical and Electronic Engineering, The University of Melbourne, Parkville, Victoria, 3010, Australia
Abstract:The asymptotic properties of temporal-difference learning algorithms with linear function approximation are analyzed in this paper. The analysis is carried out in the context of the approximation of a discounted cost-to-go function associated with an uncontrolled Markov chain with an uncountable finite-dimensional state-space. Under mild conditions, the almost sure convergence of temporal-difference learning algorithms with linear function approximation is established and an upper bound for their asymptotic approximation error is determined. The obtained results are a generalization and extension of the existing results related to the asymptotic behavior of temporal-difference learning. Moreover, they cover cases to which the existing results cannot be applied, while the adopted assumptions seem to be the weakest possible under which the almost sure convergence of temporal-difference learning algorithms is still possible to be demonstrated.
Keywords:temporal-difference learning  reinforcement learning  neuro-dynamic programming  almost sure convergence  Markov chains  positive Harris recurrence
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号