首页 | 本学科首页   官方微博 | 高级检索  
     

非线性零和微分对策的事件触发自适应动态规划算法
引用本文:崔黎黎,张勇,张欣.非线性零和微分对策的事件触发自适应动态规划算法[J].控制理论与应用,2018,35(5):610-618.
作者姓名:崔黎黎  张勇  张欣
作者单位:沈阳师范大学科信软件学院;中国石油大学(华东)信息与控制工程学院
基金项目:国家自然科学基金项目(61703289), 山东省自然科学基金项目(BX2015DX009), 辽宁省高等学校基本科研项目专项资金(LQN201720, LQN2017 02), 沈阳师范大学科技项目(L201510)资助.
摘    要:针对一类非线性零和微分对策问题,本文提出了一种事件触发自适应动态规划(event-triggered adaptive dynamic programming,ET--ADP)算法在线求解其鞍点.首先,提出一个新的自适应事件触发条件.然后,利用一个输入为采样数据的神经网络(评价网络)近似最优值函数,并设计了新型的神经网络权值更新律使得值函数、控制策略及扰动策略仅在事件触发时刻同步更新.进一步地,利用Lyapunov稳定性理论证明了所提出的算法能够在线获得非线性零和微分对策的鞍点且不会引起Zeno行为.所提出的ET--ADP算法仅在事件触发条件满足时才更新值函数、控制策略和扰动策略,因而可有效减少计算量和降低网络负荷.最后,两个仿真例子验证了所提出的ET--ADP算法的有效性.

关 键 词:自适应动态规划    非线性零和微分对策    事件触发    神经网络    最优控制
收稿时间:2017/9/15 0:00:00
修稿时间:2017/12/27 0:00:00

Event-triggered adaptive dynamic programming algorithm for the nonlinear zero-sum differential games
CUI Li-li,ZHANG Yong and ZHANG Xin.Event-triggered adaptive dynamic programming algorithm for the nonlinear zero-sum differential games[J].Control Theory & Applications,2018,35(5):610-618.
Authors:CUI Li-li  ZHANG Yong and ZHANG Xin
Affiliation:Shenyang Normal University,Shenyang Normal University,China University of Petroleum
Abstract:In this paper, an event-triggered adaptive dynamic programming algorithm (ET--ADP) is proposed to solve the saddle point of a class of nonlinear zero-sum differential games. Firstly, a new adaptive event-triggered condition is proposed. Then, a neural network (critic network) with the sampled state as its input is utilized to approximate the optimal value function. The new neural network weights updating law is designed to enable the value function, the control strategy and the disturbance strategy to be updated synchronously only at the event-triggered time. Further, the Lyapunov stability theory is used to prove that the proposed algorithm can obtain the saddle point of nonlinear zero-sum differential games online and avoid the occurrence of Zeno behavior. In the proposed ET--ADP algorithm, the value function, the control strategy and the disturbance strategy are updated only when the event-triggered condition is satisfied, as a result of which the computational burden is reduced and the network burden is eased effectively. Finally, two simulation examples validate the effectiveness of the proposed ET--ADP algorithm.
Keywords:adaptive dynamic programming  nonlinear zero-sum differential games  event-triggered  optimal control
本文献已被 CNKI 等数据库收录!
点击此处可从《控制理论与应用》浏览原始摘要信息
点击此处可从《控制理论与应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号