共查询到19条相似文献,搜索用时 162 毫秒
1.
针对一类线性时不变系统, 提出了具有反馈信息的PD-型(Proportional-derivative-type)迭代学习控制律, 利用卷积的推广的Young不等式, 分析了控制律在Lebesgue-p范数意义下的单调收敛性. 分析表明, 收敛性不但决定于系统的输入输出矩阵和控制律的微分学习增益, 而且依赖于系统的状态矩阵和控制律的比例学习增益; 进一步, 当适当选取反馈增益时, 反馈信息可加快典型的PD-型迭代学习控制律的单调收敛性. 数值仿真验证了理论分析的正确性和控制律的有效性. 相似文献
2.
3.
离散非线性系统开闭环P型迭代学习控制律及其收敛性 总被引:9,自引:3,他引:9
本文在讨论了一般开环与闭环迭代学习控制的不足后,针对一类离散非线性系统,提出了新的开闭环PG型迭代学习控制律,给出了它的收敛性证明,仿真结果表明:开闭环P型迭代律优于单纯的开环或产才环P型迭代 律。 相似文献
4.
针对不确定性系统提出一种非因果鲁棒学习控制方法。该学习控制律的非因果学习部分通过标称系统的优化指标得到,鲁棒部分通过设计鲁棒加权来实现。首先,不考虑鲁棒部分的具体形式,推导出标称系统描述的学习控制律的鲁棒收敛性条件;然后,设计与系统不确定性相关的鲁棒加权,由鲁棒收敛性条件得到鲁棒加权的设计原则;最后,通过仿真实验验证了所提出方法的有效性,并分析了不同形式不确定性系统鲁棒设计的保守性。 相似文献
5.
6.
针对一类线性广义系统,研究其P型迭代学习控制在离散频域中的收敛性态。在离散频域中,对广义系统进行奇异值分解后,利用傅里叶级数系数的性质和离散的Parseval能量等式,推演了一阶P型迭代学习控制律跟踪误差的离散能量频谱的递归关系和特性,获得了学习控制律收敛的充分条件;讨论了二阶P型迭代学习控制律的收敛条件。仿真实验验证了理论的正确性和学习律的有效性。 相似文献
7.
8.
9.
传统的迭代学习控制机理中, 积分补偿是典型的策略之一, 但其跟踪效用并不明确. 本文针对连续线性时
不变系统, 对传统的PD–型迭代学习控制律嵌入积分补偿, 利用分部积分法和推广的卷积Young不等式, 在Lebesgue-
p范数意义下, 理论分析一阶和二阶PID–型迭代学习控制律的收敛性态. 结果表明, 当比例、积分和导数学习增益满
足适当条件时, 一阶PID–型迭代学习控制律是单调收敛的, 二阶PID–型迭代学习控制律是双迭代单调收敛的. 数值
仿真验证了积分补偿可有效地提高系统的跟踪性能. 相似文献
10.
初态学习下的迭代学习控制 总被引:3,自引:1,他引:2
提出一种新的初态学习律,以放宽常规迭代学习控制方法的初始定位条件.它允许一定的定位误差,在迭代中不需要定位在某一具体位置上,使得学习控制系统具有鲁棒收敛性.针对二阶LTI系统,给出了输入学习律及初态学习律的收敛性充分条件.依据收敛性条件,学习增益的选取需系统矩阵的估计值,但在一定建模误差下,仍能保证算法的收敛性.所提出的初态学习律本身及其收敛性条件均与输入矩阵无关. 相似文献
11.
An iterative learning control scheme is presented for a class of nonlinear dynamic systems which includes holonomic systems as its subset. The control scheme is composed of two types of control methodology: a linear feedback mechanism and a feedforward learning strategy. At each iteration, the linear feedback provides stability of the system and keeps its state errors within uniform bounds. The iterative learning rule, on the other hand, tracks the entire span of a reference input over a sequence of iterations. The proposed learning control scheme takes into account the dominant system dynamics in its update algorithm in the form of scaled feedback errors. In contrast to many other learning control techniques, the proposed learning algorithm neither uses derivative terms of feedback errors nor assumes external input perturbations as a prerequisite. The convergence proof of the proposed learning scheme is given under minor conditions on the system parameters. 相似文献
12.
13.
当数据规模庞大时,深度学习模型会遇到权重调整耗时,容易陷入局部最优解的问题.为了解决这些问题,宽度学习系统应运而生,宽度学习系统不仅结构简单、训练速度快、准确率高,而且还具有增量学习的优势.介绍了宽度学习系统的产生背景和发展历程,阐述了宽度学习系统的基础理论与实现方法,对比了它与深度网络的异同;介绍了宽度学习系统在图像分类、数值回归、脑电信号处理等应用中的改进算法,分析了这些算法的优势和不足.最后总结了现有宽度学习算法存在的缺陷,并对未来研究方向进行了展望. 相似文献
14.
15.
Beatriz Pérez-Sánchez Oscar Fontenla-Romero Bertha Guijarro-Berdiñas David Martínez-Rego 《Expert systems with applications》2013,40(18):7294-7304
Many real scenarios in machine learning are of dynamic nature. Learning in these types of environments represents an important challenge for learning systems. In this context, the model used for learning should work in real time and have the ability to act and react by itself, adjusting its controlling parameters, even its structures, depending on the requirements of the process. In a previous work, the authors presented an online learning algorithm for two-layer feedforward neural networks that includes a factor that weights the errors committed in each of the samples. This method is effective in dynamic environments as well as in stationary contexts. As regards this method’s incremental feature, we raise the possibility that the network topology is adapted according to the learning needs. In this paper, we demonstrate and justify the suitability of the online learning algorithm to work with adaptive structures without significantly degrading its performance. The theoretical basis for the method is given and its performance is illustrated by means of its application to different system identification problems. The results confirm that the proposed method is able to incorporate units to its hidden layer, during the learning process, without high performance degradation. 相似文献
16.
基于协同最小二乘支持向量机的Q学习 总被引:5,自引:0,他引:5
针对强化学习系统收敛速度慢的问题, 提出一种适用于连续状态、离散动作空间的基于协同最小二乘支持向量机的Q学习. 该Q学习系统由一个最小二乘支持向量回归机(Least squares support vector regression machine, LS-SVRM)和一个最小二乘支持向量分类机(Least squares support vector classification machine, LS-SVCM)构成. LS-SVRM用于逼近状态--动作对到值函数的映射, LS-SVCM则用于逼近连续状态空间到离散动作空间的映射, 并为LS-SVRM提供实时、动态的知识或建议(建议动作值)以促进值函数的学习. 小车爬山最短时间控制仿真结果表明, 与基于单一LS-SVRM的Q学习系统相比, 该方法加快了系统的学习收敛速度, 具有较好的学习性能. 相似文献
17.
R. Andrew Russell 《Autonomous Robots》2004,16(1):81-93
The aim of the project described in this paper was to investigate robot learning at a most fundamental level. The project focused on the transition between organisms with innate behaviors and organisms that have the most rudimentary capability of learning through their personal interaction with their environment. It was assumed that the innate behaviors gave basic survival competence but no learning ability. By observing the interaction between their innate behaviors and the organism's environment it was reasoned that the organism should be able to learn how to modify its actions in a way that improves its performance. If a learning system is given more information than it requires then, when it is successful, it is difficult to say which pieces of information contribute to the success. For this reason the information available to the learning system was kept to an absolute minimum. In order to provide a practical test of the learning scheme developed in this project, the robot environment EDEN was constructed. Within EDEN a robot's actions influence its internal energy reserves. The environment incorporates sources of energy, and it also involves situations that use additional energy or reduce energy consumption. A successful learning scheme was developed purely based on the recorded history of the robot's interactions with its environment and the knowledge that the robot's innate behavior was reactive. This learning scheme allowed the robot to improve its energy management by exhibiting classical conditioning and a restricted form of operant conditioning. 相似文献
18.
Teng Liu Bin Tian Yunfeng Ai Li Li Dongpu Cao Fei-Yue Wang 《IEEE/CAA Journal of Automatica Sinica》2018,5(4):827-835
In this paper, a new machine learning framework is developed for complex system control, called parallel reinforcement learning. To overcome data deficiency of current data-driven algorithms, a parallel system is built to improve complex learning system by self-guidance. Based on the Markov chain (MC) theory, we combine the transfer learning, predictive learning, deep learning and reinforcement learning to tackle the data and action processes and to express the knowledge. Parallel reinforcement learning framework is formulated and several case studies for real-world problems are finally introduced. 相似文献
19.
The learning transient and tracking accuracy of phase lead compensation iterative learning control are determined by its three parameters: learning gain, system learnable bandwidth and lead step. Because of the model inaccuracy, the learnable bandwidth is often chosen as a conservative value, which often degrades the learning performance. In this article, the learning transient is analysed and the tuning of learnable bandwidth and lead step are developed to achieve good learning transient and tracking accuracy simultaneously. The attractive properties include that the less dependence on system model and that the tracking error during this process keeps at a very low level. Experimental results on an industrial robot are presented to verify the tuning process. 相似文献