首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 250 毫秒
1.
介绍输出概率密度函数(PDF)常规的迭代学习控制(ILC)的收敛条件,并利用此条件设计相应的迭代学习律。主要讨论如何解决输出PDF迭代学习控制(ILC)中的过迭代,收敛速度等问题。以离散输出概率密度函数(PDF)控制模型为基础,介绍了直接迭代学习控制算法收敛的必要条件,提出自适应的迭代学习参数调节方法和避免过迭代的迭代结束条件,这些措施能够保证输出PDF的迭代控制收敛且具有较快的收敛速度。仿真结果表明,输出PDF的自适应迭代学习控制具有较快的收敛速度,而学习终止条件能很好地避免过迭代。  相似文献   

2.
介绍输出概率密度函数(PDF)常规的迭代学习控制(ILC)的收敛条件,并利用此条件设计相应的迭代学习律.主要讨论如何解决输出PDF迭代学习控制(ILC)中的过迭代,收敛速度等问题.以离散输出概率密度函教(PDF)控制模型为基础,介绍了直接迭代学习控制算法收敛的必要条件,提出自适应的迭代学习参数调节方法和避免过迭代的迭代结束条件,这些措施能够保证输出PDF的迭代控制收敛且具有较快的收敛速度.仿真结果表明,输出PDF的自适应迭代学习控制具有较快的收敛速度,而学习终止条件能很好地避免过迭代.  相似文献   

3.
具有扰动的非线性系统高阶迭代学习控制   总被引:1,自引:0,他引:1  
迭代学习控制(ILC)利用系统的重复性不断改进控制性能.本文讨论一类具有扰动的非线性、时变系统高阶迭代学习控制算法及其迭代学习收敛的充分条件,并与D型迭代学习算法相比,讨论典型PD高阶ILC算法的收敛速度.仿真结果证实高阶ILC算法具有更快的收敛速度,并且当系统满足收敛条件、不确定项及输出扰动项有界时迭代学习收敛.  相似文献   

4.
带有初态学习的可变增益迭代学习控制   总被引:1,自引:0,他引:1  
曹伟  丛望  李金  郭媛 《控制与决策》2012,27(3):473-476
针对一类非线性系统提出一种新的学习控制算法,该算法在可变学习增益的迭代学习控制律基础上,增加了系统初态的迭代学习律.利用算子理论证明了系统在存在初态偏移时经过迭代学习后,其输出能够完全跟踪期望轨迹,同时得到了该算法谱半径形式的收敛条件.将该算法与传统迭代学习控制相比较可以看出,前者的收敛速度得到了较大提高,而且解决了可变学习增益迭代学习控制的初态偏移问题.仿真结果验证了该算法的有效性.  相似文献   

5.
对存在执行器故障的连续线性时变系统,给出了PID型迭代学习容错控制律的收敛条件。对连续时变故障系统设计了一种PID迭代学习容错控制律,在[λ]范数意义下给出了故障系统PID型迭代容错控制器收敛的充要条件;基于Schur补原理和不等式变换,将容错控制器收敛条件转换成线性矩阵不等式,当迭代学习收敛速度设定时,基于线性矩阵不等式能快速确定最优迭代控制增益,避免了迭代控制增益设置的盲目性。旋转控制系统的数值仿真,验证了PID迭代容错控制器优良的容错性能和跟踪性能。  相似文献   

6.
可变学习增益的迭代学习控制律   总被引:1,自引:0,他引:1       下载免费PDF全文
基于迭代学习控制理论提出了一种可变学习增益的迭代学习律,在非线性系统中对期望轨迹进行跟踪,与学习增益不变的迭代学习控制相比较,收敛速度得到很大的提高;通过对其收敛性进行严格的数学证明,得到了迭代学习律收敛的充分条件;在单机无穷大系统中,将该控制律应用于同步发电机的励磁控制,仿真结果表明该控制律的有效性,改善了控制的动态特性,有利于提高电力系统稳定性.  相似文献   

7.
针对时不变线性系统的迭代学习控制问题,提出了一种改进的时不变系统的PD型迭代学习控制算法,理论证明了系统满足收敛条件时的改进算法是收敛的。仿真实例分析表明,改进的算法利用最新算出的控制分量代替旧的控制分量,使系统的实际输出以更快的收敛速度逼近系统的理想输出。  相似文献   

8.
即时学习算法在非线性系统迭代学习控制中的应用   总被引:4,自引:1,他引:4       下载免费PDF全文
孙维  王伟  朱瑞军 《控制与决策》2003,18(3):263-266
运用即时学习算法来解决一类非线性系统的迭代学习控制初值问题。对于任何类型的迭代学习控制算法,即时学习算法都能有效地估计初始控制量,减小了初始输出误差,加快了算法的收敛速度,使得经过有限次迭代后系统输出能严格跟踪理想信号。对机器人系统的仿真结果表明了该方法的有效性。  相似文献   

9.
对倒立摆系统的平衡控制问题进行研究。在建立系统数学模型的基础上,提出指数变增益迭代学习控制律,并设计了控制器。通过系统仿真实验,结果表明:与常规迭代学习控制律相比较,本文采用的方法收敛速度大大加快,系统动态性能得到很大改善。  相似文献   

10.
基于经验数据库的迭代学习初始控制输入量的确定   总被引:6,自引:0,他引:6  
分析了初始控制输入量对迭代学习控制稳定性和收敛速度的影响,提出充分利用系统以往的控制经验来确定迭代学习初始控制输入量的思想,并给出3类确定方法——线性加权法、拟合曲线法和智能化法,对机器人对象的仿真结果表明,恰当地选取初始控制输入量,可使系统以较小的误差对新任务进行跟踪,进而减少迭代次数,提高学习控制的收敛速度,增强对新环境、新任务的适应能力。  相似文献   

11.
Auer  Peter  Long  Philip M.  Maass  Wolfgang  Woeginger  Gerhard J. 《Machine Learning》1995,18(2-3):187-230
The majority of results in computational learning theory are concerned with concept learning, i.e. with the special case of function learning for classes of functions with range {0, 1}. Much less is known about the theory of learning functions with a larger range such as or . In particular relatively few results exist about the general structure of common models for function learning, and there are only very few nontrivial function classes for which positive learning results have been exhibited in any of these models.We introduce in this paper the notion of a binary branching adversary tree for function learning, which allows us to give a somewhat surprising equivalent characterization of the optimal learning cost for learning a class of real-valued functions (in terms of a max-min definition which does not involve any learning model).Another general structural result of this paper relates the cost for learning a union of function classes to the learning costs for the individual function classes.Furthermore, we exhibit an efficient learning algorithm for learning convex piecewise linear functions from d into . Previously, the class of linear functions from d into was the only class of functions with multidimensional domain that was known to be learnable within the rigorous framework of a formal model for online learning.Finally we give a sufficient condition for an arbitrary class of functions from into that allows us to learn the class of all functions that can be written as the pointwise maximum ofk functions from . This allows us to exhibit a number of further nontrivial classes of functions from into for which there exist efficient learning algorithms.  相似文献   

12.
Kearns  Michael  Sebastian Seung  H. 《Machine Learning》1995,18(2-3):255-276
We introduce a new formal model in which a learning algorithm must combine a collection of potentially poor but statistically independent hypothesis functions in order to approximate an unknown target function arbitrarily well. Our motivation includes the question of how to make optimal use of multiple independent runs of a mediocre learning algorithm, as well as settings in which the many hypotheses are obtained by a distributed population of identical learning agents.  相似文献   

13.
In this paper we initiate an investigation of generalizations of the Probably Approximately Correct (PAC) learning model that attempt to significantly weaken the target function assumptions. The ultimate goal in this direction is informally termed agnostic learning, in which we make virtually no assumptions on the target function. The name derives from the fact that as designers of learning algorithms, we give up the belief that Nature (as represented by the target function) has a simple or succinct explanation. We give a number of positive and negative results that provide an initial outline of the possibilities for agnostic learning. Our results include hardness results for the most obvious generalization of the PAC model to an agnostic setting, an efficient and general agnostic learning method based on dynamic programming, relationships between loss functions for agnostic learning, and an algorithm for a learning problem that involves hidden variables.  相似文献   

14.
This article studies self-directed learning, a variant of the on-line (or incremental) learning model in which the learner selects the presentation order for the instances. Alternatively, one can view this model as a variation of learning with membership queries in which the learner is only charged for membership queries for which it could not predict the outcome. We give tight bounds on the complexity of self-directed learning for the concept classes of monomials, monotone DNF formulas, and axis-parallel rectangles in {0, 1, , n – 1} d . These results demonstrate that the number of mistakes under self-directed learning can be surprisingly small. We then show that learning complexity in the model of self-directed learning is less than that of all other commonly studied on-line and query learning models. Next we explore the relationship between the complexity of self-directed learning and the Vapnik-Chervonenkis (VC-)dimension. We show that, in general, the VC-dimension and the self-directed learning complexity are incomparable. However, for some special cases, we show that the VC-dimension gives a lower bound for the self-directed learning complexity. Finally, we explore a relationship between Mitchell's version space algorithm and the existence of self-directed learning algorithms that make few mistakes.  相似文献   

15.
Transfer in variable-reward hierarchical reinforcement learning   总被引:2,自引:1,他引:1  
Transfer learning seeks to leverage previously learned tasks to achieve faster learning in a new task. In this paper, we consider transfer learning in the context of related but distinct Reinforcement Learning (RL) problems. In particular, our RL problems are derived from Semi-Markov Decision Processes (SMDPs) that share the same transition dynamics but have different reward functions that are linear in a set of reward features. We formally define the transfer learning problem in the context of RL as learning an efficient algorithm to solve any SMDP drawn from a fixed distribution after experiencing a finite number of them. Furthermore, we introduce an online algorithm to solve this problem, Variable-Reward Reinforcement Learning (VRRL), that compactly stores the optimal value functions for several SMDPs, and uses them to optimally initialize the value function for a new SMDP. We generalize our method to a hierarchical RL setting where the different SMDPs share the same task hierarchy. Our experimental results in a simplified real-time strategy domain show that significant transfer learning occurs in both flat and hierarchical settings. Transfer is especially effective in the hierarchical setting where the overall value functions are decomposed into subtask value functions which are more widely amenable to transfer across different SMDPs.  相似文献   

16.
刘晓  毛宁 《数据采集与处理》2015,30(6):1310-1317
学习自动机(Learning automation,LA)是一种自适应决策器。其通过与一个随机环境不断交互学习从一个允许的动作集里选择最优的动作。在大多数传统的LA模型中,动作集总是被取作有限的。因此,对于连续参数学习问题,需要将动作空间离散化,并且学习的精度取决于离散化的粒度。本文提出一种新的连续动作集学习自动机(Continuous action set learning automaton,CALA),其动作集为一个可变区间,同时按照均匀分布方式选择输出动作。学习算法利用来自环境的二值反馈信号对动作区间的端点进行自适应更新。通过一个多模态学习问题的仿真实验,演示了新算法相对于3种现有CALA算法的优越性。  相似文献   

17.
Massive Open Online Courses (MOOCs) require individual learners to self-regulate their own learning, determining when, how and with what content and activities they engage. However, MOOCs attract a diverse range of learners, from a variety of learning and professional contexts. This study examines how a learner's current role and context influences their ability to self-regulate their learning in a MOOC: Introduction to Data Science offered by Coursera. The study compared the self-reported self-regulated learning behaviour between learners from different contexts and with different roles. Significant differences were identified between learners who were working as data professionals or studying towards a higher education degree and other learners in the MOOC. The study provides an insight into how an individual's context and role may impact their learning behaviour in MOOCs.  相似文献   

18.
19.
不同程度的监督机制在自动文本分类中的应用   总被引:1,自引:0,他引:1  
自动文本分类技术涉及信息检索、模式识别及机器学习等领域。本文以监督的程度为线索,综述了分属全监督,非监督以及半监督学习策略的若干方法-NBC(Naive Bayes Classifier),FCM(Fuzzy C-Means),SOM(Self-Organizing Map),ssFCM(serni-supervised Fuzzy C-Means)gSOM(guided Self-Organizing Map),并应用于文本分类中。其中,gSOM是我们在SOM基础上发展得到的半监督形式。并以Reuters-21578为语料,研究了监督程度对分类效果的影响,从而提出了对实际文本分类工作的建议。  相似文献   

20.
We study a model of probably exactly correct (PExact) learning that can be viewed either as the Exact model (learning from equivalence queries only) relaxed so that counterexamples to equivalence queries are distributionally drawn rather than adversarially chosen or as the probably approximately correct (PAC) model strengthened to require a perfect hypothesis. We also introduce a model of probably almost exactly correct (PAExact) learning that requires a hypothesis with negligible error and thus lies between the PExact and PAC models. Unlike the Exact and PExact models, PAExact learning is applicable to classes of functions defined over infinite instance spaces. We obtain a number of separation results between these models. Of particular note are some positive results for efficient parallel learning in the PAExact model, which stand in stark contrast to earlier negative results for efficient parallel Exact learning.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号