首页 | 本学科首页   官方微博 | 高级检索  
     


Improving Generalization with Active Learning
Authors:Cohn  David  Atlas  Les  Ladner  Richard
Affiliation:(1) Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, 02139 Cambridge, MA;(2) Department of Electrical Engineering, University of Washington, 98195 Seattle, WA;(3) Department of Computer Science and Engineering, University of Washington, 98195 Seattle, WA
Abstract:Active learning differs from ldquolearning from examplesrdquo in that the learning algorithm assumes at least some control over what part of the input domain it receives information about. In some situations, active learning is provably more powerful than learning from examples alone, giving better generalization for a fixed number of training examples.In this article, we consider the problem of learning a binary concept in the absence of noise. We describe a formalism for active concept learning calledselective sampling and show how it may be approximately implemented by a neural network. In selective sampling, a learner receives distribution information from the environment and queries an oracle on parts of the domain it considers ldquouseful.rdquo We test our implementation, called anSG-network, on three domains and observe significant improvement in generalization.A preliminary version of this article appears as Cohn et al. (1990).
Keywords:queries  active learning  generalization  version space  neural networks
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号