首页 | 本学科首页   官方微博 | 高级检索  
     


Consistency of discrete Bayesian learning
Authors:Jan Poland
Affiliation:ABB Switzerland Ltd. Corporate Research, Segelhof, CH-5405 Baden-Dättwil, Switzerland
Abstract:Bayes’ rule specifies how to obtain a posterior from a class of hypotheses endowed with a prior and the observed data. There are three fundamental ways to use this posterior for predicting the future: marginalization (integration over the hypotheses w.r.t. the posterior), MAP (taking the a posteriori most probable hypothesis), and stochastic model selection (selecting a hypothesis at random according to the posterior distribution). If the hypothesis class is countable, and contains the data generating distribution (this is termed the “realizable case”), strong consistency theorems are known for the former two methods in a sequential prediction framework, asserting almost sure convergence of the predictions to the truth as well as loss bounds. We prove corresponding results for stochastic model selection, for both discrete and continuous observation spaces. As a main technical tool, we will use the concept of a potential: this quantity, which is always positive, measures the total possible amount of future prediction errors. Precisely, in each time step, the expected potential decrease upper bounds the expected error. We introduce the entropy potential of a hypothesis class as its worst-case entropy, with regard to the true distribution. Our results are proven within a general stochastic online prediction framework, that comprises both online classification and prediction of non-i.i.d. sequences.
Keywords:Bayesian learning   Consistency   Stochastic model selection   Marginalization   MAP prediction   Almost sure convergence   Entropy potential
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号