首页 | 本学科首页   官方微博 | 高级检索  
     


Flexible latent variable models for multi-task learning
Authors:Jian Zhang  Zoubin Ghahramani  Yiming Yang
Affiliation:(1) Department of Statistics, Purdue University, West Lafayette, IN 47907, USA;(2) Department of Engineering, University of Cambridge, Cambridge, CB2 1PZ, UK;(3) School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213, USA
Abstract:Given multiple prediction problems such as regression or classification, we are interested in a joint inference framework that can effectively share information between tasks to improve the prediction accuracy, especially when the number of training examples per problem is small. In this paper we propose a probabilistic framework which can support a set of latent variable models for different multi-task learning scenarios. We show that the framework is a generalization of standard learning methods for single prediction problems and it can effectively model the shared structure among different prediction tasks. Furthermore, we present efficient algorithms for the empirical Bayes method as well as point estimation. Our experiments on both simulated datasets and real world classification datasets show the effectiveness of the proposed models in two evaluation settings: a standard multi-task learning setting and a transfer learning setting.
Keywords:Multi-task learning  Latent variable models  Hierarchical Bayesian models  Model selection  Transfer learning
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号