首页 | 本学科首页   官方微博 | 高级检索  
     


Efficient cross-validation for kernelized least-squares regression with sparse basis expansions
Authors:Tapio Pahikkala  Hanna Suominen  Jorma Boberg
Affiliation:(1) Department of Information Technology and Turku Centre for Computer Science, University of Turku, 20014 Turku, Finland;(2) NICTA Canberra Research Laboratory, Canberra, Australia;(3) College of Engineering and Computer Science, Australian National University, Canberra, Australia
Abstract:We propose an efficient algorithm for calculating hold-out and cross-validation (CV) type of estimates for sparse regularized least-squares predictors. Holding out H data points with our method requires O(min(H 2 n,Hn 2)) time provided that a predictor with n basis vectors is already trained. In addition to holding out training examples, also some of the basis vectors used to train the sparse regularized least-squares predictor with the whole training set can be removed from the basis vector set used in the hold-out computation. In our experiments, we demonstrate the speed improvements provided by our algorithm in practice, and we empirically show the benefits of removing some of the basis vectors during the CV rounds.
Keywords:
本文献已被 SpringerLink 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号