首页 | 本学科首页   官方微博 | 高级检索  
     


Learning the mean: A neural network approach
Authors:Sergio Decherchi  Mauro Parodi  Sandro Ridella
Affiliation:1. Department of Drug Discovery and Development, Italian Institute of Technology Genoa, Italy;2. DIBE – Dept. Biophysical and Electronic Engineering, University of Genoa, Via Opera Pia 11a-16145 Genova, Italy;1. School of Electronics and Information Engineering, Xi’an Jiaotong University, Xi’an, China;2. Qualcomm Research, San Diego, CA, USA;1. School of Information Science and Engineering, Dalian Polytechnic University, Dalian 116034, China;2. School of Mathematical Sciences, Dalian University of Technology, Dalian 116024, China;3. School of Mathematics and Computational Science, China University of Petroleum, Dongying 257061, China;1. The State Key Laboratory for Manufacturing System Engineering, and System Engineering Institute, Xi’an Jiaotong University, Xi’an 710049, PR China;2. School of Mathematics and Statistics, Xidian University, Xi?an 710071, PR China;3. Institute of Nonlinear Science, Xianyang Normal University, Xianyang 712000, PR China;1. College of Computer Science & Technology, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, PR China;2. Collaborative Innovation Center of Novel Software Technology and Industrialization, Nanjing 210016, PR China
Abstract:One of the key problems in machine learning theory and practice is setting the correct value of the regularization parameter; this is particularly crucial in Kernel Machines such as Support Vector Machines, Regularized Least Square or Neural Networks with Weight Decay terms. Well known methods such as Leave-One-Out (or GCV) and Evidence Maximization offer a way of predicting the regularization parameter. This work points out the failure of these methods for predicting the regularization parameter when coping with the, apparently trivial and here introduced, regularized mean problem; this is the simplest form of Tikhonov regularization, that, in turn, is the primal form of the learning algorithm Regularized Least Squares. This controlled environment gives the possibility to define oracular notions of regularization and to experiment new methodologies for predicting the regularization parameter that can be extended to the more general regression case. The analysis stems from James–Stein theory, shows the equivalence of shrinking and regularization and is carried using multiple kernels learning for regression and SVD analysis; a mean value estimator is built, first via a rational function and secondly via a balanced neural network architecture suitable for estimating statistical quantities and gaining symmetric expectations. The obtained results show that a non-linear analysis of the sample and a non-linear estimation of the mean obtained by neural networks can be profitably used to improve the accuracy of mean value estimations, especially when a small number of realizations is provided.
Keywords:
本文献已被 ScienceDirect 等数据库收录!
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号