作者
Peter Auer, Nicolo Cesa-Bianchi, Claudio Gentile
发表日期
2002/2/1
期刊
Journal of Computer and System Sciences
卷号
64
期号
1
页码范围
48-75
出版商
Academic Press
简介
We study on-line learning in the linear regression framework. Most of the performance bounds for on-line algorithms in this framework assume a constant learning rate. To achieve these bounds the learning rate must be optimized based on a posteriori information. This information depends on the whole sequence of examples and thus it is not available to any strictly on-line algorithm. We introduce new techniques for adaptively tuning the learning rate as the data sequence is progressively revealed. Our techniques allow us to prove essentially the same bounds as if we knew the optimal learning rate in advance. Moreover, such techniques apply to a wide class of on-line algorithms, including p-norm algorithms for generalized linear regression and Weighted Majority for linear regression with absolute loss. Our adaptive tunings are radically different from previous techniques, such as the so-called doubling trick …
引用总数
200120022003200420052006200720082009201020112012201320142015201620172018201920202021202220232024387121314121412917111421912912111719321918
学术搜索中的文章
P Auer, N Cesa-Bianchi, C Gentile - Journal of Computer and System Sciences, 2002