作者
Yangfan Zhou, Kaizhu Huang, Cheng Cheng, Xuguang Wang, Amir Hussain, Xin Liu
发表日期
2022/3/10
期刊
IEEE Transactions on Neural Networks and Learning Systems
卷号
34
期号
9
页码范围
6515-6529
出版商
IEEE
简介
AdaBelief, one of the current best optimizers, demonstrates superior generalization ability over the popular Adam algorithm by viewing the exponential moving average of observed gradients. AdaBelief is theoretically appealing in which it has a data-dependent regret bound when objective functions are convex, where is a time horizon. It remains, however, an open problem whether the convergence rate can be further improved without sacrificing its generalization ability. To this end, we make the first attempt in this work and design a novel optimization algorithm called FastAdaBelief that aims to exploit its strong convexity in order to achieve an even faster convergence rate. In particular, by adjusting the step size that better considers strong convexity and prevents fluctuation, our proposed FastAdaBelief demonstrates excellent generalization ability and superior convergence. As an important theoretical …
引用总数
学术搜索中的文章
Y Zhou, K Huang, C Cheng, X Wang, A Hussain, X Liu - IEEE Transactions on Neural Networks and Learning …, 2022