作者
Miguel Lázaro-Gredilla, Aníbal R Figueiras-Vidal
发表日期
2010/7/1
期刊
IEEE transactions on neural networks
卷号
21
期号
8
页码范围
1345-1351
出版商
IEEE
简介
For regression tasks, traditional neural networks (NNs) have been superseded by Gaussian processes, which provide probabilistic predictions (input-dependent error bars), improved accuracy, and virtually no overfitting. Due to their high computational cost, in scenarios with massive data sets, one has to resort to sparse Gaussian processes, which strive to achieve similar performance with much smaller computational effort. In this context, we introduce a mixture of NNs with marginalized output weights that can both provide probabilistic predictions and improve on the performance of sparse Gaussian processes, at the same computational cost. The effectiveness of this approach is shown experimentally on some representative large data sets.
引用总数
201120122013201420152016201720182019202020212022202320242116132336782
学术搜索中的文章
M Lázaro-Gredilla, AR Figueiras-Vidal - IEEE transactions on neural networks, 2010