Online learning: A comprehensive survey

SCH Hoi, D Sahoo, J Lu, P Zhao - Neurocomputing, 2021 - Elsevier
Online learning represents a family of machine learning methods, where a learner attempts
to tackle some predictive (or any type of decision-making) task by learning from a sequence …

Recent advances of large-scale linear classification

GX Yuan, CH Ho, CJ Lin - Proceedings of the IEEE, 2012 - ieeexplore.ieee.org
Linear classification is a useful tool in machine learning and data mining. For some data in a
rich dimensional space, the performance (ie, testing accuracy) of linear classifiers has …

Findings of the 2017 conference on machine translation (wmt17)

O Bojar, R Chatterjee, C Federmann, Y Graham… - 2017 - doras.dcu.ie
This paper presents the results of the WMT17 shared tasks, which included three machine
translation (MT) tasks (news, biomedical, and multimodal), two evaluation tasks (metrics and …

Ead: elastic-net attacks to deep neural networks via adversarial examples

PY Chen, Y Sharma, H Zhang, J Yi… - Proceedings of the AAAI …, 2018 - ojs.aaai.org
Recent studies have highlighted the vulnerability of deep neural networks (DNNs) to
adversarial examples—a visually indistinguishable adversarial image can easily be crafted …

Only train once: A one-shot neural network training and pruning framework

T Chen, B Ji, T Ding, B Fang, G Wang… - Advances in …, 2021 - proceedings.neurips.cc
Structured pruning is a commonly used technique in deploying deep neural networks
(DNNs) onto resource-constrained devices. However, the existing pruning methods are …

Explicit inductive bias for transfer learning with convolutional networks

LI Xuhong, Y Grandvalet… - … Conference on Machine …, 2018 - proceedings.mlr.press
In inductive transfer learning, fine-tuning pre-trained convolutional networks substantially
outperforms training from scratch. When using fine-tuning, the underlying assumption is that …

Stochastic model-based minimization of weakly convex functions

D Davis, D Drusvyatskiy - SIAM Journal on Optimization, 2019 - SIAM
We consider a family of algorithms that successively sample and minimize simple stochastic
models of the objective function. We show that under reasonable conditions on …

Less is more: Towards compact cnns

H Zhou, JM Alvarez, F Porikli - … , The Netherlands, October 11–14, 2016 …, 2016 - Springer
To attain a favorable performance on large-scale datasets, convolutional neural networks
(CNNs) are usually designed to have very high capacity involving millions of parameters. In …

A proximal stochastic gradient method with progressive variance reduction

L Xiao, T Zhang - SIAM Journal on Optimization, 2014 - SIAM
We consider the problem of minimizing the sum of two convex functions: one is the average
of a large number of smooth component functions, and the other is a general convex …

[PDF][PDF] Adaptive subgradient methods for online learning and stochastic optimization.

J Duchi, E Hazan, Y Singer - Journal of machine learning research, 2011 - jmlr.org
We present a new family of subgradient methods that dynamically incorporate knowledge of
the geometry of the data observed in earlier iterations to perform more informative gradient …