Advances in variational inference

C Zhang, J Bütepage, H Kjellström… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
Many modern unsupervised or semi-supervised machine learning algorithms rely on
Bayesian probabilistic models. These models are usually intractable and thus require …

Virtual adversarial training: a regularization method for supervised and semi-supervised learning

T Miyato, S Maeda, M Koyama… - IEEE transactions on …, 2018 - ieeexplore.ieee.org
We propose a new regularization method based on virtual adversarial loss: a new measure
of local smoothness of the conditional label distribution given input. Virtual adversarial loss …

Diversity in machine learning

Z Gong, P Zhong, W Hu - Ieee Access, 2019 - ieeexplore.ieee.org
Machine learning methods have achieved good performance and been widely applied in
various real-world applications. They can learn the model adaptively and be better fit for …

Differentiable scaffolding tree for molecular optimization

T Fu, W Gao, C Xiao, J Yasonik, CW Coley… - arXiv preprint arXiv …, 2021 - arxiv.org
The structural design of functional molecules, also called molecular optimization, is an
essential chemical science and engineering task with important applications, such as drug …

Training deep models faster with robust, approximate importance sampling

TB Johnson, C Guestrin - Advances in Neural Information …, 2018 - proceedings.neurips.cc
In theory, importance sampling speeds up stochastic gradient algorithms for supervised
learning by prioritizing training examples. In practice, the cost of computing importances …

A mango picking vision algorithm on instance segmentation and key point detection from RGB images in an open orchard

C Zheng, P Chen, J Pang, X Yang, C Chen, S Tu… - Biosystems …, 2021 - Elsevier
Highlights•Designed an end-to-end vision system for a mango picking robot.•Instance
segmentation and pick point detection are performed simultaneously.•It is robust to various …

Minimal variance sampling with provable guarantees for fast training of graph neural networks

W Cong, R Forsati, M Kandemir… - Proceedings of the 26th …, 2020 - dl.acm.org
Sampling methods (eg, node-wise, layer-wise, or subgraph) has become an indispensable
strategy to speed up training large-scale Graph Neural Networks (GNNs). However, existing …

The implicit regularization of stochastic gradient flow for least squares

A Ali, E Dobriban, R Tibshirani - International conference on …, 2020 - proceedings.mlr.press
We study the implicit regularization of mini-batch stochastic gradient descent, when applied
to the fundamental problem of least squares regression. We leverage a continuous-time …

Gradient diversity: a key ingredient for scalable distributed learning

D Yin, A Pananjady, M Lam… - International …, 2018 - proceedings.mlr.press
It has been experimentally observed that distributed implementations of mini-batch
stochastic gradient descent (SGD) algorithms exhibit speedup saturation and decaying …

Determinantal point processes in randomized numerical linear algebra

M Derezinski, MW Mahoney - Notices of the American Mathematical …, 2021 - ams.org
Randomized Numerical Linear Algebra (RandNLA) is an area which uses randomness,
most notably random sampling and random projection methods, to develop improved …