[HTML][HTML] A review of uncertainty quantification in deep learning: Techniques, applications and challenges

M Abdar, F Pourpanah, S Hussain, D Rezazadegan… - Information fusion, 2021 - Elsevier
Uncertainty quantification (UQ) methods play a pivotal role in reducing the impact of
uncertainties during both optimization and decision making processes. They have been …

Bayesian model-agnostic meta-learning

J Yoon, T Kim, O Dia, S Kim… - Advances in neural …, 2018 - proceedings.neurips.cc
Due to the inherent model uncertainty, learning to infer Bayesian posterior from a few-shot
dataset is an important step towards robust meta-learning. In this paper, we propose a novel …

Just pick a sign: Optimizing deep multitask models with gradient sign dropout

Z Chen, J Ngiam, Y Huang, T Luong… - Advances in …, 2020 - proceedings.neurips.cc
The vast majority of deep models use multiple gradient signals, typically corresponding to a
sum of multiple loss terms, to update a shared set of trainable weights. However, these …

Regularizing generative adversarial networks under limited data

HY Tseng, L Jiang, C Liu, MH Yang… - Proceedings of the …, 2021 - openaccess.thecvf.com
Recent years have witnessed the rapid progress of generative adversarial networks (GANs).
However, the success of the GAN models hinges on a large amount of training data. This …

C-mixup: Improving generalization in regression

H Yao, Y Wang, L Zhang, JY Zou… - Advances in neural …, 2022 - proceedings.neurips.cc
Improving the generalization of deep networks is an important open challenge, particularly
in domains without plentiful data. The mixup algorithm improves generalization by linearly …

Meta-learning with fewer tasks through task interpolation

H Yao, L Zhang, C Finn - arXiv preprint arXiv:2106.02695, 2021 - arxiv.org
Meta-learning enables algorithms to quickly learn a newly encountered task with just a few
labeled examples by transferring previously learned knowledge. However, the bottleneck of …

The role of deconfounding in meta-learning

Y Jiang, Z Chen, K Kuang, L Yuan… - International …, 2022 - proceedings.mlr.press
Meta-learning has emerged as a potent paradigm for quick learning of few-shot tasks, by
leveraging the meta-knowledge learned from meta-training tasks. Well-generalized meta …

Parameter-efficient mixture-of-experts architecture for pre-trained language models

ZF Gao, P Liu, WX Zhao, ZY Lu, JR Wen - arXiv preprint arXiv:2203.01104, 2022 - arxiv.org
Recently, Mixture-of-Experts (short as MoE) architecture has achieved remarkable success
in increasing the model capacity of large-scale language models. However, MoE requires …

Adversarial task up-sampling for meta-learning

Y Wu, LK Huang, Y Wei - Advances in Neural Information …, 2022 - proceedings.neurips.cc
The success of meta-learning on existing benchmarks is predicated on the assumption that
the distribution of meta-training tasks covers meta-testing tasks. Frequent violation of the …

Stochastic backpropagation: A memory efficient strategy for training video models

F Cheng, M Xu, Y Xiong, H Chen… - Proceedings of the …, 2022 - openaccess.thecvf.com
We propose a memory efficient method, named Stochastic Backpropagation (SBP), for
training deep neural networks on videos. It is based on the finding that gradients from …