Priors in bayesian deep learning: A review

V Fortuin - International Statistical Review, 2022 - Wiley Online Library
While the choice of prior is one of the most critical parts of the Bayesian inference workflow,
recent Bayesian deep learning models have often fallen back on vague priors, such as …

Deep learning for source code modeling and generation: Models, applications, and challenges

THM Le, H Chen, MA Babar - ACM Computing Surveys (CSUR), 2020 - dl.acm.org
Deep Learning (DL) techniques for Natural Language Processing have been evolving
remarkably fast. Recently, the DL advances in language modeling, machine translation, and …

Attention, please! A survey of neural attention models in deep learning

A de Santana Correia, EL Colombini - Artificial Intelligence Review, 2022 - Springer
In humans, Attention is a core property of all perceptual and cognitive operations. Given our
limited ability to process competing sources, attention mechanisms select, modulate, and …

Conditional neural processes

M Garnelo, D Rosenbaum… - International …, 2018 - proceedings.mlr.press
Deep neural networks excel at function approximation, yet they are typically trained from
scratch for each new function. On the other hand, Bayesian methods, such as Gaussian …

Attentive neural processes

H Kim, A Mnih, J Schwarz, M Garnelo, A Eslami… - arXiv preprint arXiv …, 2019 - arxiv.org
Neural Processes (NPs)(Garnelo et al 2018a; b) approach regression by learning to map a
context set of observed input-output pairs to a distribution over regression functions. Each …

VAE with a VampPrior

J Tomczak, M Welling - International conference on artificial …, 2018 - proceedings.mlr.press
Many different methods to train deep generative models have been introduced in the past. In
this paper, we propose to extend the variational auto-encoder (VAE) framework with a new …

Sample efficient adaptive text-to-speech

Y Chen, Y Assael, B Shillingford, D Budden… - arXiv preprint arXiv …, 2018 - arxiv.org
We present a meta-learning approach for adaptive text-to-speech (TTS) with few data.
During training, we learn a multi-speaker model using a shared conditional WaveNet core …

Latent alignment and variational attention

Y Deng, Y Kim, J Chiu, D Guo… - Advances in neural …, 2018 - proceedings.neurips.cc
Neural attention has become central to many state-of-the-art models in natural language
processing and related domains. Attention networks are an easy-to-train and effective …

Prototypical variational autoencoder for 3d few-shot object detection

W Tang, B Yang, X Li, YH Liu… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Few-Shot 3D Point Cloud Object Detection (FS3D) is a challenging task, aiming to
detect 3D objects of novel classes using only limited annotated samples for training …

MANNER: A variational memory-augmented model for cross domain few-shot named entity recognition

J Fang, X Wang, Z Meng, P Xie, F Huang… - Proceedings of the …, 2023 - aclanthology.org
This paper focuses on the task of cross domain few-shot named entity recognition (NER),
which aims to adapt the knowledge learned from source domain to recognize named entities …