Adaptive compositional continual meta-learning

B Wu, J Fang, X Zeng, S Liang… - … on Machine Learning, 2023 - proceedings.mlr.press
This paper focuses on continual meta-learning, where few-shot tasks are heterogeneous
and sequentially available. Recent works use a mixture model for meta-knowledge to deal …

Interpretable self-aware neural networks for robust trajectory prediction

M Itkina, M Kochenderfer - Conference on Robot Learning, 2023 - proceedings.mlr.press
Although neural networks have seen tremendous success as predictive models in a variety
of domains, they can be overly confident in their predictions on out-of-distribution (OOD) …

Failures are fated, but can be faded: Characterizing and mitigating unwanted behaviors in large-scale vision and language models

S Sagar, A Taparia, R Senanayake - arXiv preprint arXiv:2406.07145, 2024 - arxiv.org
In large deep neural networks that seem to perform surprisingly well on many tasks, we also
observe a few failures related to accuracy, social biases, and alignment with human values …

Learning discrete structured variational auto-encoder using natural evolution strategies

A Berliner, G Rotman, Y Adi, R Reichart… - arXiv preprint arXiv …, 2022 - arxiv.org
Discrete variational auto-encoders (VAEs) are able to represent semantic latent spaces in
generative learning. In many real-life settings, the discrete latent space consists of high …

MultiMax: Sparse and Multi-Modal Attention Learning

Y Zhou, M Fritz, M Keuper - arXiv preprint arXiv:2406.01189, 2024 - arxiv.org
SoftMax is a ubiquitous ingredient of modern machine learning algorithms. It maps an input
vector onto a probability simplex and reweights the input by concentrating the probability …

[PDF][PDF] Learnable Sparsity and Weak Supervision for Data-Efficient, Transparent, and Compact Neural Models

GMMA Correia - 2022 - goncalomcorreia.com
Neural network models have become ubiquitous in the machine learning literature. These
models are compositions of differentiable building blocks that compute dense …

Scalable Multi-Modal Continual Meta-Learning

B Wu, S Liang, Q Zhang - openreview.net
This paper focuses on continual meta-learning, where few-shot tasks are sequentially
available and sampled from a non-stationary distribution. Motivated by this challenging …

Adaptive Sparse Softmax: An Effective and Efficient Softmax Variant for Text Classification

Q Lv, L Geng, Z Cao, M Cao, S Li, W Li, G Fu - openreview.net
Softmax with the cross entropy loss is the standard configuration for current neural text
classification models. The gold score for a target class is supposed to be 1, but it is never …