Understanding plasticity in neural networks

C Lyle, Z Zheng, E Nikishin, BA Pires… - International …, 2023 - proceedings.mlr.press
Plasticity, the ability of a neural network to quickly change its predictions in response to new
information, is essential for the adaptability and robustness of deep reinforcement learning …

Convolutional neural operators for robust and accurate learning of PDEs

B Raonic, R Molinaro, T De Ryck… - Advances in …, 2024 - proceedings.neurips.cc
Although very successfully used in conventional machine learning, convolution based
neural network architectures--believed to be inconsistent in function space--have been …

Loss of plasticity in deep continual learning

S Dohare, JF Hernandez-Garcia, Q Lan, P Rahman… - Nature, 2024 - nature.com
Artificial neural networks, deep-learning methods and the backpropagation algorithm form
the foundation of modern machine learning and artificial intelligence. These methods are …

Interpretable and explainable logical policies via neurally guided symbolic abstraction

Q Delfosse, H Shindo, D Dhami… - Advances in Neural …, 2024 - proceedings.neurips.cc
The limited priors required by neural networks make them the dominating choice to encode
and learn policies using reinforcement learning (RL). However, they are also black-boxes …

PEDRO: Parameter-Efficient Fine-tuning with Prompt DEpenDent Representation MOdification

T Xie, T Li, W Zhu, W Han, Y Zhao - arXiv preprint arXiv:2409.17834, 2024 - arxiv.org
Due to their substantial sizes, large language models (LLMs) are typically deployed within a
single-backbone multi-tenant framework. In this setup, a single instance of an LLM …

IAPT: Instance-Aware Prompt Tuning for Large Language Models

W Zhu, A Tian, C Yin, Y Ni, X Wang… - Proceedings of the 62nd …, 2024 - aclanthology.org
Soft prompt tuning is a widely studied parameter-efficient fine-tuning method. However, it
has a clear drawback: many soft tokens must be inserted into the input sequences to …

Interpretable concept bottlenecks to align reinforcement learning agents

Q Delfosse, S Sztwiertnia, M Rothermel… - arXiv preprint arXiv …, 2024 - arxiv.org
Goal misalignment, reward sparsity and difficult credit assignment are only a few of the many
issues that make it difficult for deep reinforcement learning (RL) agents to learn optimal …

Adaptable adapters

NS Moosavi, Q Delfosse, K Kersting… - arXiv preprint arXiv …, 2022 - arxiv.org
State-of-the-art pretrained NLP models contain a hundred million to trillion parameters.
Adapters provide a parameter-efficient alternative for the full finetuning in which we can only …

Boosting object representation learning via motion and object continuity

Q Delfosse, W Stammer, T Rothenbächer… - … Conference on Machine …, 2023 - Springer
Recent unsupervised multi-object detection models have shown impressive performance
improvements, largely attributed to novel architectural inductive biases. Unfortunately …

Streaming Deep Reinforcement Learning Finally Works

M Elsayed, G Vasan, AR Mahmood - arXiv preprint arXiv:2410.14606, 2024 - arxiv.org
Natural intelligence processes experience as a continuous stream, sensing, acting, and
learning moment-by-moment in real time. Streaming learning, the modus operandi of classic …