Deep learning and the global workspace theory

R VanRullen, R Kanai - Trends in Neurosciences, 2021 - cell.com
Recent advances in deep learning have allowed artificial intelligence (AI) to reach near
human-level performance in many sensory, perceptual, linguistic, and cognitive tasks. There …

Meta-seg: A survey of meta-learning for image segmentation

S Luo, Y Li, P Gao, Y Wang, S Serikawa - Pattern Recognition, 2022 - Elsevier
A well-performed deep learning model in image segmentation relies on a large number of
labeled data. However, it is hard to obtain sufficient high-quality raw data in industrial …

Editing models with task arithmetic

G Ilharco, MT Ribeiro, M Wortsman… - arXiv preprint arXiv …, 2022 - arxiv.org
Changing how pre-trained models behave--eg, improving their performance on a
downstream task or mitigating biases learned during pre-training--is a common practice …

The power of scale for parameter-efficient prompt tuning

B Lester, R Al-Rfou, N Constant - arXiv preprint arXiv:2104.08691, 2021 - arxiv.org
In this work, we explore" prompt tuning", a simple yet effective mechanism for learning" soft
prompts" to condition frozen language models to perform specific downstream tasks. Unlike …

Efficiently identifying task groupings for multi-task learning

C Fifty, E Amid, Z Zhao, T Yu… - Advances in Neural …, 2021 - proceedings.neurips.cc
Multi-task learning can leverage information learned by one task to benefit the training of
other tasks. Despite this capacity, naively training all tasks together in one model often …

Multi-task learning with deep neural networks: A survey

M Crawshaw - arXiv preprint arXiv:2009.09796, 2020 - arxiv.org
Multi-task learning (MTL) is a subfield of machine learning in which multiple tasks are
simultaneously learned by a shared model. Such approaches offer advantages like …

Spot: Better frozen model adaptation through soft prompt transfer

T Vu, B Lester, N Constant, R Al-Rfou, D Cer - arXiv preprint arXiv …, 2021 - arxiv.org
There has been growing interest in parameter-efficient methods to apply pre-trained
language models to downstream tasks. Building on the Prompt Tuning approach of Lester et …

Merging models with fisher-weighted averaging

MS Matena, CA Raffel - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Averaging the parameters of models that have the same architecture and initialization can
provide a means of combining their respective capabilities. In this paper, we take the …

Sharp-maml: Sharpness-aware model-agnostic meta learning

M Abbas, Q Xiao, L Chen, PY Chen… - … on machine learning, 2022 - proceedings.mlr.press
Abstract Model-agnostic meta learning (MAML) is currently one of the dominating
approaches for few-shot meta-learning. Albeit its effectiveness, the optimization of MAML …

Which tasks should be learned together in multi-task learning?

T Standley, A Zamir, D Chen, L Guibas… - International …, 2020 - proceedings.mlr.press
Many computer vision applications require solving multiple tasks in real-time. A neural
network can be trained to solve multiple tasks simultaneously using multi-task learning. This …