A comprehensive survey of continual learning: theory, method and application

L Wang, X Zhang, H Su, J Zhu - IEEE Transactions on Pattern …, 2024 - ieeexplore.ieee.org
To cope with real-world dynamics, an intelligent system needs to incrementally acquire,
update, accumulate, and exploit knowledge throughout its lifetime. This ability, known as …

Online continual learning in image classification: An empirical survey

Z Mai, R Li, J Jeong, D Quispe, H Kim, S Sanner - Neurocomputing, 2022 - Elsevier
Online continual learning for image classification studies the problem of learning to classify
images from an online stream of data and tasks, where tasks may include new classes …

Audio-visual class-incremental learning

W Pian, S Mo, Y Guo, Y Tian - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
In this paper, we introduce audio-visual class-incremental learning, a class-incremental
learning scenario for audio-visual video recognition. We demonstrate that joint audio-visual …

Fine-tuned language models are continual learners

T Scialom, T Chakrabarty, S Muresan - arXiv preprint arXiv:2205.12393, 2022 - arxiv.org
Recent work on large language models relies on the intuition that most natural language
processing tasks can be described via natural language instructions. Language models …

Deepfake text detection: Limitations and opportunities

J Pu, Z Sarwar, SM Abdullah, A Rehman… - … IEEE symposium on …, 2023 - ieeexplore.ieee.org
Recent advances in generative models for language have enabled the creation of
convincing synthetic text or deepfake text. Prior work has demonstrated the potential for …

Continual learning in task-oriented dialogue systems

A Madotto, Z Lin, Z Zhou, S Moon, P Crook… - arXiv preprint arXiv …, 2020 - arxiv.org
Continual learning in task-oriented dialogue systems can allow us to add new domains and
functionalities through time without incurring the high cost of a whole system retraining. In …

Modifying memories in transformer models

C Zhu, AS Rawat, M Zaheer, S Bhojanapalli… - arXiv preprint arXiv …, 2020 - arxiv.org
Large Transformer models have achieved impressive performance in many natural
language tasks. In particular, Transformer based language models have been shown to …

Catastrophic forgetting in deep learning: A comprehensive taxonomy

EL Aleixo, JG Colonna, M Cristo… - arXiv preprint arXiv …, 2023 - arxiv.org
Deep Learning models have achieved remarkable performance in tasks such as image
classification or generation, often surpassing human accuracy. However, they can struggle …

Continual prompt tuning for dialog state tracking

Q Zhu, B Li, F Mi, X Zhu, M Huang - arXiv preprint arXiv:2203.06654, 2022 - arxiv.org
A desirable dialog system should be able to continually learn new skills without forgetting
old ones, and thereby adapt to new domains or tasks in its life cycle. However, continually …

Co-transport for class-incremental learning

DW Zhou, HJ Ye, DC Zhan - Proceedings of the 29th ACM International …, 2021 - dl.acm.org
Traditional learning systems are trained in closed-world for a fixed number of classes, and
need pre-collected datasets in advance. However, new classes often emerge in real-world …