Lifelong language pretraining with distribution-specialized experts

W Chen, Y Zhou, N Du, Y Huang… - International …, 2023 - proceedings.mlr.press
Pretraining on a large-scale corpus has become a standard method to build general
language models (LMs). Adapting a model to new data distributions targeting different …

Recent advances of foundation language models-based continual learning: A survey

Y Yang, J Zhou, X Ding, T Huai, S Liu, Q Chen… - ACM Computing …, 2024 - dl.acm.org
Recently, foundation language models (LMs) have marked significant achievements in the
domains of natural language processing (NLP) and computer vision (CV). Unlike traditional …

Continual learning for text classification with information disentanglement based regularization

Y Huang, Y Zhang, J Chen, X Wang, D Yang - arXiv preprint arXiv …, 2021 - arxiv.org
Continual learning has become increasingly important as it enables NLP models to
constantly learn and gain knowledge over time. Previous continual learning methods are …

Prototype-guided memory replay for continual learning

S Ho, M Liu, L Du, L Gao… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Continual learning (CL) is a machine learning paradigm that accumulates knowledge while
learning sequentially. The main challenge in CL is catastrophic forgetting of previously seen …

Mell: Large-scale extensible user intent classification for dialogue systems with meta lifelong learning

C Wang, H Pan, Y Liu, K Chen, M Qiu, W Zhou… - Proceedings of the 27th …, 2021 - dl.acm.org
User intent detection is vital for understanding their demands in dialogue systems. Although
the User Intent Classification (UIC) task has been widely studied, for large-scale industrial …

Prompt-based prototypical framework for continual relation extraction

H Zhang, B Liang, M Yang, H Wang… - IEEE/ACM Transactions …, 2022 - ieeexplore.ieee.org
Continual relation extraction (CRE) is an important task of continual learning, which aims to
learn incessantly emerging new relations between entities from texts. To avoid …

Power Norm Based Lifelong Learning for Paraphrase Generations

D Li, P Yang, Y Zhang, P Li - Proceedings of the 46th International ACM …, 2023 - dl.acm.org
Lifelong seq2seq language generation models are trained with multiple domains in a
lifelong learning manner, with data from each domain being observed in an online fashion. It …

Meta-learning with variational semantic memory for word sense disambiguation

Y Du, N Holla, X Zhen, CGM Snoek… - arXiv preprint arXiv …, 2021 - arxiv.org
A critical challenge faced by supervised word sense disambiguation (WSD) is the lack of
large annotated datasets with sufficient coverage of words in their diversity of senses. This …

Towards a robust experimental framework and benchmark for lifelong language learning

A Hussain, N Holla, P Mishra… - Thirty-fifth Conference …, 2021 - openreview.net
In lifelong learning, a model learns different tasks sequentially throughout its lifetime. State-
of-the-art deep learning models, however, struggle to generalize in this setting and suffer …

Prioritized experience replay method based on experience reward

J Gao, X Li, W Liu, J Zhao - 2021 International Conference on …, 2021 - ieeexplore.ieee.org
In recent years, artificial intelligence has been widely used in modern construction, and
reinforcement learning methods have played an important role in it. The experience replay …