Large, pre-trained language models (PLMs) such as BERT and GPT have drastically changed the Natural Language Processing (NLP) field. For numerous NLP tasks …
Fine-tuning large pre-trained language models on downstream tasks has become an important paradigm in NLP. However, common practice fine-tunes all of the parameters in a …
Transformer models have achieved remarkable results in various natural language tasks, but they are often prohibitively large, requiring massive memories and computational …
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of NLP. We find that the distribution shift settings in previous studies commonly lack adequate …
Pre-trained language models can be fine-tuned to solve diverse NLP tasks, including in few- shot settings. Thus fine-tuning allows the model to quickly pick up task-specific" skills," but …
Large Transformer-based models have exhibited superior performance in various natural language processing and computer vision tasks. However, these models contain enormous …
The ability to generalise well is one of the primary desiderata of natural language processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …
Abstract Edge Artificial Intelligence (AI) incorporates a network of interconnected systems and devices that receive, cache, process, and analyse data in close communication with the …
Y Peng, K Kim, F Wu, P Sridhar… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
Self-supervised speech representation learning (SSL) has shown to be effective in various downstream tasks, but SSL models are usually large and slow. Model compression …