Efficient methods for natural language processing: A survey

M Treviso, JU Lee, T Ji, B Aken, Q Cao… - Transactions of the …, 2023 - direct.mit.edu
Recent work in natural language processing (NLP) has yielded appealing results from
scaling model parameters and training data; however, using only scale to improve …

An empirical study on robustness to spurious correlations using pre-trained language models

L Tu, G Lalwani, S Gella, H He - Transactions of the Association for …, 2020 - direct.mit.edu
Recent work has shown that pre-trained language models such as BERT improve
robustness to spurious correlations in the dataset. Intrigued by these results, we find that the …

Shortcut learning of large language models in natural language understanding

M Du, F He, N Zou, D Tao, X Hu - Communications of the ACM, 2023 - dl.acm.org
Shortcut Learning of Large Language Models in Natural Language Understanding Page 1 110
COMMUNICATIONS OF THE ACM | JANUARY 2024 | VOL. 67 | NO. 1 research IMA GE B Y …

Evaluating factuality in generation with dependency-level entailment

T Goyal, G Durrett - arXiv preprint arXiv:2010.05478, 2020 - arxiv.org
Despite significant progress in text generation models, a serious limitation is their tendency
to produce text that is factually inconsistent with information in the input. Recent work has …

Towards interpreting and mitigating shortcut learning behavior of NLU models

M Du, V Manjunatha, R Jain, R Deshpande… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent studies indicate that NLU models are prone to rely on shortcut features for prediction,
without achieving true language understanding. As a result, these models fail to generalize …

Identifying and mitigating spurious correlations for improving robustness in nlp models

T Wang, R Sridhar, D Yang, X Wang - arXiv preprint arXiv:2110.07736, 2021 - arxiv.org
Recently, NLP models have achieved remarkable progress across a variety of tasks;
however, they have also been criticized for being not robust. Many robustness problems can …

Menli: Robust evaluation metrics from natural language inference

Y Chen, S Eger - Transactions of the Association for Computational …, 2023 - direct.mit.edu
Recently proposed BERT-based evaluation metrics for text generation perform well on
standard benchmarks but are vulnerable to adversarial attacks, eg, relating to information …

Increasing robustness to spurious correlations using forgettable examples

Y Yaghoobzadeh, S Mehri, R Tachet, TJ Hazen… - arXiv preprint arXiv …, 2019 - arxiv.org
Neural NLP models tend to rely on spurious correlations between labels and input features
to perform their tasks. Minority examples, ie, examples that contradict the spurious …

Evaluating gender bias in natural language inference

S Sharma, M Dey, K Sinha - arXiv preprint arXiv:2105.05541, 2021 - arxiv.org
Gender-bias stereotypes have recently raised significant ethical concerns in natural
language processing. However, progress in detection and evaluation of gender bias in …

Improving the robustness of NLI models with minimax training

M Korakakis, A Vlachos - Proceedings of the 61st Annual Meeting …, 2023 - aclanthology.org
Natural language inference (NLI) models are susceptible to learning shortcuts, ie decision
rules that spuriously correlate with the label. As a result, they achieve high in-distribution …