Using natural language processing to support peer‐feedback in the age of artificial intelligence: A cross‐disciplinary framework and a research agenda

E Bauer, M Greisel, I Kuznetsov… - British Journal of …, 2023 - Wiley Online Library
Advancements in artificial intelligence are rapidly increasing. The new‐generation large
language models, such as ChatGPT and GPT‐4, bear the potential to transform educational …

Last layer re-training is sufficient for robustness to spurious correlations

P Kirichenko, P Izmailov, AG Wilson - arXiv preprint arXiv:2204.02937, 2022 - arxiv.org
Neural network classifiers can largely rely on simple spurious features, such as
backgrounds, to make predictions. However, even in these cases, we show that they still …

A survey on automated fact-checking

Z Guo, M Schlichtkrull, A Vlachos - Transactions of the Association for …, 2022 - direct.mit.edu
Fact-checking has become increasingly important due to the speed with which both
information and misinformation can spread in the modern media ecosystem. Therefore …

Discover and cure: Concept-aware mitigation of spurious correlation

S Wu, M Yuksekgonul, L Zhang… - … Conference on Machine …, 2023 - proceedings.mlr.press
Deep neural networks often rely on spurious correlations to make predictions, which hinders
generalization beyond training environments. For instance, models that associate cats with …

Measure and improve robustness in NLP models: A survey

X Wang, H Wang, D Yang - arXiv preprint arXiv:2112.08313, 2021 - arxiv.org
As NLP models achieved state-of-the-art performances over benchmarks and gained wide
applications, it has been increasingly important to ensure the safe deployment of these …

Shortcut learning of large language models in natural language understanding

M Du, F He, N Zou, D Tao, X Hu - Communications of the ACM, 2023 - dl.acm.org
Shortcut Learning of Large Language Models in Natural Language Understanding Page 1 110
COMMUNICATIONS OF THE ACM | JANUARY 2024 | VOL. 67 | NO. 1 research IMA GE B Y …

Evading the simplicity bias: Training a diverse set of models discovers solutions with superior ood generalization

D Teney, E Abbasnejad, S Lucey… - Proceedings of the …, 2022 - openaccess.thecvf.com
Neural networks trained with SGD were recently shown to rely preferentially on linearly-
predictive features and can ignore complex, equally-predictive ones. This simplicity bias can …

Distilling model failures as directions in latent space

S Jain, H Lawrence, A Moitra, A Madry - arXiv preprint arXiv:2206.14754, 2022 - arxiv.org
Existing methods for isolating hard subpopulations and spurious correlations in datasets
often require human intervention. This can make these methods labor-intensive and dataset …

Towards interpreting and mitigating shortcut learning behavior of NLU models

M Du, V Manjunatha, R Jain, R Deshpande… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent studies indicate that NLU models are prone to rely on shortcut features for prediction,
without achieving true language understanding. As a result, these models fail to generalize …

Take Care of Your Prompt Bias! Investigating and Mitigating Prompt Bias in Factual Knowledge Extraction

Z Xu, K Peng, L Ding, D Tao, X Lu - arXiv preprint arXiv:2403.09963, 2024 - arxiv.org
Recent research shows that pre-trained language models (PLMs) suffer from" prompt bias"
in factual knowledge extraction, ie, prompts tend to introduce biases toward specific labels …