Robustness of learning from task instructions

J Gu, H Zhao, H Xu, L Nie, H Mei, W Yin - arXiv preprint arXiv:2212.03813, 2022 - arxiv.org
Traditional supervised learning mostly works on individual tasks and requires training on a
large set of task-specific examples. This paradigm seriously hinders the development of task …

How to handle different types of out-of-distribution scenarios in computational argumentation? A comprehensive and fine-grained field study

A Waldis, Y Hou, I Gurevych - … of the 62nd Annual Meeting of the …, 2024 - aclanthology.org
The advent of pre-trained Language Models (LMs) has markedly advanced natural
language processing, but their efficacy in out-of-distribution (OOD) scenarios remains a …

Automated scoring of constructed response items in math assessment using large language models

W Morris, L Holmes, JS Choi, S Crossley - International Journal of Artificial …, 2024 - Springer
Recent developments in the field of artificial intelligence allow for improved performance in
the automated assessment of extended response items in mathematics, potentially allowing …

Robust Testing of AI Language Model Resiliency with Novel Adversarial Prompts

B Hannon, Y Kumar, D Gayle, JJ Li, P Morreale - Electronics, 2024 - mdpi.com
In the rapidly advancing field of Artificial Intelligence (AI), this study presents a critical
evaluation of the resilience and cybersecurity efficacy of leading AI models, including …

On measurement validity and language models: Increasing validity and decreasing bias with instructions

M Laurer, W van Atteveldt, A Casas… - … Methods and Measures, 2024 - Taylor & Francis
Language models like BERT or GPT are becoming increasingly popular measurement tools,
but are the measurements they produce valid? Literature suggests that there is still a …

Building Efficient Universal Classifiers with Natural Language Inference

M Laurer, W van Atteveldt, A Casas… - arXiv preprint arXiv …, 2023 - arxiv.org
Generative Large Language Models (LLMs) have become the mainstream choice for
fewshot and zeroshot learning thanks to the universality of text generation. Many users …

Collaboration between intelligent agents and large language models: A novel approach for enhancing code generation capability

X Bai, S Huang, C Wei, R Wang - Expert Systems with Applications, 2025 - Elsevier
Abstract Pre-trained Large Language Models (LLMs) have demonstrated significant
potential in the Natural Language to Code (NL2Code) task. However, user-provided natural …

PROTECT: Parameter-Efficient Tuning for Few-Shot Robust Chinese Text Correction

X Feng, T Gu, L Chang, X Liu - IEEE/ACM Transactions on …, 2024 - ieeexplore.ieee.org
Non-normative texts and euphemisms are widely spread on the Internet, making it more
difficult to moderate the content. These phenomena result from misspelling errors or …

Rethinking ChatGPT's Success: Usability and Cognitive Behaviors Enabled by Auto-regressive LLMs' Prompting

X Li, M Liu - arXiv preprint arXiv:2405.10474, 2024 - arxiv.org
Over the last decade, a wide range of training and deployment strategies for Large
Language Models (LLMs) have emerged. Among these, the prompting paradigms of Auto …

Fully-fused Multi-Layer Perceptrons on Intel Data Center GPUs

K Yuan, C Bauinger, X Zhang, P Baehr… - arXiv preprint arXiv …, 2024 - arxiv.org
This paper presents a SYCL implementation of Multi-Layer Perceptrons (MLPs), which
targets and is optimized for the Intel Data Center GPU Max 1550. To increase the …