A comprehensive overview of large language models

H Naveed, AU Khan, S Qiu, M Saqib, S Anwar… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have recently demonstrated remarkable capabilities in
natural language processing tasks and beyond. This success of LLMs has led to a large …

Unleashing the power of data tsunami: A comprehensive survey on data assessment and selection for instruction tuning of language models

Y Qin, Y Yang, P Guo, G Li, H Shao, Y Shi, Z Xu… - arXiv preprint arXiv …, 2024 - arxiv.org
Instruction tuning plays a critical role in aligning large language models (LLMs) with human
preference. Despite the vast amount of open instruction datasets, naively training a LLM on …

Leak, cheat, repeat: Data contamination and evaluation malpractices in closed-source llms

S Balloccu, P Schmidtová, M Lango… - arXiv preprint arXiv …, 2024 - arxiv.org
Natural Language Processing (NLP) research is increasingly focusing on the use of Large
Language Models (LLMs), with some of the most popular ones being either fully or partially …

Rethinking semi-supervised learning with language models

Z Shi, F Tonolini, N Aletras, E Yilmaz, G Kazai… - arXiv preprint arXiv …, 2023 - arxiv.org
Semi-supervised learning (SSL) is a popular setting aiming to effectively utilize unlabelled
data to improve model performance in downstream natural language processing (NLP) …

Learning to generate instruction tuning datasets for zero-shot task adaptation

NV Nayak, Y Nan, A Trost, SH Bach - arXiv preprint arXiv:2402.18334, 2024 - arxiv.org
We introduce Bonito, an open-source model for conditional task generation: the task of
converting unannotated text into task-specific training datasets for instruction tuning. Our …

Ask the experts: sourcing a high-quality nutrition counseling dataset through Human-AI collaboration

S Balloccu, E Reiter, KJH Li, R Sargsyan… - Findings of the …, 2024 - aura.abdn.ac.uk
Abstract Large Language Models (LLMs) are being employed by end-users for various
tasks, including sensitive ones such as health counseling, disregarding potential safety …

Lexical Entrainment for Conversational Systems

Z Shi, P Sen, A Lipani - arXiv preprint arXiv:2310.09651, 2023 - arxiv.org
Conversational agents have become ubiquitous in assisting with daily tasks, and are
expected to possess human-like features. One such feature is lexical entrainment (LE), a …

Self contrastive learning for session-based recommendation

Z Shi, X Wang, A Lipani - European Conference on Information Retrieval, 2024 - Springer
Session-based recommendation, which aims to predict the next item of users' interest as per
an existing sequence interaction of items, has attracted growing applications of Contrastive …

Rethink the Effectiveness of Text Data Augmentation: An Empirical Analysis

Z Shi, A Lipani - arXiv preprint arXiv:2306.07664, 2023 - arxiv.org
In recent years, language models (LMs) have made remarkable progress in advancing the
field of natural language processing (NLP). However, the impact of data augmentation (DA) …

Understanding the Role of User Profile in the Personalization of Large Language Models

B Wu, Z Shi, HA Rahmani, V Ramineni… - arXiv preprint arXiv …, 2024 - arxiv.org
Utilizing user profiles to personalize Large Language Models (LLMs) has been shown to
enhance the performance on a wide range of tasks. However, the precise role of user …