Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

Unifying large language models and knowledge graphs: A roadmap

S Pan, L Luo, Y Wang, C Chen… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Large language models (LLMs), such as ChatGPT and GPT4, are making new waves in the
field of natural language processing and artificial intelligence, due to their emergent ability …

Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning

H Liu, D Tam, M Muqeeth, J Mohta… - Advances in …, 2022 - proceedings.neurips.cc
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …

Rlprompt: Optimizing discrete text prompts with reinforcement learning

M Deng, J Wang, CP Hsieh, Y Wang, H Guo… - arXiv preprint arXiv …, 2022 - arxiv.org
Prompting has shown impressive success in enabling large pretrained language models
(LMs) to perform diverse NLP tasks, especially when only few downstream data are …

Fine-tuning language models with just forward passes

S Malladi, T Gao, E Nichani… - Advances in …, 2023 - proceedings.neurips.cc
Fine-tuning language models (LMs) has yielded success on diverse downstream tasks, but
as LMs grow in size, backpropagation requires a prohibitively large amount of memory …

Active prompting with chain-of-thought for large language models

S Diao, P Wang, Y Lin, T Zhang - arXiv preprint arXiv:2302.12246, 2023 - arxiv.org
The increasing scale of large language models (LLMs) brings emergent abilities to various
complex tasks requiring reasoning, such as arithmetic and commonsense reasoning. It is …

Language models as black-box optimizers for vision-language models

S Liu, S Yu, Z Lin, D Pathak… - Proceedings of the …, 2024 - openaccess.thecvf.com
Vision-language models (VLMs) pre-trained on web-scale datasets have demonstrated
remarkable capabilities on downstream tasks when fine-tuned with minimal data. However …

Learn from model beyond fine-tuning: A survey

H Zheng, L Shen, A Tang, Y Luo, H Hu, B Du… - arXiv preprint arXiv …, 2023 - arxiv.org
Foundation models (FM) have demonstrated remarkable performance across a wide range
of tasks (especially in the fields of natural language processing and computer vision) …

Automatic prompt augmentation and selection with chain-of-thought from labeled data

KS Shum, S Diao, T Zhang - arXiv preprint arXiv:2302.12822, 2023 - arxiv.org
Chain-of-thought prompting (CoT) advances the reasoning abilities of large language
models (LLMs) and achieves superior performance in arithmetic, commonsense, and …

Promptboosting: Black-box text classification with ten forward passes

B Hou, J O'connor, J Andreas… - International …, 2023 - proceedings.mlr.press
We describe PromptBoosting, a query-efficient procedure for building a text classifier from a
neural language model (LM) without access to the LM's parameters, gradients, or hidden …