Instructiongpt-4: A 200-instruction paradigm for fine-tuning minigpt-4

L Wei, Z Jiang, W Huang, L Sun - arXiv preprint arXiv:2308.12067, 2023 - arxiv.org
L Wei, Z Jiang, W Huang, L Sun
arXiv preprint arXiv:2308.12067, 2023arxiv.org
Multimodal large language models acquire their instruction-following capabilities through a
two-stage training process: pre-training on image-text pairs and fine-tuning on supervised
vision-language instruction data. Recent studies have shown that large language models
can achieve satisfactory results even with a limited amount of high-quality instruction-
following data. In this paper, we introduce InstructionGPT-4, which is fine-tuned on a small
dataset comprising only 200 examples, amounting to approximately 6% of the instruction …
Multimodal large language models acquire their instruction-following capabilities through a two-stage training process: pre-training on image-text pairs and fine-tuning on supervised vision-language instruction data. Recent studies have shown that large language models can achieve satisfactory results even with a limited amount of high-quality instruction-following data. In this paper, we introduce InstructionGPT-4, which is fine-tuned on a small dataset comprising only 200 examples, amounting to approximately 6% of the instruction-following data used in the alignment dataset for MiniGPT-4. We first propose several metrics to access the quality of multimodal instruction data. Based on these metrics, we present a simple and effective data selector to automatically identify and filter low-quality vision-language data. By employing this method, InstructionGPT-4 outperforms the original MiniGPT-4 on various evaluations (e.g., visual question answering, GPT-4 preference). Overall, our findings demonstrate that less but high-quality instruction tuning data is efficient to enable multimodal large language models to generate better output.
arxiv.org
以上显示的是最相近的搜索结果。 查看全部搜索结果

Google学术搜索按钮

example.edu/paper.pdf
查找
获取 PDF 文件
引用
References