Chain-of-skills: A configurable model for open-domain question answering

K Ma, H Cheng, Y Zhang, X Liu, E Nyberg… - arXiv preprint arXiv …, 2023 - arxiv.org
arXiv preprint arXiv:2305.03130, 2023arxiv.org
The retrieval model is an indispensable component for real-world knowledge-intensive
tasks, eg, open-domain question answering (ODQA). As separate retrieval skills are
annotated for different datasets, recent work focuses on customized methods, limiting the
model transferability and scalability. In this work, we propose a modular retriever where
individual modules correspond to key skills that can be reused across datasets. Our
approach supports flexible skill configurations based on the target domain to boost …
The retrieval model is an indispensable component for real-world knowledge-intensive tasks, e.g., open-domain question answering (ODQA). As separate retrieval skills are annotated for different datasets, recent work focuses on customized methods, limiting the model transferability and scalability. In this work, we propose a modular retriever where individual modules correspond to key skills that can be reused across datasets. Our approach supports flexible skill configurations based on the target domain to boost performance. To mitigate task interference, we design a novel modularization parameterization inspired by sparse Transformer. We demonstrate that our model can benefit from self-supervised pretraining on Wikipedia and fine-tuning using multiple ODQA datasets, both in a multi-task fashion. Our approach outperforms recent self-supervised retrievers in zero-shot evaluations and achieves state-of-the-art fine-tuned retrieval performance on NQ, HotpotQA and OTT-QA.
arxiv.org
以上显示的是最相近的搜索结果。 查看全部搜索结果