I3 retriever: incorporating implicit interaction in pre-trained language models for passage retrieval

Q Dong, Y Liu, Q Ai, H Li, S Wang, Y Liu… - Proceedings of the 32nd …, 2023 - dl.acm.org
Passage retrieval is a fundamental task in many information systems, such as web search
and question answering, where both efficiency and effectiveness are critical concerns. In …

CAME: Competitively Learning a Mixture-of-Experts Model for First-stage Retrieval

J Guo, Y Cai, K Bi, Y Fan, W Chen, R Zhang… - ACM Transactions on …, 2024 - dl.acm.org
The first-stage retrieval aims to retrieve a subset of candidate documents from a huge
collection both effectively and efficiently. Since various matching patterns can exist between …

Retrieval-based Disentangled Representation Learning with Natural Language Supervision

J Zhou, X Li, L Shang, X Jiang, Q Liu… - The Twelfth International …, 2024 - openreview.net
Disentangled representation learning remains challenging as the underlying factors of
variation in the data do not naturally exist. The inherent complexity of real-world data makes …

L2R: Lifelong Learning for First-stage Retrieval with Backward-Compatible Representations

Y Cai, K Bi, Y Fan, J Guo, W Chen… - Proceedings of the 32nd …, 2023 - dl.acm.org
First-stage retrieval is a critical task that aims to retrieve relevant document candidates from
a large-scale collection. While existing retrieval models have achieved impressive …

Pre-training with Large Language Model-based Document Expansion for Dense Passage Retrieval

G Ma, X Wu, P Wang, Z Lin, S Hu - arXiv preprint arXiv:2308.08285, 2023 - arxiv.org
In this paper, we systematically study the potential of pre-training with Large Language
Model (LLM)-based document expansion for dense passage retrieval. Concretely, we …

Query-as-context Pre-training for Dense Passage Retrieval

W Xing, G Ma, W Qian, Z Lin, S Hu - Proceedings of the 2023 …, 2023 - aclanthology.org
Recently, methods have been developed to improve the performance of dense passage
retrieval by using context-supervised pre-training. These methods simply consider two …

Drop your Decoder: Pre-training with Bag-of-Word Prediction for Dense Passage Retrieval.

G Ma, X Wu, Z Lin, S Hu - Proceedings of the 47th International ACM …, 2024 - dl.acm.org
Masked auto-encoder pre-training has emerged as a prevalent technique for initializing and
enhancing dense retrieval systems. It generally utilizes additional Transformer decoder …

Improving News Recommendation via Bottlenecked Multi-task Pre-training

X Xiao, Q Li, S Liu, K Zhou - Proceedings of the 46th International ACM …, 2023 - dl.acm.org
Recent years have witnessed the boom of deep neural networks in online news
recommendation service. As news articles mainly consist of textual content, pre-trained …

Query-as-context Pre-training for Dense Passage Retrieval

X Wu, G Ma, W Qian, Z Lin, S Hu - arXiv preprint arXiv:2212.09598, 2022 - arxiv.org
Recently, methods have been developed to improve the performance of dense passage
retrieval by using context-supervised pre-training. These methods simply consider two …

Cot-mae v2: contextual masked auto-encoder with multi-view modeling for passage retrieval

X Wu, G Ma, P Wang, M Lin, Z Lin, F Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Growing techniques have been emerging to improve the performance of passage retrieval.
As an effective representation bottleneck pretraining technique, the contextual masked auto …