[图书][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
The goal of text ranking is to generate an ordered list of texts retrieved from a corpus in
response to a query. Although the most common formulation of text ranking is search …

Utilizing BERT for Information Retrieval: Survey, Applications, Resources, and Challenges

J Wang, JX Huang, X Tu, J Wang, AJ Huang… - ACM Computing …, 2024 - dl.acm.org
Recent years have witnessed a substantial increase in the use of deep learning to solve
various natural language processing (NLP) problems. Early deep learning models were …

PARADE: Passage Representation Aggregation forDocument Reranking

C Li, A Yates, S MacAvaney, B He, Y Sun - ACM Transactions on …, 2023 - dl.acm.org
Pre-trained transformer models, such as BERT and T5, have shown to be highly effective at
ad hoc passage and document ranking. Due to the inherent sequence length limits of these …

mmarco: A multilingual version of the ms marco passage ranking dataset

L Bonifacio, V Jeronymo, HQ Abonizio… - arXiv preprint arXiv …, 2021 - arxiv.org
The MS MARCO ranking dataset has been widely used for training deep learning models for
IR tasks, achieving considerable effectiveness on diverse zero-shot scenarios. However, this …

Injecting the BM25 score as text improves BERT-based re-rankers

A Askari, A Abolghasemi, G Pasi, W Kraaij… - … on Information Retrieval, 2023 - Springer
In this paper we propose a novel approach for combining first-stage lexical retrieval models
and Transformer-based re-rankers: we inject the relevance score of the lexical model as a …

Squeezing water from a stone: a bag of tricks for further improving cross-encoder effectiveness for reranking

R Pradeep, Y Liu, X Zhang, Y Li, A Yates… - European Conference on …, 2022 - Springer
While much recent work has demonstrated that hard negative mining can be used to train
better bi-encoder models, few have considered it in the context of cross-encoders, which are …

Toward best practices for training multilingual dense retrieval models

X Zhang, K Ogueji, X Ma, J Lin - ACM Transactions on Information …, 2023 - dl.acm.org
Dense retrieval models using a transformer-based bi-encoder architecture have emerged as
an active area of research. In this article, we focus on the task of monolingual retrieval in a …

Vera: Prediction techniques for reducing harmful misinformation in consumer health search

R Pradeep, X Ma, R Nogueira, J Lin - Proceedings of the 44th …, 2021 - dl.acm.org
The COVID-19 pandemic has brought about a proliferation of harmful news articles online,
with sources lacking credibility and misrepresenting scientific facts. Misinformation has real …

Can Old TREC Collections Reliably Evaluate Modern Neural Retrieval Models?

EM Voorhees, I Soboroff, J Lin - arXiv preprint arXiv:2201.11086, 2022 - arxiv.org
Neural retrieval models are generally regarded as fundamentally different from the retrieval
techniques used in the late 1990's when the TREC ad hoc test collections were constructed …

PARM: A paragraph aggregation retrieval model for dense document-to-document retrieval

S Althammer, S Hofstätter, M Sertkan… - … on Information Retrieval, 2022 - Springer
Dense passage retrieval (DPR) models show great effectiveness gains in first stage retrieval
for the web domain. However in the web domain we are in a setting with large amounts of …