Comparing score aggregation approaches for document retrieval with pretrained transformers

X Zhang, A Yates, J Lin - Advances in Information Retrieval: 43rd …, 2021 - Springer
… applying the model to document retrieval. In this … pretrained transformers that share similarities
with BERT. We find that these BERT variants are not more effective for document retrieval

Efficient document re-ranking for transformers by precomputing term representations

S MacAvaney, FM Nardini, R Perego… - … in Information Retrieval, 2020 - dl.acm.org
… To show the generality of our approach we present tests conducted also for other pretrained
transformers in Section 6.5: a version of BERT that was more effectively pre-trained, ie, …

[图书][B] Pretrained transformers for text ranking: Bert and beyond

J Lin, R Nogueira, A Yates - 2022 - books.google.com
… Historically, document expansion techniques have not been as popular as query expansion
… interest in document expansion in the context of transformers, which we cover in Chapter 4. …

How different are pre-trained transformers for text ranking?

D Rau, J Kamps - European Conference on Information Retrieval, 2022 - Springer
… In recent years, large pre-trained transformers have led to substantial gains in performance
over traditional retrieval … Does CE better rank the same documents retrieved by BM25? …

Local self-attention over long text for efficient document retrieval

S Hofstätter, H Zamani, B Mitra, N Craswell… - … in Information Retrieval, 2020 - dl.acm.org
… [16] use pretrained contextual embeddings, without fine-tuning, in downstream ranking … In
this work we proposed a solution to apply Transformers to full document re-ranking. Our TKL …

HIBERT: Document level pre-training of hierarchical bidirectional transformers for document summarization

X Zhang, F Wei, M Zhou - arXiv preprint arXiv:1905.06566, 2019 - arxiv.org
… from Transformers. We design an unsupervised method to pre-train HIBERT for document
modeling. We apply the pre-trained HIBERT to the task of document summarization and …

Webformer: Pre-training with web pages for information retrieval

Y Guo, Z Ma, J Mao, H Qian, X Zhang, H Jiang… - … in Information Retrieval, 2022 - dl.acm.org
… , we pre-train the Transformer model towards the supervised … At the finetuning stage, we use
our pre-trained Text Encoder … for effective web-document retrieval. Applied Intelligence 18, 3 …

Pre-training tasks for embedding-based large-scale retrieval

WC Chang, FX Yu, YW Chang, Y Yang… - arXiv preprint arXiv …, 2020 - arxiv.org
… We consider the large-scale query-document retrieval problem: given a query (eg, a …
Transformer model, we compare the pretraining tasks to two baselines: No Pretraining and MLM. No …

[HTML][HTML] Pre-trained transformers: an empirical comparison

S Casola, I Lauriola, A Lavelli - Machine Learning with Applications, 2022 - Elsevier
… answer to a given question in relevant documents, eg, retrieved by a search engine. The
latter finds an exact text span in a document (or typically a paragraph) containing the answer. …

MedCPT: Contrastive Pre-trained Transformers with large-scale PubMed search logs for zero-shot biomedical information retrieval

Q Jin, W Kim, Q Chen, DC Comeau, L Yeganova… - …, 2023 - academic.oup.com
… Specifically, unlike the in-batch negative documents used by the MedCPT retriever that …
documents are sampled from rank e to rank f in the top retrieved documents by the pre-trained