Exaranker: Synthetic explanations improve neural rankers

F Ferraretto, T Laitz, R Lotufo, R Nogueira - Proceedings of the 46th …, 2023 - dl.acm.org
Recent work has shown that incorporating explanations into the output generated by large
language models (LLMs) can significantly enhance performance on a broad spectrum of …

[PDF][PDF] No parameter left behind: How distillation and model size affect zero-shot retrieval

GM Rosa, L Bonifacio, V Jeronymo… - arXiv preprint arXiv …, 2022 - researchgate.net
Recent work has shown that small distilled language models are strong competitors to
models that are orders of magnitude larger and slower in a wide range of information …

In defense of cross-encoders for zero-shot retrieval

G Rosa, L Bonifacio, V Jeronymo, H Abonizio… - arXiv preprint arXiv …, 2022 - arxiv.org
Bi-encoders and cross-encoders are widely used in many state-of-the-art retrieval pipelines.
In this work we study the generalization ability of these two types of architectures on a wide …

Exaranker: Explanation-augmented neural ranker

F Ferraretto, T Laitz, R Lotufo, R Nogueira - arXiv preprint arXiv …, 2023 - arxiv.org
Recent work has shown that inducing a large language model (LLM) to generate
explanations prior to outputting an answer is an effective strategy to improve performance on …

Mmead: Ms marco entity annotations and disambiguations

C Kamphuis, A Lin, S Yang, J Lin, AP de Vries… - Proceedings of the 46th …, 2023 - dl.acm.org
MMEAD, or MS MARCO Entity Annotations and Disambiguations, is a resource for entity
links for the MS MARCO datasets. We specify a format to store and share links for both …

A study on the efficiency and generalization of light hybrid retrievers

M Luo, S Jain, A Gupta, A Einolghozati, B Oguz… - arXiv preprint arXiv …, 2022 - arxiv.org
Hybrid retrievers can take advantage of both sparse and dense retrievers. Previous hybrid
retrievers leverage indexing-heavy dense retrievers. In this work, we study" Is it possible to …

Isotropic representation can improve dense retrieval

E Jung, J Park, J Choi, S Kim, W Rhee - Pacific-Asia Conference on …, 2023 - Springer
Abstract The latest Dense Retrieval (DR) models typically encode queries and documents
using BERT and subsequently apply a cosine similarity-based scoring to determine the …

Revisiting Document Expansion and Filtering for Effective First-Stage Retrieval

W Mansour, S Zhuang, G Zuccon… - Proceedings of the 47th …, 2024 - dl.acm.org
Document expansion is a technique that aims to reduce the likelihood of term mismatch by
augmenting documents with related terms or queries. Doc2Query minus minus (Doc2Query …

Neural ranking with weak supervision for open-domain question answering: A survey

X Shen, S Vakulenko, M Del Tredici… - Findings of the …, 2023 - aclanthology.org
Neural ranking (NR) has become a key component for open-domain question-answering in
order to access external knowledge. However, training a good NR model requires …

Lossy Compression Options for Dense Index Retention

J Mackenzie, A Moffat - Proceedings of the Annual International ACM …, 2023 - dl.acm.org
Dense indexes derived from whole-of-document neural models are now more effective at
locating likely-relevant documents than are conventional term-based inverted indexes. That …