Promptagator: Few-shot dense retrieval from 8 examples

Z Dai, VY Zhao, J Ma, Y Luan, J Ni, J Lu… - arXiv preprint arXiv …, 2022 - arxiv.org
Much recent research on information retrieval has focused on how to transfer from one task
(typically with abundant supervised data) to various other tasks where supervision is limited …

Rankt5: Fine-tuning t5 for text ranking with ranking losses

H Zhuang, Z Qin, R Jagerman, K Hui, J Ma… - Proceedings of the 46th …, 2023 - dl.acm.org
Pretrained language models such as BERT have been shown to be exceptionally effective
for text ranking. However, there are limited studies on how to leverage more powerful …

Task-aware retrieval with instructions

A Asai, T Schick, P Lewis, X Chen, G Izacard… - arXiv preprint arXiv …, 2022 - arxiv.org
We study the problem of retrieval with instructions, where users of a retrieval system
explicitly describe their intent along with their queries. We aim to develop a general-purpose …

A survey of text classification with transformers: How wide? how large? how long? how accurate? how expensive? how safe?

J Fields, K Chovanec, P Madiraju - IEEE Access, 2024 - ieeexplore.ieee.org
Text classification in natural language processing (NLP) is evolving rapidly, particularly with
the surge in transformer-based models, including large language models (LLM). This paper …

Exaranker: Synthetic explanations improve neural rankers

F Ferraretto, T Laitz, R Lotufo, R Nogueira - Proceedings of the 46th …, 2023 - dl.acm.org
Recent work has shown that incorporating explanations into the output generated by large
language models (LLMs) can significantly enhance performance on a broad spectrum of …

Exaranker: Explanation-augmented neural ranker

F Ferraretto, T Laitz, R Lotufo, R Nogueira - arXiv preprint arXiv …, 2023 - arxiv.org
Recent work has shown that inducing a large language model (LLM) to generate
explanations prior to outputting an answer is an effective strategy to improve performance on …

Neuralmind-unicamp at 2022 trec neuclir: Large boring rerankers for cross-lingual retrieval

V Jeronymo, R Lotufo, R Nogueira - arXiv preprint arXiv:2303.16145, 2023 - arxiv.org
This paper reports on a study of cross-lingual information retrieval (CLIR) using the mT5-
XXL reranker on the NeuCLIR track of TREC 2022. Perhaps the biggest contribution of this …

Scaling laws for dense retrieval

Y Fang, J Zhan, Q Ai, J Mao, W Su, J Chen… - Proceedings of the 47th …, 2024 - dl.acm.org
Scaling laws have been observed in a wide range of tasks, particularly in language
generation. Previous studies have found that the performance of large language models …

Inpars toolkit: A unified and reproducible synthetic data generation pipeline for neural information retrieval

H Abonizio, L Bonifacio, V Jeronymo, R Lotufo… - arXiv preprint arXiv …, 2023 - arxiv.org
Recent work has explored Large Language Models (LLMs) to overcome the lack of training
data for Information Retrieval (IR) tasks. The generalization abilities of these models have …

Cross-lingual open-domain question answering with answer sentence generation

B Muller, L Soldaini, R Koncel-Kedziorski… - arXiv preprint arXiv …, 2021 - arxiv.org
Open-Domain Generative Question Answering has achieved impressive performance in
English by combining document-level retrieval with answer generation. These approaches …