PARADE: Passage Representation Aggregation forDocument Reranking
Pre-trained transformer models, such as BERT and T5, have shown to be highly effective at
ad hoc passage and document ranking. Due to the inherent sequence length limits of these …
ad hoc passage and document ranking. Due to the inherent sequence length limits of these …
Local self-attention over long text for efficient document retrieval
Neural networks, particularly Transformer-based architectures, have achieved significant
performance improvements on several retrieval benchmarks. When the items being …
performance improvements on several retrieval benchmarks. When the items being …
Modeling diverse relevance patterns in ad-hoc retrieval
Assessing relevance between a query and a document is challenging in ad-hoc retrieval
due to its diverse patterns, ie, a document could be relevant to a query as a whole or …
due to its diverse patterns, ie, a document could be relevant to a query as a whole or …
Intra-document cascading: Learning to select passages for neural document ranking
An emerging recipe for achieving state-of-the-art effectiveness in neural document re-
ranking involves utilizing large pre-trained language models-eg, BERT-to evaluate all …
ranking involves utilizing large pre-trained language models-eg, BERT-to evaluate all …
PageRank without hyperlinks: Structural reranking using links induced by language models
The ad hoc retrieval task is to find documents in a corpus that are relevant to a query.
Inspired by the PageRank and HITS (hubs and authorities) algorithms for Web search, we …
Inspired by the PageRank and HITS (hubs and authorities) algorithms for Web search, we …
Quality-biased ranking of web documents
Many existing retrieval approaches do not take into account the content quality of the
retrieved documents, although link-based measures such as PageRank are commonly used …
retrieved documents, although link-based measures such as PageRank are commonly used …
Unsupervised FAQ retrieval with question generation and BERT
We focus on the task of Frequently Asked Questions (FAQ) retrieval. A given user query can
be matched against the questions and/or the answers in the FAQ. We present a fully …
be matched against the questions and/or the answers in the FAQ. We present a fully …
BLADE: combining vocabulary pruning and intermediate pretraining for scaleable neural CLIR
Learning sparse representations using pretrained language models enhances the
monolingual ranking effectiveness. Such representations are sparse vectors in the …
monolingual ranking effectiveness. Such representations are sparse vectors in the …
Modeling higher-order term dependencies in information retrieval using query hypergraphs
M Bendersky, WB Croft - Proceedings of the 35th international ACM …, 2012 - dl.acm.org
Many of the recent, and more effective, retrieval models have incorporated dependencies
between the terms in the query. In this paper, we advance this query representation one step …
between the terms in the query. In this paper, we advance this query representation one step …
Finding text reuse on the web
M Bendersky, WB Croft - Proceedings of the Second ACM International …, 2009 - dl.acm.org
With the overwhelming number of reports on similar events originating from different sources
on the web, it is often hard, using existing web search paradigms, to find the original source …
on the web, it is often hard, using existing web search paradigms, to find the original source …