Learning diverse rankings with multi-armed bandits
… This leads to learned ranking functions that produce rankings with redundant … ranks is often
preferred. We present two online learning algorithms that directly learn a diverse ranking of …
preferred. We present two online learning algorithms that directly learn a diverse ranking of …
[PDF][PDF] Learning diverse rankings with multi-armed bandits
… queries often have different meanings for different users (such as jaguar). We present a new
multi-armed bandit learning algorithm that directly learns a diverse ranking of results based …
multi-armed bandit learning algorithm that directly learns a diverse ranking of results based …
Diverse user preference elicitation with multi-armed bandits
J Parapar, F Radlinski - Proceedings of the 14th ACM international …, 2021 - dl.acm.org
… a Multi-Armed Bandits based algorithm [26] that improves both thematic and item diversity
during … [43] use MABs for learning diverse rankings of web documents based on user clicks, …
during … [43] use MABs for learning diverse rankings of web documents based on user clicks, …
Monte Carlo Elites: Quality-diversity selection as a multi-armed bandit problem
… , thus applying immense pressure to newly added individuals (or cells) while the opposite
is true with curiosity-based selection; (b) parent selection in this paper is performed by ranking …
is true with curiosity-based selection; (b) parent selection in this paper is performed by ranking …
[PDF][PDF] Ranked bandits in metric spaces: learning diverse rankings over large document collections
… that explicitly takes document similarity and ranking context into account. Our … multi-armed
bandit models from the literature: ranked bandits (Radlinski et al., 2008) and Lipschitz bandits (…
bandit models from the literature: ranked bandits (Radlinski et al., 2008) and Lipschitz bandits (…
Combining difficulty ranking with multi-armed bandits to sequence educational content
… content to students in order to maximize their learning gains over … MAPLE (Multi-Armed
Bandits based Personalization for Learning … a field study in the wild where students used different …
Bandits based Personalization for Learning … a field study in the wild where students used different …
[PDF][PDF] Learning to Diversify for E-commerce Search with Multi-Armed Bandit.
… (2) We present three multi-armed bandit based algorithms for this and derive the regret and
… by using a diverse ranking strategy. This connection between diversity in ranking and CLV …
… by using a diverse ranking strategy. This connection between diversity in ranking and CLV …
Gathering additional feedback on search results by multi-armed bandits with respect to production ranking
… To test the effectiveness of exploration provided by Algorithm 1 with different scoring functions
S, we compare them with ϵ-greedy [34], which is the simplest and widely used way to add …
S, we compare them with ϵ-greedy [34], which is the simplest and widely used way to add …
Learning with limited rounds of adaptivity: Coin tossing, multi-armed bandits, and ranking from pairwise comparisons
… learning settings, active/adaptive querying is possible, but the number of rounds of adaptivity
is limited. We study … in a stochastic multi-armed bandit, and the problem of top-$ k $ ranking …
is limited. We study … in a stochastic multi-armed bandit, and the problem of top-$ k $ ranking …
Multi-armed bandits for adjudicating documents in pooling-based evaluation of information retrieval systems
… ’s satisfaction by combining relevance and diversity. Kleinberg and colleagues (Radlinski, …
and diversity and presented a multi-armed bandit algorithm that learns a diversified ranking of …
and diversity and presented a multi-armed bandit algorithm that learns a diversified ranking of …