Combinatorial bandits with relative feedback
We consider combinatorial online learning with subset choices when only relative feedback
information from subsets is available, instead of bandit or semi-bandit feedback which is
absolute. Specifically, we study two regret minimisation problems over subsets of a finite
ground set $[n] $, with subset-wise relative preference information feedback according to the
Multinomial logit choice model. In the first setting, the learner can play subsets of size
bounded by a maximum size and receives top-$ m $ rank-ordered feedback, while in the …
information from subsets is available, instead of bandit or semi-bandit feedback which is
absolute. Specifically, we study two regret minimisation problems over subsets of a finite
ground set $[n] $, with subset-wise relative preference information feedback according to the
Multinomial logit choice model. In the first setting, the learner can play subsets of size
bounded by a maximum size and receives top-$ m $ rank-ordered feedback, while in the …
Abstract
We consider combinatorial online learning with subset choices when only relative feedback information from subsets is available, instead of bandit or semi-bandit feedback which is absolute. Specifically, we study two regret minimisation problems over subsets of a finite ground set , with subset-wise relative preference information feedback according to the Multinomial logit choice model. In the first setting, the learner can play subsets of size bounded by a maximum size and receives top- rank-ordered feedback, while in the second setting the learner can play subsets of a fixed size with a full subset ranking observed as feedback. For both settings, we devise instance-dependent and order-optimal regret algorithms with regret and , respectively. We derive fundamental limits on the regret performance of online learning with subset-wise preferences, proving the tightness of our regret guarantees. Our results also show the value of eliciting more general top- rank-ordered feedback over single winner feedback (). Our theoretical results are corroborated with empirical evaluations.
proceedings.neurips.cc
以上显示的是最相近的搜索结果。 查看全部搜索结果