Distributed bandits with heterogeneous agents

L Yang, YZJ Chen, MH Hajiemaili… - … -IEEE Conference on …, 2022 - ieeexplore.ieee.org
IEEE INFOCOM 2022-IEEE Conference on Computer Communications, 2022ieeexplore.ieee.org
This paper tackles a multi-agent bandit setting where M agents cooperate together to solve
the same instance of a K-armed stochastic bandit problem. The agents are heterogeneous:
each agent has limited access to a local subset of arms and the agents are asynchronous
with different gaps between decision-making rounds. The goal for each agent is to find its
optimal local arm, and agents can cooperate by sharing their observations with others. While
cooperation between agents improves the performance of learning, it comes with an …
This paper tackles a multi-agent bandit setting where M agents cooperate together to solve the same instance of a K-armed stochastic bandit problem. The agents are heterogeneous: each agent has limited access to a local subset of arms and the agents are asynchronous with different gaps between decision-making rounds. The goal for each agent is to find its optimal local arm, and agents can cooperate by sharing their observations with others. While cooperation between agents improves the performance of learning, it comes with an additional complexity of communication between agents. For this heterogeneous multi-agent setting, we propose two learning algorithms, CO-UCB and CO-AAE. We prove that both algorithms achieve order-optimal regret, which is , where is the minimum suboptimality gap between the reward mean of arm i and any local optimal arm. In addition, a careful selection of the valuable information for cooperation, CO-AAE achieves a low communication complexity of O(log T). Last, numerical experiments verify the efficiency of both algorithms.
ieeexplore.ieee.org
以上显示的是最相近的搜索结果。 查看全部搜索结果

Google学术搜索按钮

example.edu/paper.pdf
查找
获取 PDF 文件
引用
References