Distributed multi-player bandits-a game of thrones approach

I Bistritz, A Leshem - Advances in Neural Information …, 2018 - proceedings.neurips.cc
We consider a multi-armed bandit game where N players compete for K arms for T turns.
Each player has different expected rewards for the arms, and the instantaneous rewards are …

Optimal algorithms for multiplayer multi-armed bandits

PA Wang, A Proutiere, K Ariu… - International …, 2020 - proceedings.mlr.press
The paper addresses various Multiplayer Multi-Armed Bandit (MMAB) problems, where M
decision-makers, or players, collaborate to maximize their cumulative reward. We first …

SIC-MMAB: Synchronisation involves communication in multiplayer multi-armed bandits

E Boursier, V Perchet - Advances in Neural Information …, 2019 - proceedings.neurips.cc
Motivated by cognitive radio networks, we consider the stochastic multiplayer multi-armed
bandit problem, where several players pull arms simultaneously and collisions occur if one …

A practical algorithm for multiplayer bandits when arm means vary among players

A Mehrabian, E Boursier… - International …, 2020 - proceedings.mlr.press
We study a multiplayer stochastic multi-armed bandit problem in which players cannot
communicate, and if two or more players pull the same arm, a collision occurs and the …

Cooperative stochastic bandits with asynchronous agents and constrained feedback

L Yang, YZJ Chen, S Pasteris… - Advances in …, 2021 - proceedings.neurips.cc
This paper studies a cooperative multi-armed bandit problem with $ M $ agents cooperating
together to solve the same instance of a $ K $-armed stochastic bandit problem with the goal …

Heterogeneous multi-player multi-armed bandits: Closing the gap and generalization

C Shi, W Xiong, C Shen, J Yang - Advances in neural …, 2021 - proceedings.neurips.cc
Despite the significant interests and many progresses in decentralized multi-player multi-
armed bandits (MP-MAB) problems in recent years, the regret gap to the natural centralized …

Dominate or delete: Decentralized competing bandits in serial dictatorship

A Sankararaman, S Basu… - International …, 2021 - proceedings.mlr.press
Online learning in a two-sided matching market, with demand side agents continuously
competing to be matched with supply side (arms), abstracts the complex interactions under …

Multi-player multi-armed bandits for stable allocation in heterogeneous ad-hoc networks

SJ Darak, MK Hanawal - IEEE Journal on Selected Areas in …, 2019 - ieeexplore.ieee.org
Next generation networks are expected to be ultra-dense and aim to explore spectrum
sharing paradigm that allows users to communicate in licensed, shared as well as …

Multiplayer bandits without observing collision information

G Lugosi, A Mehrabian - Mathematics of Operations …, 2022 - pubsonline.informs.org
We study multiplayer stochastic multiarmed bandit problems in which the players cannot
communicate, and if two or more players pull the same arm, a collision occurs and the …

Decentralized multi-player multi-armed bandits with no collision information

C Shi, W Xiong, C Shen, J Yang - … Conference on Artificial …, 2020 - proceedings.mlr.press
The decentralized stochastic multi-player multi-armed bandit (MP-MAB) problem, where the
collision information is not available to the players, is studied in this paper. Building on the …