Distributed Learning in Multi-Armed Bandit With Multiple Players

Type: Article

Publication Date: 2010-08-05

Citations: 407

DOI: https://doi.org/10.1109/tsp.2010.2062509

Abstract

We formulate and study a decentralized multi-armed bandit (MAB) problem. There are M distributed players competing for N independent arms. Each arm, when played, offers i.i.d. reward according to a distribution with an unknown parameter. At each time, each player chooses one arm to play without exchanging observations or any information with other players. Players choosing the same arm collide, and, depending on the collision model, either no one receives reward or the colliding players share the reward in an arbitrary way. We show that the minimum system regret of the decentralized MAB grows with time at the same logarithmic order as in the centralized counterpart where players act collectively as a single entity by exchanging observations and making decisions jointly. A decentralized policy is constructed to achieve this optimal order while ensuring fairness among players and without assuming any pre-agreement or information exchange among players. Based on a Time Division Fair Sharing (TDFS) of the M best arms, the proposed policy is constructed and its order optimality is proven under a general reward model. Furthermore, the basic structure of the TDFS policy can be used with any order-optimal single-player policy to achieve order optimality in the decentralized setting. We also establish a lower bound on the system regret growth rate for a general class of decentralized polices, to which the proposed policy belongs. This problem finds potential applications in cognitive radio networks, multi-channel communication systems, multi-agent systems, web search and advertising, and social networks.

Locations

  • IEEE Transactions on Signal Processing - View
  • arXiv (Cornell University) - View - PDF
  • DataCite API - View

Similar Works

Action Title Year Authors
+ PDF Chat Decentralized learning for multi-player multi-armed bandits 2012 Dileep Kalathil
Naumaan Nayyar
Rahul Jain
+ Decentralized Heterogeneous Multi-Player Multi-Armed Bandits with Non-Zero Rewards on Collisions 2019 Akshayaa Magesh
Venugopal V. Veeravalli
+ PDF Chat Decentralized Learning for Multiplayer Multiarmed Bandits 2014 Dileep Kalathil
Naumaan Nayyar
Rahul Jain
+ Decentralized Restless Bandit with Multiple Players and Unknown Dynamics 2011 Haoyang Liu
Keqin Liu
Qing Zhao
+ PDF Chat Decentralized Stochastic Multi-Player Multi-Armed Walking Bandits 2023 Guojun Xiong
Jian Li
+ PDF Chat Decentralized Heterogeneous Multi-Player Multi-Armed Bandits with Non-Zero Rewards on Collisions 2021 Akshayaa Magesh
Venugopal V. Veeravalli
+ PDF Chat Decentralized Online Learning Algorithms for Opportunistic Spectrum Access 2011 Yi Gai
B. Krishnamachari
+ Decentralized Online Learning Algorithms for Opportunistic Spectrum Access 2011 Yi Gai
Bhaskar Krishnamachari
+ Decentralized Stochastic Multi-Player Multi-Armed Walking Bandits 2022 Guojun Xiong
Jian Li
+ Distributed Bandits with Heterogeneous Agents 2022 Lin Yang
Yu-Zhen Janice Chen
Mohammad Hajiesmaili
John CS Lui
Don Towsley
+ Learning to coordinate without communication in multi-user multi-armed bandit problems. 2015 Orly Avner
Shie Mannor
+ Multi-Player Bandits Models Revisited 2017 Lilian Besson
Emilie Kaufmann
+ PDF Chat Distributed Bandits with Heterogeneous Agents 2022 Lin Yang
Yu-Zhen Janice Chen
Mohammad H. Hajiemaili
John C. S. Lui
Don Towsley
+ Multi-Player Multi-Armed Bandits with Finite Shareable Resources Arms: Learning Algorithms & Applications 2022 Xuchuang Wang
Hong Xie
John C. S. Lui
+ Optimal Fair Multi-Agent Bandits 2023 Amir Leshem
+ Distributed Multi-Player Bandits - a Game of Thrones Approach 2018 Ilai Bistritz
Amir Leshem
+ PDF Chat QuACK: A Multipurpose Queuing Algorithm for Cooperative $k$-Armed Bandits 2024 Benjamin Howson
Sarah Filippi
Ciara Pike-Burke
+ Multi-Player Bandits Revisited 2017 Lilian Besson
Emilie Kaufmann
+ Distributed Multi-Player Bandits - a Game of Thrones Approach 2018 Ilai Bistritz
Amir Leshem
+ An Optimal Algorithm for Multiplayer Multi-Armed Bandits 2019 Alexandre Proutière
Po-An Wang