Multi-Armed Bandits with Abstention

Type: Preprint

Publication Date: 2024-02-23

Citations: 0

DOI: https://doi.org/10.48550/arxiv.2402.15127

Abstract

We introduce a novel extension of the canonical multi-armed bandit problem that incorporates an additional strategic element: abstention. In this enhanced framework, the agent is not only tasked with selecting an arm at each time step, but also has the option to abstain from accepting the stochastic instantaneous reward before observing it. When opting for abstention, the agent either suffers a fixed regret or gains a guaranteed reward. Given this added layer of complexity, we ask whether we can develop efficient algorithms that are both asymptotically and minimax optimal. We answer this question affirmatively by designing and analyzing algorithms whose regrets meet their corresponding information-theoretic lower bounds. Our results offer valuable quantitative insights into the benefits of the abstention option, laying the groundwork for further exploration in other online decision-making problems with such an option. Numerical results further corroborate our theoretical findings.

Locations

  • arXiv (Cornell University) - View - PDF

Similar Works

Action Title Year Authors
+ Learning in Restless Multi-Armed Bandits via Adaptive Arm Sequencing Rules. 2019 Tomer Gafni
Kobi Cohen
+ Learning in Restless Multi-Armed Bandits via Adaptive Arm Sequencing Rules 2019 Tomer Gafni
Kobi Cohen
+ A Closer Look at the Worst-case Behavior of Multi-armed Bandit Algorithms 2021 Anand Kalvit
Assaf Zeevi
+ Stochastic Rising Bandits 2022 Alberto Maria Metelli
Francesco Trovò
Matteo Pirola
Marcello Restelli
+ A Closer Look at the Worst-case Behavior of Multi-armed Bandit Algorithms 2021 Anand Kalvit
Assaf Zeevi
+ A Closer Look at the Worst-case Behavior of Multi-armed Bandit Algorithms 2021 Anand Kalvit
Assaf Zeevi
+ Discrete Choice Multi-Armed Bandits 2023 Emerson Melo
David E. Muller
+ Bounded regret in stochastic multi-armed bandits 2013 Sébastien Bubeck
Vianney Perchet
Philippe Rigollet
+ PDF Chat Optimism in the Face of Ambiguity Principle for Multi-Armed Bandits 2024 Mengmeng Li
Daniel Kühn
Bahar Taşkesen
+ Perturbed-History Exploration in Stochastic Multi-Armed Bandits. 2019 Branislav Kveton
Csaba Szepesvári
Mohammad Ghavamzadeh
Craig Boutilier
+ Ballooning Multi-Armed Bandits 2020 Ganesh Ghalme
Swapnil Dhamal
Shweta Jain
Sujit Gujar
Y. Narahari
+ Perturbed-History Exploration in Stochastic Multi-Armed Bandits 2019 Branislav Kveton
Csaba Szepesvári
Mohammad Ghavamzadeh
Craig Boutilier
+ Multi-Armed Bandits with Censored Consumption of Resources 2020 Viktor Bengs
Eyke Hüllermeier
+ Perturbed-History Exploration in Stochastic Multi-Armed Bandits 2019 Branislav Kveton
Csaba Szepesvári
Mohammad Ghavamzadeh
Craig Boutilier
+ Best Arm Identification with Fixed Budget: A Large Deviation Perspective 2023 Po-An Wang
Ruo-Chun Tzeng
Alexandre Proutière
+ Optimal Exploration-Exploitation in a Multi-Armed-Bandit Problem with Non-stationary Rewards 2014 Omar Besbes
Yonatan Gur
Assaf Zeevi
+ Fair Algorithms for Multi-Agent Multi-Armed Bandits 2020 Safwan Hossain
Evi Micha
Nisarg Shah
+ PDF Chat Sparse stochastic bandits 2017 Joon Kwon
Vianney Perchet
Claire Vernade
+ PDF Chat Sparse stochastic bandits 2017 Joon Kwon
Vianney Perchet
Claire Vernade
+ Multi-Armed Bandits with Non-Stationary Rewards 2017 Corinna Cortes
Giulia DeSalvo
Vitaly Kuznetsov
Mehryar Mohri
Scott Cheng‐Hsin Yang

Works That Cite This (0)

Action Title Year Authors

Works Cited by This (0)

Action Title Year Authors