Minimizing Cost Rather Than Maximizing Reward in Restless Multi-Armed
Bandits
Minimizing Cost Rather Than Maximizing Reward in Restless Multi-Armed
Bandits
Restless Multi-Armed Bandits (RMABs) offer a powerful framework for solving resource constrained maximization problems. However, the formulation can be inappropriate for settings where the limiting constraint is a reward threshold rather than a budget. We introduce a constrained minimization problem for RMABs that balances the goal of achieving a reward …