Ask a Question

Prefer a chat interface with context about you and your work?

Minimizing Cost Rather Than Maximizing Reward in Restless Multi-Armed Bandits

Minimizing Cost Rather Than Maximizing Reward in Restless Multi-Armed Bandits

Restless Multi-Armed Bandits (RMABs) offer a powerful framework for solving resource constrained maximization problems. However, the formulation can be inappropriate for settings where the limiting constraint is a reward threshold rather than a budget. We introduce a constrained minimization problem for RMABs that balances the goal of achieving a reward …