Abstract

We consider the classic stochastic multiarmed bandit problem with a constraint that limits the total cost incurred by switching between actions to be no larger than a given switching budget. For this problem, we prove matching upper and lower bounds on the optimal (i.e., minimax) regret and provide efficient rate-optimal algorithms. Surprisingly, the optimal regret of this problem exhibits a nonconventional growth rate in terms of the time horizon and the number of arms. Consequently, we discover surprising “phase transitions” regarding how the optimal regret rate changes with respect to the switching budget: when the number of arms is fixed, there are equal-length phases, in which the optimal regret rate remains (almost) the same within each phase and exhibits abrupt changes between phases; when the number of arms grows with the time horizon, such abrupt changes become subtler and may disappear, but a generalized notion of phase transitions involving certain new measurements still exists. The results enable us to fully characterize the trade-off between the regret rate and the incurred switching cost in the stochastic multiarmed bandit problem, contributing new insights to this fundamental problem. Under the general switching cost structure, the results reveal interesting connections between bandit problems and graph traversal problems, such as the shortest Hamiltonian path problem. This paper was accepted by J. George Shanthikumar, data science. Supplemental Material: The online appendix and data are available at https://doi.org/10.1287/mnsc.2023.4755 .

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call