Header menu link for other important links
X
Ballooning multi-armed bandits
, S. Dhamal, S. Jain, S. Gujar, Y. Narahari
Published in International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS)
2020
Volume: 2020-May
   
Pages: 1849 - 1851
Abstract
We introduce ballooning multi-armed bandits (BL-MAB), a novel extension to the classical stochastic MAB model. In the BL-MAB model, the set of available arms grows (or balloons) over time. The regret in a BL-MAB setting is computed with respect to the best available arm at each time. We first observe that the existing stochastic MAB algorithms are not regret-optimal for the BL-MAB model. We show that if the best arm is equally likely to arrive at any time, a sub-linear regret cannot be achieved, irrespective of the arrival of the other arms. We further show that if the best arm is more likely to arrive in the early rounds, one can achieve sub-linear regret. Making reasonable assumptions on the arrival distribution of the best arm in terms of the thinness of the distribution's tail, we prove that the proposed algorithm achieves sub-linear instance-independent regret. We further quantify explicit dependence of regret on the arrival distribution parameters. © 2020 International Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS). All rights reserved.
About the journal
JournalProceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, AAMAS
PublisherInternational Foundation for Autonomous Agents and Multiagent Systems (IFAAMAS)
ISSN15488403