Abstract
The following coins problem is a version of a multi-armed bandit problem where one has to select from among a set of objects, say classifiers, after an experimentation phase that is constrained by a time or cost budget. The question is how to spend the budget. The problem involves pure exploration only, differentiating it from typical multi-armed bandit problems involving an exploration/exploitation tradeoff [BF85]. It is an abstraction of the following scenarios: choosing from among a set of alternative treatments after a fixed number of clinical trials, determining the best parameter settings for a program given a deadline that only allows a fixed number of runs; or choosing a life partner in the bachelor/bachelorette TV show where time is limited. We are interested in the computational complexity of the coins problem and/or efficient algorithms with approximation guarantees.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Berry, D., Fristedt, B.: Bandit Problems: Sequential Allocation of Experiments. Chapman and Hall, NewYork (1985)
Lizotte, D., Madani, O., Greiner, R.: Budgeted learning of Naive Bayes classifiers. In: UAI 2003 (2003)
Madani, O., Lizotte, D., Greiner, R.: Active model selection (submitted). Technical report, University of Alberta and AICML (2004), http://www.cs.ualberta.ca/~madani/budget.html
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Madani, O., Lizotte, D.J., Greiner, R. (2004). The Budgeted Multi-armed Bandit Problem. In: Shawe-Taylor, J., Singer, Y. (eds) Learning Theory. COLT 2004. Lecture Notes in Computer Science(), vol 3120. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-27819-1_46
Download citation
DOI: https://doi.org/10.1007/978-3-540-27819-1_46
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22282-8
Online ISBN: 978-3-540-27819-1
eBook Packages: Springer Book Archive