Authors
Joannes Vermorel, Mehryar Mohri
Publication date
2005/10/3
Book
European conference on machine learning
Pages
437-448
Publisher
Springer Berlin Heidelberg
Description
The multi-armed bandit problem for a gambler is to decide which arm of a K-slot machine to pull to maximize his total reward in a series of trials. Many real-world learning and optimization problems can be modeled in this way. Several strategies or algorithms have been proposed as a solution to this problem in the last two decades, but, to our knowledge, there has been no common evaluation of these algorithms.
This paper provides a preliminary empirical evaluation of several multi-armed bandit algorithms. It also describes and analyzes a new algorithm, Poker (Price Of Knowledge and Estimated Reward) whose performance compares favorably to that of other existing algorithms in several experiments. One remarkable outcome of our experiments is that the most naive approach, the ε-greedy strategy, proves to be often hard to beat.
Total citations
20062007200820092010201120122013201420152016201720182019202020212022202320243111518241844364749474054696293597130
Scholar articles
J Vermorel, M Mohri - European conference on machine learning, 2005