Authors
Aurélien Garivier, Eric Moulines
Publication date
2011/10/5
Book
International conference on algorithmic learning theory
Pages
174-188
Publisher
Springer Berlin Heidelberg
Description
Many problems, such as cognitive radio, parameter control of a scanning tunnelling microscope or internet advertisement, can be modelled as non-stationary bandit problems where the distributions of rewards changes abruptly at unknown time instants. In this paper, we analyze two algorithms designed for solving this issue: discounted UCB (D-UCB) and sliding-window UCB (SW-UCB). We establish an upper-bound for the expected regret by upper-bounding the expectation of the number of times suboptimal arms are played. The proof relies on an interesting Hoeffding type inequality for self normalized deviations with a random number of summands. We establish a lower-bound for the regret in presence of abrupt changes in the arms reward distributions. We show that the discounted UCB and the sliding-window UCB both match the lower-bound up to a logarithmic factor. Numerical simulations show that …
Total citations
2012201320142015201620172018201920202021202220232024531324163433667283829665
Scholar articles
A Garivier, E Moulines - International conference on algorithmic learning theory, 2011