Authors
Aleksandrs Slivkins, Filip Radlinski, Sreenivas Gollapudi
Publication date
2013
Journal
Journal of Machine Learning Research
Volume
14
Issue
Feb
Pages
399-436
Publisher
Conference version: ICML 2010 (Intl. Conf. on Machine Learning)
Description
Most learning to rank research has assumed that the utility of different documents is independent, which results in learned ranking functions that return redundant results. The few approaches that avoid this have rather unsatisfyingly lacked theoretical foundations, or do not scale. We present a learning-to-rank formulation that optimizes the fraction of satisfied users, with several scalable algorithms that explicitly takes document similarity and ranking context into account. Our formulation is a non-trivial common generalization of two multi-armed bandit models from the literature: ranked bandits (Radlinski et al., 2008) and Lipschitz bandits (Kleinberg et al., 2008b). We present theoretical justifications for this approach, as well as a near-optimal algorithm. Our evaluation adds optimizations that improve empirical performance, and shows that our algorithms learn orders of magnitude more quickly than previous approaches.
Total citations
201120122013201420152016201720182019202020212022202320241179141991219171481046
Scholar articles
A Slivkins, F Radlinski, S Gollapudi - The Journal of Machine Learning Research, 2013
A Slivkins, F Radlinski, S Gollapudi - Proc. of the 27th International Conference on Machine …, 2010
F Radlinski, A Slivkins, S Gollapudi - Advances in Ranking