Sciweavers

WSDM
2016
ACM

Multileave Gradient Descent for Fast Online Learning to Rank

8 years 7 months ago
Multileave Gradient Descent for Fast Online Learning to Rank
Modern search systems are based on dozens or even hundreds of ranking features. The dueling bandit gradient descent (DBGD) algorithm has been shown to effectively learn combinations of these features solely from user interactions. DBGD explores the search space by comparing a possibly improved ranker to the current production ranker. To this end, it uses interleaved comparison methods, which can infer with high sensitivity a preference between two rankings based only on interaction data. A limiting factor is that it can compare only to a single exploratory ranker. We propose an online learning to rank algorithm called multileave gradient descent (MGD) that extends DBGD to learn from so-called multileaved comparison methods that can compare a set of rankings instead of merely a pair. We show experimentally that MGD allows for better selection of candidates than DBGD without the need for more comparisons involving users. An important implication of our results is that orders of magnitud...
Anne Schuth, Harrie Oosterhuis, Shimon Whiteson, M
Added 12 Apr 2016
Updated 12 Apr 2016
Type Journal
Year 2016
Where WSDM
Authors Anne Schuth, Harrie Oosterhuis, Shimon Whiteson, Maarten de Rijke
Comments (0)