Collaborative Filtering as a Multi-Armed Bandit

Frédéric Guillou 1, 2, * Romaric Gaudel 1, 2 Philippe Preux 1, 2
* Corresponding author
2 SEQUEL - Sequential Learning
Inria Lille - Nord Europe, CRIStAL - Centre de Recherche en Informatique, Signal et Automatique de Lille (CRIStAL) - UMR 9189
Abstract : Recommender Systems (RS) aim at suggesting to users one or several items in which they might have interest. Following the feedback they receive from the user, these systems have to adapt their model in order to improve future recommendations. The repetition of these steps defines the RS as a sequential process. This sequential aspect raises an exploration-exploitation dilemma, which is surprisingly rarely taken into account for RS without contextual information. In this paper we present an explore-exploit collaborative filtering RS, based on Matrix Factor-ization and Bandits algorithms. Using experiments on artificial and real datasets, we show the importance and practicability of using sequential approaches to perform recommendation. We also study the impact of the model update on both the quality and the computation time of the recommendation procedure.
Document type :
Conference papers
Complete list of metadatas
Contributor : Romaric Gaudel <>
Submitted on : Thursday, January 14, 2016 - 3:33:09 PM
Last modification on : Friday, March 22, 2019 - 1:35:16 AM
Long-term archiving on : Friday, November 11, 2016 - 6:07:44 AM


Files produced by the author(s)


  • HAL Id : hal-01256254, version 1


Frédéric Guillou, Romaric Gaudel, Philippe Preux. Collaborative Filtering as a Multi-Armed Bandit. NIPS'15 Workshop: Machine Learning for eCommerce, Dec 2015, Montréal, Canada. ⟨hal-01256254⟩



Record views


Files downloads