International audienceMultiple-play bandits aim at displaying relevant items at relevant positions on a web page. We introduce a new bandit-based algorithm, PB-MHB, for online recommender systems which uses the Thompson sampling framework with Metropolis-Hastings approximation. This algorithm handles a display setting governed by the positionbased model. Our sampling method does not require as input the probability of a user to look at a given position in the web page which is difficult to obtain in some applications. Experiments on simulated and real datasets show that our method, with fewer prior information, delivers better recommendations than state-of-the-art algorithms
Matrix factorization (MF) collaborative filtering is an effective and widely used method in recommen...
In our ongoing work we extend the Thompson Sampling (TS) bandit policy for orchestrating the collect...
We study recommendation in scenarios where there's no prior information about the quality of content...
International audienceMultiple-play bandits aim at displaying relevant items at relevant positions o...
The aim of the research presented in this dissertation is to construct a model for personalised item...
We tackle the online learning to rank problem of assigning L items to K predefined positions on a we...
International audienceWe tackle the online ranking problem of assigning L items to K positions on a ...
Multi-Armed Bandit (MAB) framework has been successfully applied in many web applications. However, ...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
In this work, we address the combinatorial optimization problem in the stochastic bandit setting wit...
Ranking system is the core part of modern retrieval and recommender systems, where the goal is to ra...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
The cold-start problem has attracted extensive attention among various online services that provide ...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
Matrix factorization (MF) collaborative filtering is an effective and widely used method in recommen...
In our ongoing work we extend the Thompson Sampling (TS) bandit policy for orchestrating the collect...
We study recommendation in scenarios where there's no prior information about the quality of content...
International audienceMultiple-play bandits aim at displaying relevant items at relevant positions o...
The aim of the research presented in this dissertation is to construct a model for personalised item...
We tackle the online learning to rank problem of assigning L items to K predefined positions on a we...
International audienceWe tackle the online ranking problem of assigning L items to K positions on a ...
Multi-Armed Bandit (MAB) framework has been successfully applied in many web applications. However, ...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
In this work, we address the combinatorial optimization problem in the stochastic bandit setting wit...
Ranking system is the core part of modern retrieval and recommender systems, where the goal is to ra...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
The cold-start problem has attracted extensive attention among various online services that provide ...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
Matrix factorization (MF) collaborative filtering is an effective and widely used method in recommen...
In our ongoing work we extend the Thompson Sampling (TS) bandit policy for orchestrating the collect...
We study recommendation in scenarios where there's no prior information about the quality of content...