Abstract. To follow the dynamicity of the user’s content, researchers have recently started to model interactions between users and the Context-Aware Recommender Systems (CARS) as a bandit problem where the system needs to deal with exploration and exploitation dilemma. In this sense, we propose to study the freshness of the user’s content in CARS through the bandit problem. We introduce in this paper an algo-rithm named Freshness-Aware Thompson Sampling (FA-TS) that man-ages the recommendation of fresh document according to the user’s risk of the situation. The intensive evaluation and the detailed analysis of the experimental results reveals several important discoveries in the explo-ration/exploitation (exr/exp) behaviour
Current recommender systems need to recommend items that are relevant to users (exploitation), but t...
In multi-objective reinforcement learning (MORL), much attention is paid to generating optimal solut...
Graduate School of Artificial Intelligence ArtificiMulti-armed bandit is a well-formulated test bed ...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
The aim of the research presented in this dissertation is to construct a model for personalised item...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
Abstract. Mobile Context-Aware Recommender Systems can be natu-rally modelled as an exploration/expl...
In our ongoing work we extend the Thompson Sampling (TS) bandit policy for orchestrating the collect...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study recommendation in scenarios where there's no prior information about the quality of content...
In this work, we address the combinatorial optimization problem in the stochastic bandit setting wit...
In sequential decision problems in an unknown environment, the decision maker often faces a dilemma ...
A challenging aspect of the bandit problem is that a stochastic reward is observed only for the chos...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
Current recommender systems need to recommend items that are relevant to users (exploitation), but t...
In multi-objective reinforcement learning (MORL), much attention is paid to generating optimal solut...
Graduate School of Artificial Intelligence ArtificiMulti-armed bandit is a well-formulated test bed ...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
The aim of the research presented in this dissertation is to construct a model for personalised item...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
Abstract. Mobile Context-Aware Recommender Systems can be natu-rally modelled as an exploration/expl...
In our ongoing work we extend the Thompson Sampling (TS) bandit policy for orchestrating the collect...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study recommendation in scenarios where there's no prior information about the quality of content...
In this work, we address the combinatorial optimization problem in the stochastic bandit setting wit...
In sequential decision problems in an unknown environment, the decision maker often faces a dilemma ...
A challenging aspect of the bandit problem is that a stochastic reward is observed only for the chos...
Multi-armed bandit (MAB) problem is derived from slot machines in the casino. It is about how a gamb...
Current recommender systems need to recommend items that are relevant to users (exploitation), but t...
In multi-objective reinforcement learning (MORL), much attention is paid to generating optimal solut...
Graduate School of Artificial Intelligence ArtificiMulti-armed bandit is a well-formulated test bed ...