We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of exploration-exploitation (“bandit”) strategies. We provide a sharp regret analysis of this algorithm in a standard stochastic noise set-ting, demonstrate its scalability properties, and prove its effectiveness on a number of synthetic and real-world datasets. Our experiments show a significant increase in prediction performance over state-of-the-art methods for bandit prob-lems. 1
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of ...
Classical collaborative filtering, and content-based filtering methods try to learn a static recomme...
Multi-armed bandit problems are receiving a great deal of attention because they adequately formaliz...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
With the rapid growth in velocity and volume, streaming data compels decision support systems to pre...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
We study identifying user clusters in contextual multi-armed bandits (MAB). Contextual MAB is an eff...
We perform online interactive recommendation via a factorization-based bandit algorithm. Low-rank ma...
We study recommendation in scenarios where there's no prior information about the quality of content...
We consider a new setting of online clustering of contextual cascading bandits, an online learning p...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of ...
Classical collaborative filtering, and content-based filtering methods try to learn a static recomme...
Multi-armed bandit problems are receiving a great deal of attention because they adequately formaliz...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
With the rapid growth in velocity and volume, streaming data compels decision support systems to pre...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
We study identifying user clusters in contextual multi-armed bandits (MAB). Contextual MAB is an eff...
We perform online interactive recommendation via a factorization-based bandit algorithm. Low-rank ma...
We study recommendation in scenarios where there's no prior information about the quality of content...
We consider a new setting of online clustering of contextual cascading bandits, an online learning p...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...