We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of exploration-exploitation (“bandit”) strategies. We provide a sharp regret analysis of this algorithm in a standard stochastic noise set-ting, demonstrate its scalability properties, and prove its effectiveness on a number of artificial and real-world datasets. Our experiments show a significant increase in prediction performance over state-of-the-art methods for bandit prob-lems. 1
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of ...
Classical collaborative filtering, and content-based filtering methods try to learn a static recomme...
Multi-armed bandit problems are receiving a great deal of attention because they adequately formaliz...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
With the rapid growth in velocity and volume, streaming data compels decision support systems to pre...
We study identifying user clusters in contextual multi-armed bandits (MAB). Contextual MAB is an eff...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
We perform online interactive recommendation via a factorization-based bandit algorithm. Low-rank ma...
We consider a new setting of online clustering of contextual cascading bandits, an online learning p...
We study recommendation in scenarios where there's no prior information about the quality of content...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...
We introduce a novel algorithmic approach to content recommendation based on adaptive clustering of ...
Classical collaborative filtering, and content-based filtering methods try to learn a static recomme...
Multi-armed bandit problems are receiving a great deal of attention because they adequately formaliz...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
Multi-armed bandit problems formalize the exploration-exploitation trade-offs arising in several ind...
With the rapid growth in velocity and volume, streaming data compels decision support systems to pre...
We study identifying user clusters in contextual multi-armed bandits (MAB). Contextual MAB is an eff...
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection ...
We perform online interactive recommendation via a factorization-based bandit algorithm. Low-rank ma...
We consider a new setting of online clustering of contextual cascading bandits, an online learning p...
We study recommendation in scenarios where there's no prior information about the quality of content...
The cold-start problem has attracted extensive attention among various online services that provide ...
We study the problem of decision-making under uncertainty in the bandit setting. This thesis goes be...
International audienceThis paper addresses the on-line recommendation problem facing new users and n...
Stochastic bandit algorithms are increasingly being used in the domain of recommender systems, when ...