The probabilistic ranking principle advocates ranking documents in order of de-creasing probability of relevance to a query, independent of how other documents are ranked. The result is that similar documents are often ranked at similar po-sitions. In contrast, empirical studies have shown that a diverse set of results is often preferable over one containing redundant results, as typical web queries of-ten have different meanings for different users (such as jaguar). We present a new multi-armed bandit learning algorithm that directly learns a diverse ranking of results based on users ’ clicking behavior. In particular, it maximizes the proba-bility that a relevant document is found in the top k positions of a ranking. After T presentations...
International audienceWe tackle, in the multiple-play bandit setting, the online ranking problem of ...
We study the problem of online rank elicitation, assuming that rankings of a set of alternatives obe...
We tackle the online learning to rank problem of assigning L items to K predefined positions on a we...
Multi-Armed Bandit (MAB) framework has been successfully applied in many web applications. However, ...
International audienceAlgorithms for learning to rank Web documents, display ads, or other types of ...
Ranking system is the core part of modern retrieval and recommender systems, where the goal is to ra...
International audienceIn subset ranking, the goal is to learn a ranking function that approximates a...
International audienceWe tackle the online ranking problem of assigning L items to K positions on a ...
Given a repeatedly issued query and a document with a not-yet-confirmed potential to satisfy the use...
Identifying the topic of a search query is a challenging problem, for which a solution would be valu...
Abstract. We propose an approach for discovering in an automatic way formulas for ranking arms while...
We study a ranking problem in the contextual multi-armed bandit setting. A learning agent selects an...
In this paper, as a novel approach, we learn Markov chain transition probabilities for ranking of mu...
International audienceWe tackle, in the multiple-play bandit setting, the online ranking problem of ...
We study the problem of online rank elicitation, assuming that rankings of a set of alternatives obe...
We tackle the online learning to rank problem of assigning L items to K predefined positions on a we...
Multi-Armed Bandit (MAB) framework has been successfully applied in many web applications. However, ...
International audienceAlgorithms for learning to rank Web documents, display ads, or other types of ...
Ranking system is the core part of modern retrieval and recommender systems, where the goal is to ra...
International audienceIn subset ranking, the goal is to learn a ranking function that approximates a...
International audienceWe tackle the online ranking problem of assigning L items to K positions on a ...
Given a repeatedly issued query and a document with a not-yet-confirmed potential to satisfy the use...
Identifying the topic of a search query is a challenging problem, for which a solution would be valu...
Abstract. We propose an approach for discovering in an automatic way formulas for ranking arms while...
We study a ranking problem in the contextual multi-armed bandit setting. A learning agent selects an...
In this paper, as a novel approach, we learn Markov chain transition probabilities for ranking of mu...
International audienceWe tackle, in the multiple-play bandit setting, the online ranking problem of ...
We study the problem of online rank elicitation, assuming that rankings of a set of alternatives obe...
We tackle the online learning to rank problem of assigning L items to K predefined positions on a we...