We study a multi-armed bandit problem in a setting where covariates are available. We take a nonparametric approach to estimate the functional relationship between the response (reward) and the covariates. The estimated relationships and appropriate randomization are used to select a good arm to play for a greater expected reward. Randomization helps balance the tendency to trust the currently most promising arm with further exploration of other arms. It is shown that, with some familiar nonparametric methods (e.g., histogram), the proposed strategy is strongly consistent in the sense that the accumulated reward is asymptotically equivalent to that based on the best arm (which depends on the covariates) almost surely
We consider a multiarmed bandit problem where the expected reward of each arm is a linear function o...
We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realiza...
Published version of an article from Lecture Notes in Computer Science. Also available at SpringerLi...
We consider a bandit problem which involves sequential sampling from two populations (arms). Each ar...
Multi-armed bandit problem is an important optimization game that requires an exploration-exploitati...
University of Minnesota Ph.D. dissertation. July 2014. Major: Statistics. Advisor: Yuhong Yang. 1 co...
Consider a Bayesian sequential allocation problem that incorporates a covariate. The goal is to maxi...
In this paper, we propose a set of allocation strategies to deal with the multi-armed bandit problem...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
This paper studies an important sequential decision making problem known as the multi-armed stochast...
Different allocation strategies can be found in the literature to deal with the multi-armed bandit p...
International audienceWe consider the problem of best arm identification in the multi-armed bandit m...
A Multi-Armed Bandits (MAB) is a learning problem where an agent sequentially chooses an action amon...
International audienceWe consider the problem of finding the best arm in a stochastic multi-armed ba...
In this thesis, we study strategies for sequential resource allocation, under the so-called stochast...
We consider a multiarmed bandit problem where the expected reward of each arm is a linear function o...
We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realiza...
Published version of an article from Lecture Notes in Computer Science. Also available at SpringerLi...
We consider a bandit problem which involves sequential sampling from two populations (arms). Each ar...
Multi-armed bandit problem is an important optimization game that requires an exploration-exploitati...
University of Minnesota Ph.D. dissertation. July 2014. Major: Statistics. Advisor: Yuhong Yang. 1 co...
Consider a Bayesian sequential allocation problem that incorporates a covariate. The goal is to maxi...
In this paper, we propose a set of allocation strategies to deal with the multi-armed bandit problem...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
This paper studies an important sequential decision making problem known as the multi-armed stochast...
Different allocation strategies can be found in the literature to deal with the multi-armed bandit p...
International audienceWe consider the problem of best arm identification in the multi-armed bandit m...
A Multi-Armed Bandits (MAB) is a learning problem where an agent sequentially chooses an action amon...
International audienceWe consider the problem of finding the best arm in a stochastic multi-armed ba...
In this thesis, we study strategies for sequential resource allocation, under the so-called stochast...
We consider a multiarmed bandit problem where the expected reward of each arm is a linear function o...
We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realiza...
Published version of an article from Lecture Notes in Computer Science. Also available at SpringerLi...