We study the multi-armed bandit problem with multiple plays and a budget constraint for both the stochastic and the adversarial setting. At each round, exactly K out of N possible arms have to be played (with 1 ≤ K <= N). In addition to observing the individual rewards for each arm played, the player also learns a vector of costs which has to be covered with an a-priori defined budget B. The game ends when the sum of current costs associated with the played arms exceeds the remaining budget. Firstly, we analyze this setting for the stochastic case, for which we assume each arm to have an underlying cost and reward distribution with support [cmin, 1] and [0, 1], respectively. We derive an Upper Confidence Bound (UCB) algorithm which achie...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
This paper investigates stochastic and adversarial combinatorial multi-armed bandit problems. In the...
Abstract — In this paper, we consider the problem of multi-armed bandits with a large, possibly infi...
International audienceWe study a generalization of the multi-armed bandit problem with multiple p...
We study the infinitely many-armed bandit problem with budget constraints, where the number of arms ...
In a multi-armed bandit (MAB) problem a gambler needs to choose at each round of play one of K arms,...
We consider multi-armed bandit problems where the number of arms is larger than the possible number ...
International audienceWe consider multi-armed bandit problems where the number of arms is larger tha...
We study the multi-armed bandit problems with budget constraint and variable costs (MAB-BV). In this...
We study a two-player stochastic multi-armed bandit (MAB) problem with different expected rewards fo...
We introduce the budget-limited multi-armed bandit (MAB), which captures situations where a learners...
In the multi-armed bandit problem, a gambler must decide which arm of non-identical slot machines t...
We introduce the budget–limited multi–armed bandit (MAB), which captures situations where a learner’...
We consider a multiarmed bandit problem where the expected reward of each arm is a linear function o...
We consider the stochastic multi-armed bandit (MAB) problem in a setting where a player can pay to p...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
This paper investigates stochastic and adversarial combinatorial multi-armed bandit problems. In the...
Abstract — In this paper, we consider the problem of multi-armed bandits with a large, possibly infi...
International audienceWe study a generalization of the multi-armed bandit problem with multiple p...
We study the infinitely many-armed bandit problem with budget constraints, where the number of arms ...
In a multi-armed bandit (MAB) problem a gambler needs to choose at each round of play one of K arms,...
We consider multi-armed bandit problems where the number of arms is larger than the possible number ...
International audienceWe consider multi-armed bandit problems where the number of arms is larger tha...
We study the multi-armed bandit problems with budget constraint and variable costs (MAB-BV). In this...
We study a two-player stochastic multi-armed bandit (MAB) problem with different expected rewards fo...
We introduce the budget-limited multi-armed bandit (MAB), which captures situations where a learners...
In the multi-armed bandit problem, a gambler must decide which arm of non-identical slot machines t...
We introduce the budget–limited multi–armed bandit (MAB), which captures situations where a learner’...
We consider a multiarmed bandit problem where the expected reward of each arm is a linear function o...
We consider the stochastic multi-armed bandit (MAB) problem in a setting where a player can pay to p...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
This paper investigates stochastic and adversarial combinatorial multi-armed bandit problems. In the...
Abstract — In this paper, we consider the problem of multi-armed bandits with a large, possibly infi...