How can we take advantage of opportunities for experimental parallelization in exploration-exploitation tradeoffs? In many experimental scenarios, it is often desirable to execute experiments simultaneously or in batches, rather than only performing one at a time. Additionally, observations may be both noisy and expensive. We introduce Gaussian Process Batch Upper Confidence Bound (GP-BUCB), an upper confidence bound-based algorithm, which models the reward function as a sample from a Gaussian process and which can select batches of experiments to run in parallel. We prove a general regret bound for GP-BUCB, as well as the surprising result that for some common kernels, the asymptotic average regret can be made independent of the batch size...
International audienceBoltzmann exploration is a classic strategy for sequential decision-making und...
Bandit algorithms are concerned with trading exploration with exploitation where a number of options...
International audienceWe consider a generalization of stochastic bandits where the set of arms, $\cX...
How can we take advantage of opportunities for experimental parallelization in exploration-exploitat...
Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We fo...
Gaussian processes (GP) are one of the most successful frameworks to model uncertainty. However, GP ...
Abstract. In this paper, we consider the challenge of maximizing an unknown function f for which eva...
Many applications require optimizing an un-known, noisy function that is expensive to evaluate. We f...
International audienceBandit algorithms are concerned with trading exploration with exploitation whe...
Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We fo...
International audienceGaussian processes (GP) are a stochastic processes, used as Bayesian approach ...
Kernel-based bandit is an extensively studied black-box optimization problem, in which the objective...
International audienceThis work addresses the problem of regret minimization in non-stochastic multi...
We consider sequential decision problems under uncertainty, where we seek to optimize an unknown fun...
Deliverable no. 2.1.1-BThe sequential sampling strategies based on Gaussian processes are widely use...
International audienceBoltzmann exploration is a classic strategy for sequential decision-making und...
Bandit algorithms are concerned with trading exploration with exploitation where a number of options...
International audienceWe consider a generalization of stochastic bandits where the set of arms, $\cX...
How can we take advantage of opportunities for experimental parallelization in exploration-exploitat...
Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We fo...
Gaussian processes (GP) are one of the most successful frameworks to model uncertainty. However, GP ...
Abstract. In this paper, we consider the challenge of maximizing an unknown function f for which eva...
Many applications require optimizing an un-known, noisy function that is expensive to evaluate. We f...
International audienceBandit algorithms are concerned with trading exploration with exploitation whe...
Many applications require optimizing an unknown, noisy function that is expensive to evaluate. We fo...
International audienceGaussian processes (GP) are a stochastic processes, used as Bayesian approach ...
Kernel-based bandit is an extensively studied black-box optimization problem, in which the objective...
International audienceThis work addresses the problem of regret minimization in non-stochastic multi...
We consider sequential decision problems under uncertainty, where we seek to optimize an unknown fun...
Deliverable no. 2.1.1-BThe sequential sampling strategies based on Gaussian processes are widely use...
International audienceBoltzmann exploration is a classic strategy for sequential decision-making und...
Bandit algorithms are concerned with trading exploration with exploitation where a number of options...
International audienceWe consider a generalization of stochastic bandits where the set of arms, $\cX...