A multi-armed bandit problem models an agent that simultaneously attempts to acquire new information (exploration) and optimizes the decisions based on existing knowledge (exploitation). In clinical trials, this framework applies to Bayesian multi-armed randomized adaptive designs. The allocation rule of experimental units involves the posterior probability of each treatment being the best. The trade-off between gain in information and selection of the most promising treatment is modulated by a quantity γ, typically prefixed or linearly increasing with accumulating sample size. We propose a predictive criterion for selecting γ that also allows its progressive reassessment based on interim analyses data
How humans achieve long-term goals in an uncertain environment, via repeated trials and noisy observ...
How people achieve long-term goals in an imperfectly known environment, via repeated tries and noisy...
We consider the problem of sequentially choosing between a set of unbiased Monte Carlo estimators to...
Suppose two treatments with binary responses are available for patients with some disease. Sequentia...
We study learning in a bandit task in which the outcome probabilities of six arms switch (“jump”) ov...
In experiments, researchers commonly allocate subjects randomly and equally to the different treatme...
Adaptive designs for multi-armed clinical trials have become increasingly popular recently because o...
Multi-armed bandit, a popular framework for sequential decision-making problems, has recently gained...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Background Adaptive designs offer added flexibility in the execution of clinical tri...
Bayesian experimental design (BED) is a methodology to identify designs that are expected to yield i...
Multi-armed bandit problems (MABPs) are a special type of optimal control problem that has been stud...
In field experiments, researchers commonly allocate subjects to different treatment conditions befor...
The bandit problem is a dynamic decision-making task that is simply described, well-suited to contro...
How people achieve long-term goals in an imperfectly known environment, via repeated tries and noisy...
How humans achieve long-term goals in an uncertain environment, via repeated trials and noisy observ...
How people achieve long-term goals in an imperfectly known environment, via repeated tries and noisy...
We consider the problem of sequentially choosing between a set of unbiased Monte Carlo estimators to...
Suppose two treatments with binary responses are available for patients with some disease. Sequentia...
We study learning in a bandit task in which the outcome probabilities of six arms switch (“jump”) ov...
In experiments, researchers commonly allocate subjects randomly and equally to the different treatme...
Adaptive designs for multi-armed clinical trials have become increasingly popular recently because o...
Multi-armed bandit, a popular framework for sequential decision-making problems, has recently gained...
Abstract—We present a formal model of human decision-making in explore-exploit tasks using the conte...
Background Adaptive designs offer added flexibility in the execution of clinical tri...
Bayesian experimental design (BED) is a methodology to identify designs that are expected to yield i...
Multi-armed bandit problems (MABPs) are a special type of optimal control problem that has been stud...
In field experiments, researchers commonly allocate subjects to different treatment conditions befor...
The bandit problem is a dynamic decision-making task that is simply described, well-suited to contro...
How people achieve long-term goals in an imperfectly known environment, via repeated tries and noisy...
How humans achieve long-term goals in an uncertain environment, via repeated trials and noisy observ...
How people achieve long-term goals in an imperfectly known environment, via repeated tries and noisy...
We consider the problem of sequentially choosing between a set of unbiased Monte Carlo estimators to...