When comparing the performance of multi-armed bandit algorithms, the potential impact of missing data is often overlooked. In practice, it also affects their implementation where the simplest approach to overcome this is to continue to sample according to the original bandit algorithm, ignoring missing outcomes. We investigate the impact on performance of this approach to deal with missing data for several bandit algorithms through an extensive simulation study assuming the rewards are missing at random. We focus on two-armed bandit algorithms with binary outcomes in the context of patient allocation for clinical trials with relatively small sample sizes. However, our results apply to other applications of bandit algorithms where missing da...
International audienceAlgorithms based on upper confidence bounds for balancing exploration and expl...
The performance evaluation of imputation algorithms often involves the generation of missing values...
This thesis considers the multi-armed bandit (MAB) problem, both the traditional bandit feedback and...
The impact of missing data on ONS varies among different algorithms. Generally, bandit algorithms ou...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
The multi-armed bandit (MAB) problem refers to the task of sequentially assigning treatments to expe...
We explore a new model of bandit experiments where a potentially nonstationary sequence of contexts ...
Single-case experiments have become increasingly popular in psychological and educational research. ...
In real-life situations, we often encounter data sets containing missing observations. Statistical m...
Currently, a growing number of programs become available in statistical software for multiple imputa...
The multi-armed bandit (MAB) problem provides a convenient abstraction for many online decision prob...
An empirical comparative study is made of a sample of action selection policies on a test suite of t...
Multi-armed bandit, a popular framework for sequential decision-making problems, has recently gained...
Multi-armed bandit problems (MABPs) are a special type of optimal control problem that has been stud...
International audienceWe consider the problem of finding the best arm in a stochastic multi-armed ba...
International audienceAlgorithms based on upper confidence bounds for balancing exploration and expl...
The performance evaluation of imputation algorithms often involves the generation of missing values...
This thesis considers the multi-armed bandit (MAB) problem, both the traditional bandit feedback and...
The impact of missing data on ONS varies among different algorithms. Generally, bandit algorithms ou...
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploit...
The multi-armed bandit (MAB) problem refers to the task of sequentially assigning treatments to expe...
We explore a new model of bandit experiments where a potentially nonstationary sequence of contexts ...
Single-case experiments have become increasingly popular in psychological and educational research. ...
In real-life situations, we often encounter data sets containing missing observations. Statistical m...
Currently, a growing number of programs become available in statistical software for multiple imputa...
The multi-armed bandit (MAB) problem provides a convenient abstraction for many online decision prob...
An empirical comparative study is made of a sample of action selection policies on a test suite of t...
Multi-armed bandit, a popular framework for sequential decision-making problems, has recently gained...
Multi-armed bandit problems (MABPs) are a special type of optimal control problem that has been stud...
International audienceWe consider the problem of finding the best arm in a stochastic multi-armed ba...
International audienceAlgorithms based on upper confidence bounds for balancing exploration and expl...
The performance evaluation of imputation algorithms often involves the generation of missing values...
This thesis considers the multi-armed bandit (MAB) problem, both the traditional bandit feedback and...