Bandits and Markov Decision Processes are powerful sequential decision making paradigms that have been widely applied to solve real world problems. However, existing algorithms often suffer from high sample complexity due to the large action space. In this thesis, we present several contributions to reduce the sample complexity by exploiting the problem structure. In the first part, we study how to utilize the given causal information represented as a causal graph along with associated conditional distributions for bandit problems. We propose two algorithms, causal upper confidence bound (C-UCB) and causal Thompson Sampling (C-TS), that enjoy improved cumulative regret bounds compared with algorithms that do not use causal information. Furt...
Causal inference provides a set of principles and tools that allows one to combine data and knowledg...
I consider the type of statistical experiment commonly referred to as adaptive trials, in which the ...
The field of causal learning has grown in the past decade, establishing itself as a major focus in a...
Bandits and Markov Decision Processes are powerful sequential decision making paradigms that have be...
Modern learning systems like recommendation engines, computational advertising systems, online param...
We study the problem of using causal models to improve the rate at which good interventions can be l...
Sequentially making-decision abounds in real-world problems ranging from robots needing to interact ...
In combinatorial causal bandits (CCB), the learning agent chooses at most K variables in each round ...
This thesis investigates sequential decision making tasks that fall in the framework of reinforcemen...
We study the problem of determining the best intervention in a Causal Bayesian Network (CBN) specifi...
Purpose Sampling an action according to the probability that the action is believed to be the optima...
The Markov assumption (MA) is fundamental to the empirical validity of reinforcement learning. In th...
Inferring the causal structure that links n observables is usually based upon detecting statistical ...
We study learning algorithms for the classical Markovian bandit problem with discount. We explain ho...
In this work we consider probabilistic approaches to sequential decision making. The ultimate goal i...
Causal inference provides a set of principles and tools that allows one to combine data and knowledg...
I consider the type of statistical experiment commonly referred to as adaptive trials, in which the ...
The field of causal learning has grown in the past decade, establishing itself as a major focus in a...
Bandits and Markov Decision Processes are powerful sequential decision making paradigms that have be...
Modern learning systems like recommendation engines, computational advertising systems, online param...
We study the problem of using causal models to improve the rate at which good interventions can be l...
Sequentially making-decision abounds in real-world problems ranging from robots needing to interact ...
In combinatorial causal bandits (CCB), the learning agent chooses at most K variables in each round ...
This thesis investigates sequential decision making tasks that fall in the framework of reinforcemen...
We study the problem of determining the best intervention in a Causal Bayesian Network (CBN) specifi...
Purpose Sampling an action according to the probability that the action is believed to be the optima...
The Markov assumption (MA) is fundamental to the empirical validity of reinforcement learning. In th...
Inferring the causal structure that links n observables is usually based upon detecting statistical ...
We study learning algorithms for the classical Markovian bandit problem with discount. We explain ho...
In this work we consider probabilistic approaches to sequential decision making. The ultimate goal i...
Causal inference provides a set of principles and tools that allows one to combine data and knowledg...
I consider the type of statistical experiment commonly referred to as adaptive trials, in which the ...
The field of causal learning has grown in the past decade, establishing itself as a major focus in a...