A recently proposed formulation of the stochastic planning and control problem as one of parameter estimation for suitable artificial statistical models has led to the adoption of inference algorithms for this notoriously hard problem. At the algorithmic level, the focus has been on developing Expectation-Maximization (EM) algorithms. In this paper, we begin by making the crucial observation that the stochastic control problem can be reinterpreted as one of trans-dimensional inference. With this new interpretation, we are able to propose a novel reversible jump Markov chain Monte Carlo (MCMC) algorithm that is more efficient than its EM counterparts. Moreover, it enables us to implement full Bayesian policy search, without the need for grad...
a b s t r a c t We present an overview of Markov chain Monte Carlo, a sampling method for model infe...
Using the expression for the unnormalized nonlinear filter for a hidden Markov model, we develop a d...
This thesis addresses several issues appearing in Bayesian statistics. Firstly, computations for app...
Reinforcement learning methods are being applied to control problems in robotics domain. These algor...
Policy search algorithms have facilitated application of Reinforcement Learning (RL) to dynamic syst...
Markov ChainMonte Carlo (MCMC) and sequentialMonte Carlo (SMC) methods are the two most popular clas...
We consider the problem of "optimal learning" for Markov decision processes with uncertain...
Application of learning algorithms to robotics and control problems with highly nonlinear dynamics t...
This work addresses the problem of estimating the optimal value function in a Markov Decision Proces...
methods are the two most popular classes of algorithms used to sample from general high-dimensional ...
We propose a new class of learning algorithms that combines variational approximation and Markov cha...
We consider the inverse reinforcement learning problem, that is, the problem of learning from, and t...
Bayesian planning is a formally elegant approach to learning optimal behaviour under model uncertain...
Bayesian planning is a formally elegant approach to learning optimal behaviour under model uncertain...
This article considers Markov chain computational methods for incorporating uncertainty about the d...
a b s t r a c t We present an overview of Markov chain Monte Carlo, a sampling method for model infe...
Using the expression for the unnormalized nonlinear filter for a hidden Markov model, we develop a d...
This thesis addresses several issues appearing in Bayesian statistics. Firstly, computations for app...
Reinforcement learning methods are being applied to control problems in robotics domain. These algor...
Policy search algorithms have facilitated application of Reinforcement Learning (RL) to dynamic syst...
Markov ChainMonte Carlo (MCMC) and sequentialMonte Carlo (SMC) methods are the two most popular clas...
We consider the problem of "optimal learning" for Markov decision processes with uncertain...
Application of learning algorithms to robotics and control problems with highly nonlinear dynamics t...
This work addresses the problem of estimating the optimal value function in a Markov Decision Proces...
methods are the two most popular classes of algorithms used to sample from general high-dimensional ...
We propose a new class of learning algorithms that combines variational approximation and Markov cha...
We consider the inverse reinforcement learning problem, that is, the problem of learning from, and t...
Bayesian planning is a formally elegant approach to learning optimal behaviour under model uncertain...
Bayesian planning is a formally elegant approach to learning optimal behaviour under model uncertain...
This article considers Markov chain computational methods for incorporating uncertainty about the d...
a b s t r a c t We present an overview of Markov chain Monte Carlo, a sampling method for model infe...
Using the expression for the unnormalized nonlinear filter for a hidden Markov model, we develop a d...
This thesis addresses several issues appearing in Bayesian statistics. Firstly, computations for app...