We introduce a class of Markov decision problems (MDPs) which greatly simplify Reinforcement Learning. These MDPs have discrete state spaces and continuous control spaces. The controls have the effect of scaling the transition probabilities of an underlying Markov chain. A control cost penalizing KL divergence between controlled and uncontrolled transition probabilities makes the minimization problem convex and allows analytical computation of the optimal controls given the optimal value function. An exponential transformation of the optimal value function makes the minimized Bellman equation linear. Apart from their theoretical signi cance, the new MDPs enable ef cient approximations to traditional MDPs. Shortest path problems are approxim...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
This study is concerned with finite Markov decision processes (MDPs) whose state are exactly observa...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
Increasing attention has been paid to reinforcement learning algorithms in recent years, partly due ...
University of Minnesota M.S. thesis. June 2012. Major: Computer science. Advisor: Prof. Paul Schrate...
We present a hierarchical reinforcement learning framework that formulates each task in the hierarch...
Semi-Markov Decision Problems are continuous time generalizations of discrete time Markov Decision P...
We propose a novel approach for solving continuous and hybrid Markov Decision Processes (MDPs) based...
Solving Markov decision processes (MDPs) efficiently is challenging in many cases, for example, when...
A large class of problems of sequential decision making under uncertainty, of which the underlying p...
www.cs.tu-berlin.de\∼geibel Abstract. In this article, I will consider Markov Decision Processes wit...
We present a framework to address a class of sequential decision making problems. Our framework feat...
In this paper, we study the problem of transferring the available Markov Decision Process (MDP) mode...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
This study is concerned with finite Markov decision processes (MDPs) whose state are exactly observa...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
We introduce a class of MPDs which greatly simplify Reinforcement Learning. They have discrete state...
Increasing attention has been paid to reinforcement learning algorithms in recent years, partly due ...
University of Minnesota M.S. thesis. June 2012. Major: Computer science. Advisor: Prof. Paul Schrate...
We present a hierarchical reinforcement learning framework that formulates each task in the hierarch...
Semi-Markov Decision Problems are continuous time generalizations of discrete time Markov Decision P...
We propose a novel approach for solving continuous and hybrid Markov Decision Processes (MDPs) based...
Solving Markov decision processes (MDPs) efficiently is challenging in many cases, for example, when...
A large class of problems of sequential decision making under uncertainty, of which the underlying p...
www.cs.tu-berlin.de\∼geibel Abstract. In this article, I will consider Markov Decision Processes wit...
We present a framework to address a class of sequential decision making problems. Our framework feat...
In this paper, we study the problem of transferring the available Markov Decision Process (MDP) mode...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
This study is concerned with finite Markov decision processes (MDPs) whose state are exactly observa...
Address email We present an approximation scheme for solving Markov Decision Processes (MDPs) in whi...