This paper considers sequential decision making problems under uncertainty, the tradeoff between the expected return and the risk of high loss, and methods that use dynamic programming to find optimal policies. It is argued that using Bellman's Principle determines how risk considerations on the return can be incorporated. The discussion centers around returns generated by Markov Decision Processes and conclusions concern a large class of methods in Reinforcement Learning
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Decision theory addresses the task of choosing an action; it provides robust decision-making criteri...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Stochastic sequential decision-making problems are generally modeled and solved as Markov decision p...
In this paper, we give a brief review of Markov Decision Processes (MDPs), and how Reinforcement Lea...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
In this paper, we consider Markov Decision Processes (MDPs) with error states. Error states are thos...
This work addresses the problem of inverse reinforcement learning in Markov decision processes where...
Replicating the human ability to solve complex planning problems based on minimal prior knowledge ha...
This thesis dives into the theory of discrete time stochastic optimal control through exploring dyna...
We consider the dilemma of taking sequential action within a nebulous and costly stochastic system. ...
We develop a framework for risk-sensitive behaviour in reinforcement learning (RL) due to uncertaint...
What are the functionals of the reward that can be computed and optimized exactly in Markov Decision...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
peer reviewedClassical reinforcement learning (RL) techniques are generally concerned with the desig...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Decision theory addresses the task of choosing an action; it provides robust decision-making criteri...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Stochastic sequential decision-making problems are generally modeled and solved as Markov decision p...
In this paper, we give a brief review of Markov Decision Processes (MDPs), and how Reinforcement Lea...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
In this paper, we consider Markov Decision Processes (MDPs) with error states. Error states are thos...
This work addresses the problem of inverse reinforcement learning in Markov decision processes where...
Replicating the human ability to solve complex planning problems based on minimal prior knowledge ha...
This thesis dives into the theory of discrete time stochastic optimal control through exploring dyna...
We consider the dilemma of taking sequential action within a nebulous and costly stochastic system. ...
We develop a framework for risk-sensitive behaviour in reinforcement learning (RL) due to uncertaint...
What are the functionals of the reward that can be computed and optimized exactly in Markov Decision...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
peer reviewedClassical reinforcement learning (RL) techniques are generally concerned with the desig...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Decision theory addresses the task of choosing an action; it provides robust decision-making criteri...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...