We consider robust Markov Decision Processes with Borel state and action spaces, unbounded cost and finite time horizon. Our formulation leads to a Stackelberg game against nature. Under integrability, continuity and compactness assumptions we derive a robust cost iteration for a fixed policy of the decision maker and a value iteration for the robust optimization problem. Moreover, we show the existence of deterministic optimal policies for both players. This is in contrast to classical zero-sum games. In case the state space is the real line we show under some convexity assumptions that the interchange of supremum and infimum is possible with the help of Sion's minimax Theorem. Further, we consider the problem with special ambiguity sets. ...
Decision making formulated as finding a strategy that maximizes a utility function depends critic...
AbstractThis paper studies the minimizing risk problems in Markov decision processes with countable ...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
The main goal of this paper is to discuss several approaches to formulation of distributionally robu...
The authors consider the fundamental problem of finding good policies in uncertain models. It is dem...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Thesis (Ph.D.)--University of Washington, 2018Many applications in decision-making use a dynamic opt...
The authors consider the fundamental problem of nding good policies in uncertain models. It is dem...
Thesis (Ph.D.)--University of Washington, 2018Many applications in decision-making use a dynamic opt...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Decision making formulated as finding a strategy that maximizes a utility function depends critic...
AbstractThis paper studies the minimizing risk problems in Markov decision processes with countable ...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
The main goal of this paper is to discuss several approaches to formulation of distributionally robu...
The authors consider the fundamental problem of finding good policies in uncertain models. It is dem...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Thesis (Ph.D.)--University of Washington, 2018Many applications in decision-making use a dynamic opt...
The authors consider the fundamental problem of nding good policies in uncertain models. It is dem...
Thesis (Ph.D.)--University of Washington, 2018Many applications in decision-making use a dynamic opt...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Decision making formulated as finding a strategy that maximizes a utility function depends critic...
AbstractThis paper studies the minimizing risk problems in Markov decision processes with countable ...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...