Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environments. However, the solutions of MDPs are of limited practical use because of their sensitivity to distributional model parameters, which are typically unknown and have to be estimated by the decision maker. To counter the detrimental effects of estimation errors, we consider robust MDPs that offer probabilistic guarantees in view of the unknown parameters. To this end, we assume that an observation history of the MDP is available. Based on this history, we derive a confidence region that contains the unknown parameters with a prespecified probability 1-β. Afterward, we determine a policy that attains the highest worst-case performance over t...
In recent years, robust Markov decision processes (MDPs) have emerged as a prominent modeling framew...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
Markov decision process (MDP) is a decision making framework where a decision maker is interested in...
Markov Decision Processes (MDPs) are a popular class of models suitable for solving control decision...
Markov decision processes (MDPs) are a common approach to model dynamic optimization problems in man...
Abstract — This paper presents a new robust decision-making algorithm that accounts for model uncert...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
In recent years, robust Markov decision processes (MDPs) have emerged as a prominent modeling framew...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
Markov decision process (MDP) is a decision making framework where a decision maker is interested in...
Markov Decision Processes (MDPs) are a popular class of models suitable for solving control decision...
Markov decision processes (MDPs) are a common approach to model dynamic optimization problems in man...
Abstract — This paper presents a new robust decision-making algorithm that accounts for model uncert...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
In recent years, robust Markov decision processes (MDPs) have emerged as a prominent modeling framew...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...