Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs capture the stochasticity that may arise, for instance, from imprecise actuators via probabilities in the transition function. However, in data-driven applications, deriving precise probabilities from (limited) data introduces statistical errors that may lead to unexpected or undesirable outcomes. Uncertain MDPs (uMDPs) do not require precise probabilities but instead use so-called uncertainty sets in the transitions, accounting for such limited data. Tools from the formal verification community efficiently compute robust policies that provably adhere to formal specifications, like safety constraints, under the worst-case instance in the unc...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Reinforcement learning (RL) has become a highly successful framework for learning in Markov decision...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
We present a method for designing a robust control policy for an uncertain system subject to tempora...
Markov decision processes (MDPs) are a common approach to model dynamic optimization problems in man...
We present a method for designing robust controllers for dynamical systems with linear temporal logi...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
Thesis (Ph.D.)--University of Washington, 2018Markov decision processes (MDPs) model a class of stoc...
Markov Decision Processes (MDPs) constitute a mathematical framework for modelling systems featuring...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Reinforcement learning (RL) has become a highly successful framework for learning in Markov decision...
Markov decision processes (MDPs) are formal models commonly used in sequential decision-making. MDPs...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
We present a method for designing a robust control policy for an uncertain system subject to tempora...
Markov decision processes (MDPs) are a common approach to model dynamic optimization problems in man...
We present a method for designing robust controllers for dynamical systems with linear temporal logi...
Thesis (Ph. D.)--Massachusetts Institute of Technology, Sloan School of Management, Operations Resea...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
Thesis (Ph.D.)--University of Washington, 2018Markov decision processes (MDPs) model a class of stoc...
Markov Decision Processes (MDPs) constitute a mathematical framework for modelling systems featuring...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Reinforcement learning (RL) has become a highly successful framework for learning in Markov decision...