In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by finding a policy that optimizes the worst-case performance over an uncertainty set of MDPs. While much of the literature has focused on discounted MDPs, robust average-reward MDPs remain largely unexplored. In this paper, we focus on robust average-reward MDPs, where the goal is to find a policy that optimizes the worst-case average reward over an uncertainty set. We first take an approach that approximates average-reward MDPs using discounted MDPs. We prove that the robust discounted value function converges to the robust average-reward as the discount factor goes to 1, and moreover when it is large, any optimal policy of the robust discoun...
© 2017 AI Access Foundation. All rights reserved. Markov Decision Processes (MDPs) are an effective ...
We consider robust Markov Decision Processes with Borel state and action spaces, unbounded cost and ...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Bounded parameter Markov Decision Processes (BMDPs) address the issue of dealing with uncertainty in...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
In this paper, approximate dynamic programming (ADP) problems are modeled by discounted infinite-hor...
Robust Markov decision processes (MDPs) provide a general framework to model decision problems where...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
The precise specification of reward functions for Markov decision processes (MDPs) is often extremel...
Computation of a satisfactory control policy for a Markov decision process when the parameters of th...
We study the synthesis of robust optimal control policies for Markov decision processes with transit...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
We present an algorithm called Optimistic Linear Programming (OLP) for learning to optimize average ...
© 2017 AI Access Foundation. All rights reserved. Markov Decision Processes (MDPs) are an effective ...
We consider robust Markov Decision Processes with Borel state and action spaces, unbounded cost and ...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...
Bounded parameter Markov Decision Processes (BMDPs) address the issue of dealing with uncertainty in...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
In this paper, approximate dynamic programming (ADP) problems are modeled by discounted infinite-hor...
Robust Markov decision processes (MDPs) provide a general framework to model decision problems where...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
The precise specification of reward functions for Markov decision processes (MDPs) is often extremel...
Computation of a satisfactory control policy for a Markov decision process when the parameters of th...
We study the synthesis of robust optimal control policies for Markov decision processes with transit...
We consider Markov decision processes where the values of the parameters are uncertain. This uncerta...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
Optimal solutions to Markov Decision Problems (MDPs) are very sensitive with respect to the state tr...
We present an algorithm called Optimistic Linear Programming (OLP) for learning to optimize average ...
© 2017 AI Access Foundation. All rights reserved. Markov Decision Processes (MDPs) are an effective ...
We consider robust Markov Decision Processes with Borel state and action spaces, unbounded cost and ...
We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robu...