Abstract: We study the synthesis of robust optimal control policies for Markov decision processes with transition uncertainty (UMDPs) and subject to two types of constraints: (i) constraints on the worst-case, maximal total cost and (ii) safety-threshold constraints that bound the worst-case probability of visiting a set of error states. For maximal total cost constraints, we propose a state-augmentation method and a two-step synthesis algorithm to generate deterministic, memoryless optimal policies given the reward to be maximized. For safety threshold constraints, we introduce a new cost function and provide an approximately optimal solution by a reduction to an uncertain Markov decision process under a maximal total cost constraint. The ...
This paper analyzes a connection between risk-sensitive and minimax criteria for discrete-time, fini...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
We consider the problem of designing policies for partially observable Markov decision processes (PO...
We study the synthesis of robust optimal control policies for Markov decision processes with transit...
We present a method for designing a robust control policy for an uncertain system subject to tempora...
We present a method for designing robust controllers for dynamical systems with linear temporal logi...
Abstract — We present a method for designing a robust control policy for an uncertain system subject...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
In this paper, approximate dynamic programming (ADP) problems are modeled by discounted infinite-hor...
The authors consider the fundamental problem of finding good policies in uncertain models. It is dem...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
Abstract — This paper presents a new robust decision-making algorithm that accounts for model uncert...
This paper analyzes a connection between risk-sensitive and minimax criteria for discrete-time, fini...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
We consider the problem of designing policies for partially observable Markov decision processes (PO...
We study the synthesis of robust optimal control policies for Markov decision processes with transit...
We present a method for designing a robust control policy for an uncertain system subject to tempora...
We present a method for designing robust controllers for dynamical systems with linear temporal logi...
Abstract — We present a method for designing a robust control policy for an uncertain system subject...
Optimal solutions to Markov decision problems may be very sensitive with respect to the state transi...
In robust Markov decision processes (MDPs), the uncertainty in the transition kernel is addressed by...
In this paper, we seek robust policies for uncertain Markov Decision Processes (MDPs). Most robust o...
In this paper, approximate dynamic programming (ADP) problems are modeled by discounted infinite-hor...
The authors consider the fundamental problem of finding good policies in uncertain models. It is dem...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
Abstract — This paper presents a new robust decision-making algorithm that accounts for model uncert...
This paper analyzes a connection between risk-sensitive and minimax criteria for discrete-time, fini...
The authors consider the fundamental problem of nding good poli-cies in uncertain models. It is demo...
We consider the problem of designing policies for partially observable Markov decision processes (PO...