We consider the problem of designing policies for partially observable Markov decision processes (POMDPs) with dynamic coherent risk objectives. Synthesizing risk-averse optimal policies for POMDPs requires infinite memory and thus undecidable. To overcome this difficulty, we propose a method based on bounded policy iteration for designing stochastic but finite state (memory) controllers, which takes advantage of standard convex optimization methods. Given a memory budget and optimality criterion, the proposed method modifies the stochastic finite state controller leading to sub-optimal solutions with lower coherent risk
In this paper, we address the problem of risk-aware conditional planning where the goal is generatin...
Partially observable Markov decision processes (POMDP) can be used as a model for planning in stocha...
We study planning problems for dynamical systems with uncertainty caused by measurement and process ...
We consider the problem of designing policies for partially observable Markov decision processes (PO...
Uncertain partially observable Markov decision processes (uPOMDPs) allow the probabilistic transitio...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2...
We present a memory-bounded optimization approach for solving infinite-horizon decen-tralized POMDPs...
AbstractIn this paper, we bring techniques from operations research to bear on the problem of choosi...
A partially observable Markov decision process (POMDP) is a model of planning and control that enabl...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
We consider the problem of designing policies for Markov decision processes (MDPs) with dynamic cohe...
When controlling dynamic systems such as mobile robots in uncertain environments, there is a trade o...
We propose a new method for learning policies for large, partially observable Markov decision proces...
This paper analyzes a connection between risk-sensitive and minimaxcriteria for discrete-time, finit...
This paper presents a model-based planner called the Probabilistic Sulu Planner or the p-Sulu Planne...
In this paper, we address the problem of risk-aware conditional planning where the goal is generatin...
Partially observable Markov decision processes (POMDP) can be used as a model for planning in stocha...
We study planning problems for dynamical systems with uncertainty caused by measurement and process ...
We consider the problem of designing policies for partially observable Markov decision processes (PO...
Uncertain partially observable Markov decision processes (uPOMDPs) allow the probabilistic transitio...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Aeronautics and Astronautics, 2...
We present a memory-bounded optimization approach for solving infinite-horizon decen-tralized POMDPs...
AbstractIn this paper, we bring techniques from operations research to bear on the problem of choosi...
A partially observable Markov decision process (POMDP) is a model of planning and control that enabl...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
We consider the problem of designing policies for Markov decision processes (MDPs) with dynamic cohe...
When controlling dynamic systems such as mobile robots in uncertain environments, there is a trade o...
We propose a new method for learning policies for large, partially observable Markov decision proces...
This paper analyzes a connection between risk-sensitive and minimaxcriteria for discrete-time, finit...
This paper presents a model-based planner called the Probabilistic Sulu Planner or the p-Sulu Planne...
In this paper, we address the problem of risk-aware conditional planning where the goal is generatin...
Partially observable Markov decision processes (POMDP) can be used as a model for planning in stocha...
We study planning problems for dynamical systems with uncertainty caused by measurement and process ...