Optimal policy computation in finite-horizon Markov decision processes is a classical problem in optimization with lots of pratical applications. For stationary policies and infinite horizon it is known to be solvable in polynomial time by linear programming, whereas for finite-horizon it is a longstanding open problem. We consider this problem for a slightly generalized model, namely partially-observable Markov decision processes (POMDPs). We show that it is NP-complete and that -- unless P = NP -- the optimal policy cannot be polynomial time "-approximated for any " ! 1. A similar result is shown for the average policy. The problem of whether the average performance is positive is shown to be PPcomplete for stationary policies...
A partially observable Markov decision process (POMDP) is a model of planning and control that enabl...
AbstractIn the paper we consider the complexity of constructing optimal policies (strategies) for so...
There is much interest in using partially observable Markov decision processes (POMDPs) as a formal ...
A partially-observable Markov decision process (POMDP) is a generalization of a Markov decision proc...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
We study the problem of approximation of optimal values in partially-observable Markov decision proc...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
Partially Observable Markov Decision Processes (POMDPs) provide a rich representation for agents act...
The concept of partially observable Markov decision processes was born to handle the problem of lack...
We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specif...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
Partially observable Markov decision processes (POMDP) can be used as a model for planning in stocha...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
A partially observable Markov decision process (POMDP) is a model of planning and control that enabl...
AbstractIn the paper we consider the complexity of constructing optimal policies (strategies) for so...
There is much interest in using partially observable Markov decision processes (POMDPs) as a formal ...
A partially-observable Markov decision process (POMDP) is a generalization of a Markov decision proc...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
Partially observable Markov decision processes (POMDPs) provide a natural and principled framework t...
We study the problem of approximation of optimal values in partially-observable Markov decision proc...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
Partially Observable Markov Decision Processes (POMDPs) provide a rich representation for agents act...
The concept of partially observable Markov decision processes was born to handle the problem of lack...
We consider partially observable Markov decision processes (POMDPs) with ω-regular conditions specif...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
Partially observable Markov decision processes (POMDP) can be used as a model for planning in stocha...
We consider partially observable Markov decision processes (POMDPs) with a set of target states and ...
A partially observable Markov decision process (POMDP) is a model of planning and control that enabl...
AbstractIn the paper we consider the complexity of constructing optimal policies (strategies) for so...
There is much interest in using partially observable Markov decision processes (POMDPs) as a formal ...