POMDP algorithms have made significant progress in recent years by allowing practitioners to find good solutions to increasingly large problems. Most approaches (including point-based and policy iteration techniques) operate by refining a lower bound of the optimal value function. Several approaches (e.g., HSVI2, SARSOP, grid-based approaches and online forward search) also refine an upper bound. However, approximating the optimal value function by an upper bound is computationally expensive and therefore tightness is often sacrificed to improve efficiency (e.g., sawtooth approximation). In this paper, we describe a new approach to efficiently compute tighter bounds by i) conducting a prioritized breadth first search over the reachable ...
Abstract. Recent scaling up of POMDP solvers towards realistic applications is largely due to point-...
Constrained partially observable Markov decision processes (CPOMDPs) have been used to model various...
This paper presents and evaluates two pruning techniques to reinforce the efficiency of constraint o...
POMDP algorithms have made significant progress in re-cent years by allowing practitioners to find g...
We present a novel POMDP planning algorithm called heuristic search value iteration (HSVI). HSVI is ...
In planning with partially observable Markov decision processes, pre-compiled policies are often rep...
In many fields in computational sustainability, applications of POMDPs are inhibited by the complexi...
Abstract—Recent scaling up of POMDP solvers towards re-alistic applications is largely due to point-...
In many POMDP applications in computational sustainability, it is important that the computed pol-ic...
Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conferenc
Partially Observable Markov Decision Processes (POMDPs) are powerful models for planning under uncer...
Planning in partially observable environments remains a challenging problem, de-spite significant re...
Partially Observable Markov Decision Processes (pomdps) are gen-eral models of sequential decision p...
This paper introduces the even-odd POMDP an approximation to POMDPs Partially Observable Markov Deci...
Value iteration is a popular algorithm for finding near optimal policies for POMDPs. It is inefficie...
Abstract. Recent scaling up of POMDP solvers towards realistic applications is largely due to point-...
Constrained partially observable Markov decision processes (CPOMDPs) have been used to model various...
This paper presents and evaluates two pruning techniques to reinforce the efficiency of constraint o...
POMDP algorithms have made significant progress in re-cent years by allowing practitioners to find g...
We present a novel POMDP planning algorithm called heuristic search value iteration (HSVI). HSVI is ...
In planning with partially observable Markov decision processes, pre-compiled policies are often rep...
In many fields in computational sustainability, applications of POMDPs are inhibited by the complexi...
Abstract—Recent scaling up of POMDP solvers towards re-alistic applications is largely due to point-...
In many POMDP applications in computational sustainability, it is important that the computed pol-ic...
Advances in Neural Information Processing Systems 20 - Proceedings of the 2007 Conferenc
Partially Observable Markov Decision Processes (POMDPs) are powerful models for planning under uncer...
Planning in partially observable environments remains a challenging problem, de-spite significant re...
Partially Observable Markov Decision Processes (pomdps) are gen-eral models of sequential decision p...
This paper introduces the even-odd POMDP an approximation to POMDPs Partially Observable Markov Deci...
Value iteration is a popular algorithm for finding near optimal policies for POMDPs. It is inefficie...
Abstract. Recent scaling up of POMDP solvers towards realistic applications is largely due to point-...
Constrained partially observable Markov decision processes (CPOMDPs) have been used to model various...
This paper presents and evaluates two pruning techniques to reinforce the efficiency of constraint o...