In contrast to previous competitions, where the problems were goal-based, the 2011 International Probabilistic Planning Competition (IPPC-2011) emphasized finite-horizon reward maximization problems with large branching factors. These MDPs modeled more realistic planning scenarios and presented challenges to the previous state-of-the-art planners (e.g., those from IPPC-2008), which were primarily based on domain determinization — a technique more suited to goal-oriented MDPs with small branching factors. Moreover, large branching factors render the existing implementations of RTDP- and LAO-style algorithms inefficient as well. In this paper we present GLUTTON, our planner at IPPC-2011 that performed well on these challenging MDPs. The...
We describe the version of the GPT planner to be used in the planning competition. This version, cal...
We explore approximate policy iteration, replacing the usual costfunction learning step with a learn...
Solving multiagent planning problems modeled as DEC-POMDPs is an important challenge. These models ...
Recent domain-determinization techniques have been very successful in many probabilistic planning pr...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
Markov decision processes (MDP) offer a rich model that has been extensively used by the AI communit...
International audienceOver the past few years, attempts to scale up infinite-horizon DECPOMDPs are m...
We address the class of probabilistic planning problems where the objective is to maximize the proba...
We introduce a family of MDP reduced models characterized by two parameters: the maximum number of p...
We address the class of probabilistic planning problems where the objective is to maximize the proba...
Most traditional approaches to probabilistic planning in relationally specified MDPs rely on groundi...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
International audienceWe consider the problem of online planning in a Markov decision process with d...
Recent advances in deep learning have enabled optimization of deep reactive policies (DRPs) for cont...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
We describe the version of the GPT planner to be used in the planning competition. This version, cal...
We explore approximate policy iteration, replacing the usual costfunction learning step with a learn...
Solving multiagent planning problems modeled as DEC-POMDPs is an important challenge. These models ...
Recent domain-determinization techniques have been very successful in many probabilistic planning pr...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
Markov decision processes (MDP) offer a rich model that has been extensively used by the AI communit...
International audienceOver the past few years, attempts to scale up infinite-horizon DECPOMDPs are m...
We address the class of probabilistic planning problems where the objective is to maximize the proba...
We introduce a family of MDP reduced models characterized by two parameters: the maximum number of p...
We address the class of probabilistic planning problems where the objective is to maximize the proba...
Most traditional approaches to probabilistic planning in relationally specified MDPs rely on groundi...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
International audienceWe consider the problem of online planning in a Markov decision process with d...
Recent advances in deep learning have enabled optimization of deep reactive policies (DRPs) for cont...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
We describe the version of the GPT planner to be used in the planning competition. This version, cal...
We explore approximate policy iteration, replacing the usual costfunction learning step with a learn...
Solving multiagent planning problems modeled as DEC-POMDPs is an important challenge. These models ...