We consider large-scale Markov decision processes (MDPs) with parameter un-certainty, under the robust MDP paradigm. Previous studies showed that robust MDPs, based on a minimax approach to handle uncertainty, can be solved us-ing dynamic programming for small to medium sized problems. However, due to the “curse of dimensionality”, MDPs that model real-life problems are typically prohibitively large for such approaches. In this work we employ a reinforcement learning approach to tackle this planning problem: we develop a robust approx-imate dynamic programming method based on a projected fixed point equation to approximately solve large scale robust MDPs. We show that the proposed method provably succeeds under certain technical conditions,...
When modeling real-world decision-theoretic planning problems in the Markov decision process (MDP) f...
When modeling real-world decision-theoretic planning problems in the Markov Decision Process (MDP) f...
Real-world planning problems frequently involve mixtures of continuous and discrete state variables ...
We consider large-scale Markov decision pro-cesses (MDPs) with parameter uncertainty, un-der the rob...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
We describe how to use robust Markov decision processes for value function ap-proximation with state...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
A large class of sequential decision-making problems undl uncertainty can bemodB3z as Markovand ...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
When modeling real-world decision-theoretic planning problems in the Markov decision process (MDP) f...
When modeling real-world decision-theoretic planning problems in the Markov Decision Process (MDP) f...
Real-world planning problems frequently involve mixtures of continuous and discrete state variables ...
We consider large-scale Markov decision pro-cesses (MDPs) with parameter uncertainty, un-der the rob...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDP) is a standard modeling tool for sequential decision making in a dyna...
Markov decision processes (MDPs) are the defacto framework for sequential decision making in the pre...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
This paper addresses the problem of planning under uncertainty in large Markov Decision Processes (M...
We describe how to use robust Markov decision processes for value function ap-proximation with state...
Markov decision processes (MDPs) are powerful tools for decision making in uncertain dynamic environ...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
A large class of sequential decision-making problems undl uncertainty can bemodB3z as Markovand ...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
While Markov Decision Processes (MDPs) have been shown to be effective models for planning under unc...
When modeling real-world decision-theoretic planning problems in the Markov decision process (MDP) f...
When modeling real-world decision-theoretic planning problems in the Markov Decision Process (MDP) f...
Real-world planning problems frequently involve mixtures of continuous and discrete state variables ...