We are interested in the problem of determining a course of action to achieve a desired objective in a non-deterministic environment. Markov decision processes (MDPs) provide a framework for repre-senting this action selection problem, and there are a number of algorithms that learn optimal policies within this formulation. This framework has also been used to study state space abstraction, problem decomposition, and policy reuse. These techniques sacrifice optimality of their solution for improved learning speed. In this paper we examine the sub-optimality of reusing policies that are solutions to subproblems. This is done within a restricted class of MDPs, namely those where non-zero reward is received only upon reaching a goal state. We ...
The problem of making optimal decisions in uncertain conditions is central to Artificial Intelligenc...
Planning plays an important role in the broad class of decision theory. Planning has drawn much atte...
Permission is hereby granted to the University of Alberta Libraries to reproduce single copies of th...
We are interested in the problem of determining a course of action to achieve a desired objective in...
This paper provides new techniques for abstracting the state space of a Markov Decision Process (MD...
A Markov decision process (MDP) relies on the notions of state, describing the current situation of ...
Infinite-horizon non-stationary Markov decision processes provide a general framework to model many ...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
What are the functionals of the reward that can be computed and optimized exactly in Markov Decision...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
We introduce a class of Markov decision problems (MDPs) which greatly simplify Reinforcement Learnin...
International audienceWe consider an agent interacting with an environment in a single stream of act...
We consider an agent interacting with an environment in a single stream of actions, observations, an...
Markov Decision Processes (MDPs) and Partially Observable MDPs (POMDPs) have been proposed as a fram...
The problem of making optimal decisions in uncertain conditions is central to Artificial Intelligenc...
Planning plays an important role in the broad class of decision theory. Planning has drawn much atte...
Permission is hereby granted to the University of Alberta Libraries to reproduce single copies of th...
We are interested in the problem of determining a course of action to achieve a desired objective in...
This paper provides new techniques for abstracting the state space of a Markov Decision Process (MD...
A Markov decision process (MDP) relies on the notions of state, describing the current situation of ...
Infinite-horizon non-stationary Markov decision processes provide a general framework to model many ...
Problems involving optimal sequential making in uncertain dynamic systems arise in domains such as e...
What are the functionals of the reward that can be computed and optimized exactly in Markov Decision...
summary:In this note we focus attention on identifying optimal policies and on elimination suboptima...
Markov decision problems (MDPs) provide the foundations for a number of problems of interest to AI r...
We introduce a class of Markov decision problems (MDPs) which greatly simplify Reinforcement Learnin...
International audienceWe consider an agent interacting with an environment in a single stream of act...
We consider an agent interacting with an environment in a single stream of actions, observations, an...
Markov Decision Processes (MDPs) and Partially Observable MDPs (POMDPs) have been proposed as a fram...
The problem of making optimal decisions in uncertain conditions is central to Artificial Intelligenc...
Planning plays an important role in the broad class of decision theory. Planning has drawn much atte...
Permission is hereby granted to the University of Alberta Libraries to reproduce single copies of th...