Applying Q-Learning to multidimensional, real-valued state spaces is time-consuming in most cases. In this article, we deal with the assumption that a coarse partition of the state space is sufficient for learning good or even optimal policies. An algorithm is presented which constructs proper policies for abstract state spaces using an incremental procedure without approximating a Qfunction. By combining an approach similar to dynamic programming and a search for policies, we can speed up the learning process. To provide empirical evidence, we use a cart-pole system. Experiments were conducted for a simulated environment as well as for a real plant. 1
Abstract—A multiresolution state-space discretization method is developed for the episodic unsupervi...
Abstract — Q-learning is a technique used to compute an opti-mal policy for a controlled Markov chai...
We address the conflict between identification and control or alternatively, the conflict be-tween e...
Q-learning as well as other learning paradigms depend strongly on the representation of the underlyi...
Value-based approaches to reinforcement learning (RL) maintain a value function that measures the lo...
Autonomous automata should not only be able to learn how to behave efficiently in any predefined int...
The paper analyzes one of the main reinforcement learning methods - Q-learning, which is actively us...
Q-learning (Watkins, 1989) is a simple way for agents to learn how to act optimally in controlled Ma...
In this article, we consider learning problems in which the learning agent has only imprecise inform...
© 2016 The Authors and IOS Press. Q-learning associates states and actions of a Markov Decision Proc...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
Reinforcement Learning is carried out on-line, through trial-and-error interactions of the agent wit...
Reinforcement learning har proven to be very successful for finding optimal policies on uncertian an...
International audienceWe present a novel approach to state space discretization for constructivist a...
This paper presents a novel incremental algorithm that combines Q-learning, a well-known dynamic-pro...
Abstract—A multiresolution state-space discretization method is developed for the episodic unsupervi...
Abstract — Q-learning is a technique used to compute an opti-mal policy for a controlled Markov chai...
We address the conflict between identification and control or alternatively, the conflict be-tween e...
Q-learning as well as other learning paradigms depend strongly on the representation of the underlyi...
Value-based approaches to reinforcement learning (RL) maintain a value function that measures the lo...
Autonomous automata should not only be able to learn how to behave efficiently in any predefined int...
The paper analyzes one of the main reinforcement learning methods - Q-learning, which is actively us...
Q-learning (Watkins, 1989) is a simple way for agents to learn how to act optimally in controlled Ma...
In this article, we consider learning problems in which the learning agent has only imprecise inform...
© 2016 The Authors and IOS Press. Q-learning associates states and actions of a Markov Decision Proc...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
Reinforcement Learning is carried out on-line, through trial-and-error interactions of the agent wit...
Reinforcement learning har proven to be very successful for finding optimal policies on uncertian an...
International audienceWe present a novel approach to state space discretization for constructivist a...
This paper presents a novel incremental algorithm that combines Q-learning, a well-known dynamic-pro...
Abstract—A multiresolution state-space discretization method is developed for the episodic unsupervi...
Abstract — Q-learning is a technique used to compute an opti-mal policy for a controlled Markov chai...
We address the conflict between identification and control or alternatively, the conflict be-tween e...