Q-learning as well as other learning paradigms depend strongly on the representation of the underlying state space. As a special case of the hidden state problem we investigate the effect of a self-organizing discretization of the state space in a simple control problem. We apply the neural gas algorithm with adaptation of learning rate and neighborhood range to a simulated cart-pole problem. The learning parameters are determined by the ambiguity of successful actions inside each cell
This thesis involves the use of a reinforcement learning algorithm (RL) called Q-learning to train a...
Value-based approaches to reinforcement learning (RL) maintain a value function that measures the lo...
The convergence property of reinforcement learning has been extensively investigated in the field of...
Q-learning as well as other learning paradigms depend strongly on the representation of the underlyi...
Autonomous automata should not only be able to learn how to behave efficiently in any predefined int...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
Applying Q-Learning to multidimensional, real-valued state spaces is time-consuming in most cases. I...
We present a variant of the Q-learning algorithm with automatic control of the exploration rate by a...
Q-learning can be used to find an optimal action-selection policy for any given finite Markov Decisi...
Proceeding of: RoboCup-99: Robot Soccer World Cup III, July 27 to August 6, 1999, Stockholm, SwedenR...
International audienceIn this paper, we propose a contribution in the field of Reinforcement Learnin...
A self-organising architecture, loosely based upon a particular implementation of adaptive resonance...
Reinforcement learning has proven to be a set of successful techniques for nding optimal policies ...
We address the conflict between identification and control or alternatively, the conflict be-tween e...
Abstract — Q-learning is a technique used to compute an opti-mal policy for a controlled Markov chai...
This thesis involves the use of a reinforcement learning algorithm (RL) called Q-learning to train a...
Value-based approaches to reinforcement learning (RL) maintain a value function that measures the lo...
The convergence property of reinforcement learning has been extensively investigated in the field of...
Q-learning as well as other learning paradigms depend strongly on the representation of the underlyi...
Autonomous automata should not only be able to learn how to behave efficiently in any predefined int...
Abstract. Q-learning can be used to learn a control policy that max-imises a scalar reward through i...
Applying Q-Learning to multidimensional, real-valued state spaces is time-consuming in most cases. I...
We present a variant of the Q-learning algorithm with automatic control of the exploration rate by a...
Q-learning can be used to find an optimal action-selection policy for any given finite Markov Decisi...
Proceeding of: RoboCup-99: Robot Soccer World Cup III, July 27 to August 6, 1999, Stockholm, SwedenR...
International audienceIn this paper, we propose a contribution in the field of Reinforcement Learnin...
A self-organising architecture, loosely based upon a particular implementation of adaptive resonance...
Reinforcement learning has proven to be a set of successful techniques for nding optimal policies ...
We address the conflict between identification and control or alternatively, the conflict be-tween e...
Abstract — Q-learning is a technique used to compute an opti-mal policy for a controlled Markov chai...
This thesis involves the use of a reinforcement learning algorithm (RL) called Q-learning to train a...
Value-based approaches to reinforcement learning (RL) maintain a value function that measures the lo...
The convergence property of reinforcement learning has been extensively investigated in the field of...