Model-Based Reinforcement Learning (MBRL) can greatly profit from using world models for estimating the consequences of selecting particular actions: an animat can construct such a model from its experiences and use it for computing rewarding behavior. We study the problem of collecting useful experiences through exploration in stochastic environments. Towards this end we use MBRL to maximize exploration rewards (in addition to environmental rewards) for visits of states that promise information gain. We also combine MBRL and the Interval Estimation algorithm (Kaelbling, 1993). Experimental results demonstrate the advantages of our approaches
Equipping artificial agents with useful exploration mechanisms remains a challenge to this day. Huma...
Recent Reinforcement Learning (RL) algorithms, such as R-MAX, make (with high probability) only a sm...
We generalise the problem of reward modelling (RM) for reinforcement learning (RL) to handle non-Mar...
Model-Based Reinforcement Learning (MBRL) can greatly profit from using world models for estimating...
Reinforcement learning can greatly prot from world models updated by experience and used for computi...
Reinforcement learning can greatly profit from world models updated by experience and used for comp...
Formal exploration approaches in model-based reinforcement learning estimate the accuracy of the cur...
Reinforcement learning systems are often concerned with balancing exploration of untested actions ag...
Formal exploration approaches in model-based reinforcement learning estimate the accuracy of the cur...
International audienceReinforcement learning (RL) is a paradigm for learning sequential decision mak...
One problem of current Reinforcement Learning algorithms is finding a balance between exploitation o...
International audienceRealistic environments often provide agents with very limited feedback. When t...
The impetus for exploration in reinforcement learning (RL) is decreasing uncertainty about the envir...
Reinforcement learning with sparse rewards is still an open challenge. Classic methods rely on getti...
This paper discusses parameter-based exploration methods for reinforcement learning. Parameter-based...
Equipping artificial agents with useful exploration mechanisms remains a challenge to this day. Huma...
Recent Reinforcement Learning (RL) algorithms, such as R-MAX, make (with high probability) only a sm...
We generalise the problem of reward modelling (RM) for reinforcement learning (RL) to handle non-Mar...
Model-Based Reinforcement Learning (MBRL) can greatly profit from using world models for estimating...
Reinforcement learning can greatly prot from world models updated by experience and used for computi...
Reinforcement learning can greatly profit from world models updated by experience and used for comp...
Formal exploration approaches in model-based reinforcement learning estimate the accuracy of the cur...
Reinforcement learning systems are often concerned with balancing exploration of untested actions ag...
Formal exploration approaches in model-based reinforcement learning estimate the accuracy of the cur...
International audienceReinforcement learning (RL) is a paradigm for learning sequential decision mak...
One problem of current Reinforcement Learning algorithms is finding a balance between exploitation o...
International audienceRealistic environments often provide agents with very limited feedback. When t...
The impetus for exploration in reinforcement learning (RL) is decreasing uncertainty about the envir...
Reinforcement learning with sparse rewards is still an open challenge. Classic methods rely on getti...
This paper discusses parameter-based exploration methods for reinforcement learning. Parameter-based...
Equipping artificial agents with useful exploration mechanisms remains a challenge to this day. Huma...
Recent Reinforcement Learning (RL) algorithms, such as R-MAX, make (with high probability) only a sm...
We generalise the problem of reward modelling (RM) for reinforcement learning (RL) to handle non-Mar...