Transfer in Reinforcement Learning aims to improve learning performance on target tasks using knowledge from experienced source tasks. Successor features (SF) are a prominent transfer mechanism in domains where the reward function changes between tasks. They reevaluate the expected return of previously learned policies in a new target task and to transfer their knowledge. A limiting factor of the SF framework is its assumption that rewards linearly decompose into successor features and a reward weight vector. We propose a novel SF mechanism, $\xi$-learning, based on learning the cumulative discounted probability of successor features. Crucially, $\xi$-learning allows to reevaluate the expected return of policies for general reward functions...
We present an algorithm for Inverse Reinforcement Learning (IRL) from expert state observations only...
Theories of reward learning in neuroscience have focused on two families of algorithms thought to ca...
International audienceThis article addresses a particular Transfer Reinforcement Learning (RL) probl...
source code available at https://gitlab.inria.fr/robotlearn/xi_learningTransfer in Reinforcement Lea...
Abstract: Successor Features stand at the boundary between modelfree and model-based Reinforcement L...
Transfer in reinforcement learning refers to the notion that generalization should occur not only wi...
While reinforcement learning algorithms provide automated acquisition of optimal policies, practical...
Successor-style representations have many advantages for reinforcement learning: for example, they c...
A longstanding goal in reinforcement learning is to build intelligent agents that show fast learning...
A common setting in multitask reinforcement learning (RL) demands that an agent rapidly adapt to var...
In reinforcement learning, reward is used to guide the learning process. The reward is often designe...
Reinforcement Learning (RL) has seen exponential performance improvements over the past decade, achi...
We consider how to transfer knowledge from previous tasks (MDPs) to a current task in long-lived and...
We propose a method for meta-learning reinforcement learning algorithms by searching over the space ...
In the search for more sample-efficient reinforcement-learning (RL) algorithms, a promising directio...
We present an algorithm for Inverse Reinforcement Learning (IRL) from expert state observations only...
Theories of reward learning in neuroscience have focused on two families of algorithms thought to ca...
International audienceThis article addresses a particular Transfer Reinforcement Learning (RL) probl...
source code available at https://gitlab.inria.fr/robotlearn/xi_learningTransfer in Reinforcement Lea...
Abstract: Successor Features stand at the boundary between modelfree and model-based Reinforcement L...
Transfer in reinforcement learning refers to the notion that generalization should occur not only wi...
While reinforcement learning algorithms provide automated acquisition of optimal policies, practical...
Successor-style representations have many advantages for reinforcement learning: for example, they c...
A longstanding goal in reinforcement learning is to build intelligent agents that show fast learning...
A common setting in multitask reinforcement learning (RL) demands that an agent rapidly adapt to var...
In reinforcement learning, reward is used to guide the learning process. The reward is often designe...
Reinforcement Learning (RL) has seen exponential performance improvements over the past decade, achi...
We consider how to transfer knowledge from previous tasks (MDPs) to a current task in long-lived and...
We propose a method for meta-learning reinforcement learning algorithms by searching over the space ...
In the search for more sample-efficient reinforcement-learning (RL) algorithms, a promising directio...
We present an algorithm for Inverse Reinforcement Learning (IRL) from expert state observations only...
Theories of reward learning in neuroscience have focused on two families of algorithms thought to ca...
International audienceThis article addresses a particular Transfer Reinforcement Learning (RL) probl...