Theories of reward learning in neuroscience have focused on two families of algorithms thought to capture deliberative versus habitual choice. ‘Model-based’ algorithms compute the value of candidate actions from scratch, whereas ‘model-free’ algorithms make choice more efficient but less flexible by storing pre-computed action values. We examine an intermediate algorithmic family, the successor representation, which balances flexibility and efficiency by storing partially computed action values: predictions about future events. These pre-computation strategies differ in how they update their choices following changes in a task. The successor representation’s reliance on stored predictions about future states predicts a unique signature of i...
International audienceMany of the decisions we make in our everyday lives are sequential and entail ...
To decide optimally between available options, organisms need to learn the values associated with th...
How do we use our memories of the past to guide decisions we’ve never had to make before? Although e...
Theories of reward learning in neuroscience have focused on two families of algorithms, thought to c...
Humans and animals are capable of evaluating actions by considering their long-run future rewards th...
Many of the decisions we make in our everyday lives are sequential and entail sparse rewards. While ...
The ability to predict long-term future rewards is crucial for survival. Some animals may have to ...
National audienceAccounting for behavioural capabilities and flexibilities experimentally observed i...
SummaryReinforcement learning (RL) uses sequential experience with situations (“states”) and outcome...
Humans and animals are capable of evaluating actions by considering their long-run future rewards th...
Making sequential decisions to harvest rewards is a notoriously difficult problem. One difficulty is...
The ability to integrate past and current feedback associated with di↵erent environmental stimuli is...
Balancing exploration and exploitation is one of the central problems in reinforcement learning. We ...
For adaptive real-time behavior in real-world contexts, the brain needs to allow past information ov...
<p>To behave adaptively, we must learn from the consequences of our actions. Doing so is difficult w...
International audienceMany of the decisions we make in our everyday lives are sequential and entail ...
To decide optimally between available options, organisms need to learn the values associated with th...
How do we use our memories of the past to guide decisions we’ve never had to make before? Although e...
Theories of reward learning in neuroscience have focused on two families of algorithms, thought to c...
Humans and animals are capable of evaluating actions by considering their long-run future rewards th...
Many of the decisions we make in our everyday lives are sequential and entail sparse rewards. While ...
The ability to predict long-term future rewards is crucial for survival. Some animals may have to ...
National audienceAccounting for behavioural capabilities and flexibilities experimentally observed i...
SummaryReinforcement learning (RL) uses sequential experience with situations (“states”) and outcome...
Humans and animals are capable of evaluating actions by considering their long-run future rewards th...
Making sequential decisions to harvest rewards is a notoriously difficult problem. One difficulty is...
The ability to integrate past and current feedback associated with di↵erent environmental stimuli is...
Balancing exploration and exploitation is one of the central problems in reinforcement learning. We ...
For adaptive real-time behavior in real-world contexts, the brain needs to allow past information ov...
<p>To behave adaptively, we must learn from the consequences of our actions. Doing so is difficult w...
International audienceMany of the decisions we make in our everyday lives are sequential and entail ...
To decide optimally between available options, organisms need to learn the values associated with th...
How do we use our memories of the past to guide decisions we’ve never had to make before? Although e...