Many previous studies of the brain areas involved in reward prediction errors have not accounted for the downstream projections of dopamine areas when interpreting these results. We propose that paradigms like conditioned inhibition, which involves pairing a rewarded CS with an inhibitor that always cancels the reward, can reduce this confound and allow for further specification of the computational role different brain regions play into the RPE signal. Further predictions of the role of different brain areas in reward learning and how positive and negative valence learning interact in the brain are inspired by the PVLV model, a more biologically plausible alternative to TD learning that uses two parameters, learned value and primary value,...
Reinforcement learning describes motivated behavior in terms of two abstract signals. The representa...
Predictions help guide learning. As we encounter objects in our environment, we make predictions abo...
Reward learning depends on accurate reward associations with potential choices. These associations c...
Compared to our understanding of positive prediction error signals occurring due to unexpected rewar...
International audienceWhether maximizing rewards and minimizing punishments rely on distinct brain s...
Prediction-error signals consistent with formal models of “reinforcement learning” (RL) have repeate...
Learning occurs when an outcome deviates from expectation (prediction error). According to formal le...
Reinforcement learning (RL) in simple instrumental tasks is usually modeled as a monolithic process ...
International audienceOne of the earliest attempts to understand how animals learn involved pairing ...
Dopamine neurons are thought to facilitate learning by signaling reward prediction errors (RPEs), th...
Prediction-error signals consistent with formal models of "reinforcement learning" (RL) have repeate...
Compared to our understanding of positive prediction error signals occurring due to unexpected rewar...
It has been suggested that the midbrain dopamine (DA) neurons, receiving inputs from the cortico-bas...
Dopamine (DA) neurons in the ventral tegmental area (VTA) are thought to encode reward prediction er...
Previous reports have described that neural activities in midbrain dopamine areas are sensitive to u...
Reinforcement learning describes motivated behavior in terms of two abstract signals. The representa...
Predictions help guide learning. As we encounter objects in our environment, we make predictions abo...
Reward learning depends on accurate reward associations with potential choices. These associations c...
Compared to our understanding of positive prediction error signals occurring due to unexpected rewar...
International audienceWhether maximizing rewards and minimizing punishments rely on distinct brain s...
Prediction-error signals consistent with formal models of “reinforcement learning” (RL) have repeate...
Learning occurs when an outcome deviates from expectation (prediction error). According to formal le...
Reinforcement learning (RL) in simple instrumental tasks is usually modeled as a monolithic process ...
International audienceOne of the earliest attempts to understand how animals learn involved pairing ...
Dopamine neurons are thought to facilitate learning by signaling reward prediction errors (RPEs), th...
Prediction-error signals consistent with formal models of "reinforcement learning" (RL) have repeate...
Compared to our understanding of positive prediction error signals occurring due to unexpected rewar...
It has been suggested that the midbrain dopamine (DA) neurons, receiving inputs from the cortico-bas...
Dopamine (DA) neurons in the ventral tegmental area (VTA) are thought to encode reward prediction er...
Previous reports have described that neural activities in midbrain dopamine areas are sensitive to u...
Reinforcement learning describes motivated behavior in terms of two abstract signals. The representa...
Predictions help guide learning. As we encounter objects in our environment, we make predictions abo...
Reward learning depends on accurate reward associations with potential choices. These associations c...