Efficient coding of cognitive variables underlies dopamine response and choice behavior
Reward expectations based on internal knowledge of the external environment are a core
component of adaptive behavior. However, internal knowledge may be inaccurate or …
component of adaptive behavior. However, internal knowledge may be inaccurate or …
[HTML][HTML] Dopamine-independent effect of rewards on choices through hidden-state inference
M Blanco-Pozo, T Akam, ME Walton - Nature Neuroscience, 2024 - nature.com
Dopamine is implicated in adaptive behavior through reward prediction error (RPE) signals
that update value estimates. There is also accumulating evidence that animals in structured …
that update value estimates. There is also accumulating evidence that animals in structured …
[HTML][HTML] Stable representations of decision variables for flexible behavior
Decisions occur in dynamic environments. In the framework of reinforcement learning, the
probability of performing an action is influenced by decision variables. Discrepancies …
probability of performing an action is influenced by decision variables. Discrepancies …
[HTML][HTML] Reward-predictive representations generalize across tasks in reinforcement learning
In computer science, reinforcement learning is a powerful framework with which artificial
agents can learn to maximize their performance for any given Markov decision process …
agents can learn to maximize their performance for any given Markov decision process …
[HTML][HTML] Fidelity of the representation of value in decision-making
PM Bays, BA Dowding - PLoS computational biology, 2017 - journals.plos.org
The ability to make optimal decisions depends on evaluating the expected rewards
associated with different potential actions. This process is critically dependent on the fidelity …
associated with different potential actions. This process is critically dependent on the fidelity …
Multiplexing signals in reinforcement learning with internal models and dopamine
H Nakahara - Current opinion in neurobiology, 2014 - Elsevier
Highlights•Decision-making involves various uses of internal models, such as reflecting
reward structures.•Generalized prediction errors are used to improve value-based decision …
reward structures.•Generalized prediction errors are used to improve value-based decision …
Learning to represent reward structure: A key to adapting to complex environments
H Nakahara, O Hikosaka - Neuroscience research, 2012 - Elsevier
Predicting outcomes is a critical ability of humans and animals. The dopamine reward
prediction error hypothesis, the driving force behind the recent progress in neural “value …
prediction error hypothesis, the driving force behind the recent progress in neural “value …
Computational noise in reward-guided learning drives behavioral variability in volatile environments
C Findling, V Skvortsova, R Dromnelle… - Nature …, 2019 - nature.com
When learning the value of actions in volatile environments, humans often make seemingly
irrational decisions that fail to maximize expected value. We reasoned that these 'non …
irrational decisions that fail to maximize expected value. We reasoned that these 'non …
Reward-bases: dopaminergic mechanisms for adaptive acquisition of multiple reward types
Animals can adapt their preferences for different types for reward according to physiological
state, such as hunger or thirst. To describe this ability, we propose a simple extension of …
state, such as hunger or thirst. To describe this ability, we propose a simple extension of …
[HTML][HTML] Belief state representation in the dopamine system
Learning to predict future outcomes is critical for driving appropriate behaviors.
Reinforcement learning (RL) models have successfully accounted for such learning, relying …
Reinforcement learning (RL) models have successfully accounted for such learning, relying …