Computational Neuroscience Seminar - LCN


Friday, December 17th, 2010, 11h15, BC01

Kerstin Preuschoff, University of Zurich (homepage)

Q-learning the learning rate

Abstract:

In reinforcement learning, the learning rate is a fundamental parameter that determines how past prediction errors affect future predictions. Traditionally, the learning rate is kept constant, yet behaviorally, the learning rate is known to change both within and across contexts. Here, we propose a model-free approach to setting the learning rate. Key to our proposal is that one think about the learning rate as an action to be chosen to minimize a loss (prediction risk). This differentiates our proposal from learning models where the learning rate is adapted in a mechanistic way or based on a model of the task at hand. We use Q-learning to achieve prediction risk minimization and to simultaneously learn the prediction risk. Our algorithm produces learning rates that are a function of both risk (uncertainty in stable environments) and volatility (likelihood of changes in the environment). Learning rates decrease with risk and increase in volatility. The same functional dependence emerges in model-based approaches to setting the learning rate. We discuss behavioral evidence that these results parallel human and animal behavior. Imaging of the dopaminergic system, insula and anterior cingulate cortex of the primate brain supports the premise behind our account, namely, that reward learning is uncertainty-sensitive.

back