Not logged in.

Contribution Details

Type Working Paper
Scope Discipline-based scholarship
Title Model-free reinforcement learning operates over information stored in working-memory to drive human choices
Organization Unit
Authors
  • Carolina Feher da Silva
  • Yuan-Wei Yao
  • Todd Anthony Hare
Language
  • English
Institution University of Zurich
Series Name bioRxiv
Number 107698
Number of Pages 24
Date 2017
Abstract Text Model-free learning creates stimulus-response associations, but are there limits to the types of stimuli it can operate over? Most experiments on reward-learning have used discrete sensory stimuli, but there is no algorithmic reason to restrict model-free learning to external stimuli, and theories suggest that model-free processes may operate over highly abstract concepts and goals. Our study aimed to determine whether model-free learning can operate over environmental states defined by information held in working memory. We compared the data from human participants in two conditions that presented learning cues either simultaneously or as a temporal sequence that required working memory. There was a significant influence of model-free learning in the working memory condition. Moreover, both groups showed greater model-free effects than simulated model-based agents. Thus, we show that model-free learning processes operate not just in parallel, but also in cooperation with canonical executive functions such as working memory to support behavior.
Digital Object Identifier 10.1101/107698
PDF File Download from ZORA
Export BibTeX
EP3 XML (ZORA)