Predictive representations can link model-based reinforcement learning to model-free mechanisms
Author(s): Russek, Evan M.; Momennejad, Ida; Botvinick, Matthew M.; Gershman, Samuel J.; Daw, Nathaniel D.
DownloadTo refer to this page use:
http://arks.princeton.edu/ark:/88435/pr1348k
Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Russek, Evan M. | - |
dc.contributor.author | Momennejad, Ida | - |
dc.contributor.author | Botvinick, Matthew M. | - |
dc.contributor.author | Gershman, Samuel J. | - |
dc.contributor.author | Daw, Nathaniel D. | - |
dc.date.accessioned | 2020-02-19T21:59:09Z | - |
dc.date.available | 2020-02-19T21:59:09Z | - |
dc.date.issued | 2017-09-25 | en_US |
dc.identifier.citation | Russek, Evan M, Momennejad, Ida, Botvinick, Matthew M, Gershman, Samuel J, Daw, Nathaniel D. (2017). Predictive representations can link model-based reinforcement learning to model-free mechanisms.. PLoS computational biology, 13 (9), e1005768 - ?. doi:10.1371/journal.pcbi.1005768 | en_US |
dc.identifier.issn | 1553-734X | - |
dc.identifier.uri | http://arks.princeton.edu/ark:/88435/pr1348k | - |
dc.description.abstract | Humans and animals are capable of evaluating actions by considering their long-run future rewards through a process described using model-based reinforcement learning (RL) algorithms. The mechanisms by which neural circuits perform the computations prescribed by model-based RL remain largely unknown; however, multiple lines of evidence suggest that neural circuits supporting model-based behavior are structurally homologous to and overlapping with those thought to carry out model-free temporal difference (TD) learning. Here, we lay out a family of approaches by which model-based computation may be built upon a core of TD learning. The foundation of this framework is the successor representation, a predictive state representation that, when combined with TD learning of value predictions, can produce a subset of the behaviors associated with model-based learning, while requiring less decision-time computation than dynamic programming. Using simulations, we delineate the precise behavioral capabilities enabled by evaluating actions using this approach, and compare them to those demonstrated by biological organisms. We then introduce two new algorithms that build upon the successor representation while progressively mitigating its limitations. Because this framework can account for the full range of observed putatively model-based behaviors while still utilizing a core TD framework, we suggest that it represents a neurally plausible family of mechanisms for model-based evaluation. | en_US |
dc.language | eng | en_US |
dc.language.iso | en_US | en_US |
dc.relation.ispartof | PLoS computational biology | en_US |
dc.rights | Final published version. This is an open access article. | en_US |
dc.title | Predictive representations can link model-based reinforcement learning to model-free mechanisms | en_US |
dc.type | Journal Article | en_US |
dc.identifier.doi | doi:10.1371/journal.pcbi.1005768 | - |
dc.identifier.eissn | 1553-7358 | - |
pu.type.symplectic | http://www.symplectic.co.uk/publications/atom-terms/1.0/journal-article | en_US |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
journal.pcbi.1005768.pdf | 4.75 MB | Adobe PDF | View/Download |
Items in OAR@Princeton are protected by copyright, with all rights reserved, unless otherwise indicated.