Literature DB >> 32251444

Discovery of hierarchical representations for efficient planning.

Momchil S Tomov1,2, Samyukta Yagati3, Agni Kumar3, Wanqian Yang4, Samuel J Gershman2.   

Abstract

We propose that humans spontaneously organize environments into clusters of states that support hierarchical planning, enabling them to tackle challenging problems by breaking them down into sub-problems at various levels of abstraction. People constantly rely on such hierarchical presentations to accomplish tasks big and small-from planning one's day, to organizing a wedding, to getting a PhD-often succeeding on the very first attempt. We formalize a Bayesian model of hierarchy discovery that explains how humans discover such useful abstractions. Building on principles developed in structure learning and robotics, the model predicts that hierarchy discovery should be sensitive to the topological structure, reward distribution, and distribution of tasks in the environment. In five simulations, we show that the model accounts for previously reported effects of environment structure on planning behavior, such as detection of bottleneck states and transitions. We then test the novel predictions of the model in eight behavioral experiments, demonstrating how the distribution of tasks and rewards can influence planning behavior via the discovered hierarchy, sometimes facilitating and sometimes hindering performance. We find evidence that the hierarchy discovery process unfolds incrementally across trials. Finally, we propose how hierarchy discovery and hierarchical planning might be implemented in the brain. Together, these findings present an important advance in our understanding of how the brain might use Bayesian inference to discover and exploit the hidden hierarchical structure of the environment.

Entities:  

Year:  2020        PMID: 32251444      PMCID: PMC7162548          DOI: 10.1371/journal.pcbi.1007594

Source DB:  PubMed          Journal:  PLoS Comput Biol        ISSN: 1553-734X            Impact factor:   4.475


  43 in total

1.  Multiple model-based reinforcement learning.

Authors:  Kenji Doya; Kazuyuki Samejima; Ken-ichi Katagiri; Mitsuo Kawato
Journal:  Neural Comput       Date:  2002-06       Impact factor: 2.026

2.  A rational analysis of rule-based concept learning.

Authors:  Noah D Goodman; Joshua B Tenenbaum; Jacob Feldman; Thomas L Griffiths
Journal:  Cogn Sci       Date:  2008-01-02

3.  Multistability and perceptual inference.

Authors:  Samuel J Gershman; Edward Vul; Joshua B Tenenbaum
Journal:  Neural Comput       Date:  2011-10-24       Impact factor: 2.026

4.  Optogenetic Editing Reveals the Hierarchical Organization of Learned Action Sequences.

Authors:  Claire E Geddes; Hao Li; Xin Jin
Journal:  Cell       Date:  2018-06-28       Impact factor: 41.582

Review 5.  Goal-directed instrumental action: contingency and incentive learning and their cortical substrates.

Authors:  B W Balleine; A Dickinson
Journal:  Neuropharmacology       Date:  1998 Apr-May       Impact factor: 5.250

Review 6.  Whatever next? Predictive brains, situated agents, and the future of cognitive science.

Authors:  Andy Clark
Journal:  Behav Brain Sci       Date:  2013-05-10       Impact factor: 12.579

7.  The successor representation in human reinforcement learning.

Authors:  I Momennejad; E M Russek; J H Cheong; M M Botvinick; N D Daw; S J Gershman
Journal:  Nat Hum Behav       Date:  2017-08-28

Review 8.  Habits, action sequences and reinforcement learning.

Authors:  Amir Dezfouli; Bernard W Balleine
Journal:  Eur J Neurosci       Date:  2012-04       Impact factor: 3.386

9.  Human-level control through deep reinforcement learning.

Authors:  Volodymyr Mnih; Koray Kavukcuoglu; David Silver; Andrei A Rusu; Joel Veness; Marc G Bellemare; Alex Graves; Martin Riedmiller; Andreas K Fidjeland; Georg Ostrovski; Stig Petersen; Charles Beattie; Amir Sadik; Ioannis Antonoglou; Helen King; Dharshan Kumaran; Daan Wierstra; Shane Legg; Demis Hassabis
Journal:  Nature       Date:  2015-02-26       Impact factor: 49.962

10.  Bonsai trees in your head: how the pavlovian system sculpts goal-directed choices by pruning decision trees.

Authors:  Quentin J M Huys; Neir Eshel; Elizabeth O'Nions; Luke Sheridan; Peter Dayan; Jonathan P Roiser
Journal:  PLoS Comput Biol       Date:  2012-03-08       Impact factor: 4.475

View more
  7 in total

1.  Computational evidence for hierarchically structured reinforcement learning in humans.

Authors:  Maria K Eckstein; Anne G E Collins
Journal:  Proc Natl Acad Sci U S A       Date:  2020-11-24       Impact factor: 11.205

Review 2.  Hierarchical Reinforcement Learning, Sequential Behavior, and the Dorsal Frontostriatal System.

Authors:  Miriam Janssen; Christopher LeWarne; Diana Burk; Bruno B Averbeck
Journal:  J Cogn Neurosci       Date:  2022-07-01       Impact factor: 3.420

3.  A weighted constraint satisfaction approach to human goal-directed decision making.

Authors:  Yuxuan Li; James L McClelland
Journal:  PLoS Comput Biol       Date:  2022-06-16       Impact factor: 4.779

4.  Eye movements reveal spatiotemporal dynamics of visually-informed planning in navigation.

Authors:  Seren Zhu; Kaushik J Lakshminarasimhan; Nastaran Arfaei; Dora E Angelaki
Journal:  Elife       Date:  2022-05-03       Impact factor: 8.713

5.  Learning compositional sequences with multiple time scales through a hierarchical network of spiking neurons.

Authors:  Amadeus Maes; Mauricio Barahona; Claudia Clopath
Journal:  PLoS Comput Biol       Date:  2021-03-25       Impact factor: 4.475

6.  Clone-structured graph representations enable flexible learning and vicarious evaluation of cognitive maps.

Authors:  Dileep George; Rajeev V Rikhye; Nishad Gothoskar; J Swaroop Guntupalli; Antoine Dedieu; Miguel Lázaro-Gredilla
Journal:  Nat Commun       Date:  2021-04-22       Impact factor: 14.919

7.  Sleep targets highly connected global and local nodes to aid consolidation of learned graph networks.

Authors:  G B Feld; M Bernard; A B Rawson; H J Spiers
Journal:  Sci Rep       Date:  2022-09-05       Impact factor: 4.996

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.