| Literature DB >> 25675480 |
Quentin J M Huys1, Níall Lally2, Paul Faulkner3, Neir Eshel4, Erich Seifritz5, Samuel J Gershman6, Peter Dayan7, Jonathan P Roiser8.
Abstract
Humans routinely formulate plans in domains so complex that even the most powerful computers are taxed. To do so, they seem to avail themselves of many strategies and heuristics that efficiently simplify, approximate, and hierarchically decompose hard tasks into simpler subtasks. Theoretical and cognitive research has revealed several such strategies; however, little is known about their establishment, interaction, and efficiency. Here, we use model-based behavioral analysis to provide a detailed examination of the performance of human subjects in a moderately deep planning task. We find that subjects exploit the structure of the domain to establish subgoals in a way that achieves a nearly maximal reduction in the cost of computing values of choices, but then combine partial searches with greedy local steps to solve subtasks, and maladaptively prune the decision trees of subtasks in a reflexive manner upon encountering salient losses. Subjects come idiosyncratically to favor particular sequences of actions to achieve subgoals, creating novel complex actions or "options."Entities:
Keywords: hierarchical reinforcement learning; memoization; planning; pruning
Mesh:
Year: 2015 PMID: 25675480 PMCID: PMC4364207 DOI: 10.1073/pnas.1414219112
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 11.205