| Literature DB >> 35294284 |
Frederick Callaway1, Yash Raj Jain2, Bas van Opheusden1, Priyam Das3, Gabriela Iwama2, Sayan Gul4, Paul M Krueger5, Frederic Becker2, Thomas L Griffiths1,5, Falk Lieder2.
Abstract
SignificanceMany bad decisions and their devastating consequences could be avoided if people used optimal decision strategies. Here, we introduce a principled computational approach to improving human decision making. The basic idea is to give people feedback on how they reach their decisions. We develop a method that leverages artificial intelligence to generate this feedback in such a way that people quickly discover the best possible decision strategies. Our empirical findings suggest that a principled computational approach leads to improvements in decision-making competence that transfer to more difficult decisions in more complex environments. In the long run, this line of work might lead to apps that teach people clever strategies for decision making, reasoning, goal setting, planning, and goal achievement.Entities:
Keywords: bounded rationality; cognitive training; heuristics; rationality enhancement
Mesh:
Year: 2022 PMID: 35294284 PMCID: PMC8944825 DOI: 10.1073/pnas.2117432119
Source DB: PubMed Journal: Proc Natl Acad Sci U S A ISSN: 0027-8424 Impact factor: 11.205
Fig. 1.The Mouselab-MDP paradigm. (A) Participants click to reveal the rewards at future states to construct a plan. (B) Metacognitive feedback penalizes suboptimal decision-making operations (clicks) with a delay and provides instruction on what operation(s) should have been taken instead.
Fig. 2.Metacognitive feedback accelerates learning and improves performance. (A) Average score in the test block for each condition. (B) Proportion of participants who started by inspecting a potential final outcome split by condition. Here and in all future plots, the error bars and shaded areas convey 95% confidence intervals produced by 1,000 bootstrap samples. Scores are bootstrapped over participant means. Asterisks indicate significance of the permutation test reported in the main text as follows: **P < 0.01.
Fig. 3.The benefits of metacognitive feedback transfer to more difficult problems and are retained for at least 24 h. (A) The near-transfer task is a five-step sequential decision problem where the rewards are normally distributed with a variance that increases exponentially from the first step to the last step. (B) Average performance on the transfer task given immediately after training. (C) The same, but with a 24-h delay between training and test. **P < 0.01, ***P < 0.001.
Fig. 4.Metacognitive feedback improved people’s performance in an environment where the rewards are independently and identically distributed across all locations. *P < 0.05, **P < 0.01.
Fig. 5.Far transfer. (A) The task environment shares the core property that rewards are more variable in distant states, but bears little resemblance to the training task. (B) Transfer performance in each condition. (C) Proportion of participants planning backward on each trial. The dashed line indicates the switch to the transfer task.
Fig. 6.Test performance with different subsets of the two components of metacognitive feedback (delay penalties and information about the optimal heuristic).