| Literature DB >> 29066078 |
Wouter van den Bos1, Rasmus Bruckner2, Matthew R Nassar3, Rui Mata4, Ben Eppinger5.
Abstract
In recent years, the application of computational modeling in studies on age-related changes in decision making and learning has gained in popularity. One advantage of computational models is that they provide access to latent variables that cannot be directly observed from behavior. In combination with experimental manipulations, these latent variables can help to test hypotheses about age-related changes in behavioral and neurobiological measures at a level of specificity that is not achievable with descriptive analysis approaches alone. This level of specificity can in turn be beneficial to establish the identity of the corresponding behavioral and neurobiological mechanisms. In this paper, we will illustrate applications of computational methods using examples of lifespan research on risk taking, strategy selection and reinforcement learning. We will elaborate on problems that can occur when computational neuroscience methods are applied to data of different age groups. Finally, we will discuss potential targets for future applications and outline general shortcomings of computational neuroscience methods for research on human lifespan development.Entities:
Keywords: Brain development; Computational neuroscience; Decision-making; Identification; Reinforcement learning; Risk-taking; Strategies
Mesh:
Year: 2017 PMID: 29066078 PMCID: PMC5916502 DOI: 10.1016/j.dcn.2017.09.008
Source DB: PubMed Journal: Dev Cogn Neurosci ISSN: 1878-9293 Impact factor: 6.464
Fig. 1A) Marr’s levels of analysis. B) Cartoon of a child choosing between two wheels of fortune. A risky option with a 75% chance of winning 20 Euro and 25% chance of losing 4 Euro versus a safe option with 100% chance of winning 14 Euro. Choosing the option with the highest outcome variance (the risky option) is often considered risk seeking behavior in the context of these tasks (even when total expected value is the same).
Fig. 2When presented with a simple binary choice gamble there are different strategies. The hardmax choice rule deterministically chooses the option with the higher expected value. Other strategies, such as minimax and maximax, only use part of the information that is presented. Maximax tries to maximize to maximum possible gain and minimax tries to minimize the maximum possible loss. Although they may yield seemingly similar behavior, they rely on different mental processes. Furthermore, even though these strategies may sometimes be captured by a parametric model such as cumulative prospect theory parameterizations (Pachur et al., 2017), they make different predictions about what happens on algorithmic and neural level (e.g., no representation of expected utility, no integration of possible outcomes).
Overview of typical problems and solutions when applying the strategy view.
| Problem | Possible Solution(s) |
|---|---|
| Use a pre-defined and constrained set of strategies that have been validated in past research ( | |
| Validate the use of several strategies using additional process data, such as search ( | |
| Adopt Bayesian modeling to help quantify the trade-off between flexibility in the number of hypothesized strategies and descriptive adequacy ( | |
| Explicitly model the mixture of strategies, and use parameter estimation to establish the contributions of individual strategies ( | |
| Design experimental paradigms so as to include critical tests that allow maximizing differences between hypothesized strategies ( | |
| Calculate power a priori based on hypothesized strategies and collect appropriately large samples | |
| Collapse across individuals in a meaningful way, for example, across similar strategy types, such as compensatory and non-compensatory strategies ( | |
| Design the task to elicit specific strategies ( | |
| Instruct or train individuals to execute specific strategies ( | |
| Design the statistical structure of the task to elicit specific strategies ( | |
| Compare different statistical structures and assess individual or developmental differences in adaptivity ( | |
| Adopt computational models that estimate strategy execution errors ( | |
| Estimate developmental effects on strategy selection and execution directly using choice vs. no-choice method, that is, comparing experimental conditions in which individuals can select (choice) or simply execute (no-choice) particular strategies ( |
Fig. 3Reinforcement learning (RL) and model-based fMRI analyses. A) Example of a two-armed bandit task. Participants are required to choose between the blue and red option, which is followed by a reward or a punishment. B) Reward distribution of the blue and the red option and application of an RL model. On average, the blue option is associated with reward, the red option is associated with punishment. Choices indicate that the model is able to learn that the blue option is associated with higher reward. As a consequence of a preference for the blue option, the model receives rewards on most trials. Across trials, the model approximately learns the expected value (EV) of both options. The prediction error (PE) indicates the difference between received rewards and EVs and can be utilized to adjust EVs. Finally, model-parameters can be used as parametric regressors in neuroimaging analyses. Note that the predicted BOLD signal of rewards and PEs can go in opposite directions.
Fig. 4Simulations using a reinforcement learning (RL) model with different learning rate but equal exploration parameters. A) Rapidly learning RL model with a learning rate α = 0.6 and exploration term θ = 5. B) Slowly learning model with a learning rate α = 0.05. C) Although models have different learning rate parameters, mean performance across 500 simulations with 30 trials each is similar. Error bars represent the standard deviation of the mean between the simulations.