| Literature DB >> 31664035 |
Romy Frömer1, Carolyn K Dean Wolf2, Amitai Shenhav3.
Abstract
When choosing between options, whether menu items or career paths, we can evaluate how rewarding each one will be, or how congruent it is with our current choice goal (e.g., to point out the best option or the worst one.). Past decision-making research interpreted findings through the former lens, but in these experiments the most rewarding option was always most congruent with the task goal (choosing the best option). It is therefore unclear to what extent expected reward vs. goal congruency can account for choice value findings. To deconfound these two variables, we performed three behavioral studies and an fMRI study in which the task goal varied between identifying the best vs. the worst option. Contrary to prevailing accounts, we find that goal congruency dominates choice behavior and neural activity. We separately identify dissociable signals of expected reward. Our findings call for a reinterpretation of previous research on value-based choice.Entities:
Mesh:
Substances:
Year: 2019 PMID: 31664035 PMCID: PMC6820735 DOI: 10.1038/s41467-019-12931-x
Source DB: PubMed Journal: Nat Commun ISSN: 2041-1723 Impact factor: 14.919
Fig. 1Overall value effects on RT are driven by goal congruency rather than reward value. a After evaluating each item in isolation (left), participants saw sets of four options and (in separate blocks) were instructed to choose either the best or the worst option (right). The same example is shown for both blocks but each choice sets was only viewed once in a session. b Top: A reward-based account predicts that RTs should decrease with overall value of the set, irrespective of the choice goal. Bottom: A goal congruency account predicts that RTs should decrease with overall value in Choose-Best blocks but instead increase with overall value in Choose-Worst Blocks. c Both Study 1 (behavioral) and Study 2 (fMRI) find the task-specific RT reversal predicted by a goal congruency account (see also Supplementary Study 1, Supplementary Discussion). Shaded error bars show 95% confidence intervals. d Our empirical findings were captured by an LCA model that took goal values (rather than reward values) as inputs
Comparison of overall reward vs overall goal value effects on log RT
|
| Study 1 | Study 2 | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
|
| |
|
| ||||||||||
| (Intercept) | 1.42 | 1.36–1.49 | 41.93 | 31 |
| 1.52 | 1.44–1.59 | 38.93 | 31 |
|
| Value Difference | −0.49 | −0.55–−0.43 | −16.48 | 3430 |
| −0.43 | −0.52–−0.34 | −9.18 | 31 |
|
| Overall Value | 0.05 | −0.01–0.11 | 1.61 | 3448 | 0.108 | −0.01 | −0.06–0.05 | −0.24 | 4222 | 0.809 |
| Best - Worst Condition | −0.07 | −0.14–0.00 | −1.90 | 31 | 0.066 | −0.01 | −0.09–0.06 | −0.33 | 31 | 0.740 |
| Overall Value: Best - Worst | −0.75 | −0.86–−0.63 | −12.66 | 3296 |
| −0.76 | −0.86–−0.65 | −14.39 | 3914 |
|
| (Intercept) | 1.42 | 1.36–1.49 | 41.97 | 31 |
| 1.52 | 1.44–1.59 | 38.93 | 31 |
|
| Value Difference | −0.49 | −0.55–−0.43 | −16.48 | 3430 |
| −0.43 | −0.52–−0.34 | −9.18 | 31 |
|
| Overall Goal Value | −0.37 | −0.43–−0.32 | −12.66 | 3296 |
| −0.38 | −0.43–−0.33 | −14.39 | 3914 |
|
| Overall Reward Value | 0.05 | −0.01–0.11 | 1.61 | 3448 | 0.108 | −0.01 | −0.06–0.05 | −0.24 | 4222 | 0.809 |
| Best - Worst Condition | −0.08 | −0.15–−0.01 | −2.21 | 31 |
| −0.03 | −0.11–0.04 | −0.85 | 31 | 0.402 |
Significant effects are highlighted in bold.
Model comparison for OVreward and OVgoal effects on RT across studies
|
| Study 1 | Study 2 | ||||||||
|---|---|---|---|---|---|---|---|---|---|---|
|
|
|
| X2 |
|
|
|
| X2 |
| |
| Model 0 (baseline): VD + C | 0.22 | 4112 | 0.19 | 4396 | ||||||
| Model 1: baseline + OV | 0.22 | 4112 | 0 | 2.72 | 0.099 | 0.19 | 4398 | 2 | 0.04 | 0.847 |
| Model 2: baseline + C | 0.26 | 3956 | −156 | 157.68 |
| 0.25 | 4197 | −202 | 203.31 |
|
| Model 3: baseline + OV | 0.26 | 3957 | 1 | 2.59 | 0.108 | 0.25 | 4195 | −2 | 0.06 | 0.809 |
For each study, models are compared sequentially, and dAIC is the difference in AIC of each model to the previous model. VD Value Difference, OV overall value, significant effects are highlighted in bold
Model comparison for reward and goal value – related BOLD activity in the valuation network ROI
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|
| Cb-w + RT (baseline) | 0.062 | 11976 | 12040 | |||
| GLM-1: baseline + RVreward + OVreward | 0.066 | 11969 | 12045 | −7 | 11 |
|
| GLM-2: baseline + RVgoal + OVgoal | 0.065 | 11963 | 12040 | −6 | 5.68 |
|
| baseline + RVgoal + OVgoal + OVreward | 0.069 | 11955 | 12038 | −8 | 10.2 |
|
| GLM-3: baseline + RVgoal + OVgoal + RVreward + OVreward | 0.069 | 11957 | 12046 | 2 | 0.2 | 0.657 |
RV Relative Value, OV = Overall Value, significant effects are highlighted in bold
Fixed effects summary GLM-3: Reward and goal values
|
|
|
|
|
|
|
|---|---|---|---|---|---|
| (Intercept) | 0.00 | −0.08–0.08 | 0.01 | 31.00 | 0.992 |
| Best - Worst Condition | −0.01 | −0.10–0.09 | −0.13 | 42.00 | 0.896 |
| Overall Reward Value | 0.21 | 0.08–0.34 | 3.15 | 3844.00 |
|
| Relative Reward Value | −0.03 | −0.16–0.10 | −0.44 | 4225.00 | 0.657 |
| Overall Goal Value | 0.16 | 0.03–0.29 | 2.44 | 1632.00 |
|
| Relative Goal Value | 0.22 | 0.09–0.36 | 3.37 | 4173.00 |
|
| RT | 0.02 | −0.08–0.12 | 0.37 | 32.00 | 0.715 |
Significant effects are highlighted in bold
Fig. 2The valuation network tracks goal values and overall, but not relative reward value. a Valuation network mask. b Mixed-effects regression coefficients show that the valuation network ROI defined a priori based on ref. [15]. tracks both overall and relative goal value, and also tracks overall (but not relative) reward value. Error bars show standard error of the mean. *p < 0.05, **p < 0.01
Fig. 3Reward and goal value dissociate across the striatum’s dorsal-ventral axis. a Whole brain results for relative goal value (blue) and overall reward value (green), thresholded at voxelwise p < 0.001 and cluster-corrected p < 0.05. Despite being correlated with activity in our a priori value network ROI, overall goal value did not survive the whole-brain threshold used for these follow-up exploratory analyses. b To interrogate our findings across regions of striatum, we selected independent bilateral ROIs previously used as seeds for distinct resting-state networks:[49] Dorsal Caudate (dark orange; [x, y, z] = ±12, 10, 8), Ventral Striatum, superior (orange; ±8, 10, 1), and Ventral Striatum, inferior (yellow; ±10, 11, −9). c Mixed-effects regression coefficients across these ROIs demonstrate a dorsal-ventral dissociation, with dorsal regions more sensitive to overall reward value and the ventralmost region more sensitive to relative goal value. Error bars show standard error of the mean. **p < 0.01, ***p < 0.001