Daisy Crawley1, Lei Zhang2,3,4, Emily J H Jones5, Jumana Ahmad1,6, Bethany Oakley1, Antonia San José Cáceres1,7, Tony Charman8,9, Jan K Buitelaar10,11,12, Declan G M Murphy1,9,13, Christopher Chatham4, Hanneke den Ouden10, Eva Loth1,13. 1. Department of Forensic and Neurodevelopmental Sciences, Institute of Psychiatry, Psychology & Neuroscience, King's College London, London, United Kingdom. 2. Institute of Systems Neuroscience, University Medical Center Hamburg-Eppendorf, Hamburg, Germany. 3. Neuropsychopharmacology and Biopsychology Unit, Department of Cognition, Emotion, and Methods in Psychology, Faculty of Psychology, University of Vienna, Vienna, Austria. 4. F. Hoffmann La Roche, Innovation Center Basel, Basel, Switzerland. 5. Centre for Brain and Cognitive Development, Birkbeck, University of London, London, United Kingdom. 6. Department of Psychology, Social Work and Counselling, University of Greenwich, London, United Kingdom. 7. Instituto de Investigación Sanitaria Gregorio Marañón, Departamento de Psiquiatría del Niño y del Adolescente, Hospital General Universitario Gregorio Marañón, Madrid, Spain. 8. Department of Psychology, Institute of Psychiatry, Psychology & Neuroscience, King's College London, London, United Kingdom. 9. South London and Maudsley NHS Foundation Trust (SLaM), London, United Kingdom. 10. Donders Institute for Brain, Cognition and Behaviour, Centre for Cognitive Neuroimaging, Radboud University, Nijmegen, the Netherlands. 11. Department of Cognitive Neuroscience, Radboud University Nijmegen Medical Center, Nijmegen, the Netherlands. 12. Karakter Child and Adolescent Psychiatry University Centre, Nijmegen, the Netherlands. 13. Sackler Institute for Translational Neurodevelopment, Institute of Psychiatry, Psychology & Neuroscience, King's College London, London, United Kingdom.
Abstract
Flexible behavior is critical for everyday decision-making and has been implicated in restricted, repetitive behaviors (RRB) in autism spectrum disorder (ASD). However, how flexible behavior changes developmentally in ASD remains largely unknown. Here, we used a developmental approach and examined flexible behavior on a probabilistic reversal learning task in 572 children, adolescents, and adults (ASD N = 321; typical development [TD] N = 251). Using computational modeling, we quantified latent variables that index mechanisms underlying perseveration and feedback sensitivity. We then assessed these variables in relation to diagnosis, developmental stage, core autism symptomatology, and associated psychiatric symptoms. Autistic individuals showed on average more perseveration and less feedback sensitivity than TD individuals, and, across cases and controls, older age groups showed more feedback sensitivity than younger age groups. Computational modeling revealed that dominant learning mechanisms underpinning flexible behavior differed across developmental stages and reduced flexible behavior in ASD was driven by less optimal learning on average within each age group. In autistic children, perseverative errors were positively related to anxiety symptoms, and in autistic adults, perseveration (indexed by both task errors and model parameter estimates) was positively related to RRB. These findings provide novel insights into reduced flexible behavior in relation to clinical symptoms in ASD.
Flexible behavior is critical for everyday decision-making and has been implicated in restricted, repetitive behaviors (RRB) in autism spectrum disorder (ASD). However, how flexible behavior changes developmentally in ASD remains largely unknown. Here, we used a developmental approach and examined flexible behavior on a probabilistic reversal learning task in 572 children, adolescents, and adults (ASD N = 321; typical development [TD] N = 251). Using computational modeling, we quantified latent variables that index mechanisms underlying perseveration and feedback sensitivity. We then assessed these variables in relation to diagnosis, developmental stage, core autism symptomatology, and associated psychiatric symptoms. Autistic individuals showed on average more perseveration and less feedback sensitivity than TD individuals, and, across cases and controls, older age groups showed more feedback sensitivity than younger age groups. Computational modeling revealed that dominant learning mechanisms underpinning flexible behavior differed across developmental stages and reduced flexible behavior in ASD was driven by less optimal learning on average within each age group. In autistic children, perseverative errors were positively related to anxiety symptoms, and in autistic adults, perseveration (indexed by both task errors and model parameter estimates) was positively related to RRB. These findings provide novel insights into reduced flexible behavior in relation to clinical symptoms in ASD.
Flexible behavior is a fundamental part of everyday life. It requires learning from feedback to guide decisions and adapting responses when feedback changes. These cognitive processes are implicated in a range of neurodevelopmental and neuropsychiatric conditions, including autism spectrum disorder (ASD; [1]), as well as attention-deficit hyperactivity disorder (ADHD) and anxiety, both of which frequently co-occur in ASD [2-5]. In particular, reduced flexible behavior is suggested to underpin core features of restricted, repetitive behaviors (RRB) in ASD, such as insistence on sameness. However, current evidence is inconclusive, and the mechanisms by which these impairments arise remain unclear [6, 7]. Studies of neurotypical individuals show that the cognitive processes underlying flexible behavior and reinforcement learning change through childhood and adolescence into adulthood [8, 9]. Therefore, a developmental approach within ASD that characterizes component learning processes is likely to bring us closer to understanding mechanisms of (in)flexible behavior and identifying therapeutic targets.Probabilistic reversal learning (PRL) paradigms require individuals to find a balance between learning structure in an uncertain environment while remaining flexible to change [10]. Typically, participants must learn using feedback which of a set of stimuli is most rewarded and adapt their responses when the rule changes, in order to maximize favorable outcomes. PRL paradigms therefore provide a direct assessment of flexible choice behavior (in addition to tapping reinforcement learning), as they require information to be integrated over a number of trials in order to detect true changes, and—much like interacting with our environment—this trial-and-error learning is continually updated throughout the task. Furthermore, PRL paradigms do not require tracking of extradimensional shifts, thereby constraining the recruitment of additional cognitive domains [11, 12].Previous literature has reported reduced reversal learning in ASD relative to controls and a positive relationship between reversal errors and RRB [1, 13]. In contrast, others have reported poorer overall task performance but unspecific to reversal adaptation [14, 15], or no differences in reversal learning nor any associations with ASD symptomatology [16, 17]. It is worth noting that these inconsistencies in ASD-related changes in cognitive flexibility are also reflected in the broader literature using alternative paradigms (see [7, 18] for reviews).With respect to reinforcement learning, studies of reward processing suggest atypical or diminished neural responses to rewards in ASD [19-22], though results from adolescent studies are less consistent [23-25]. If reinforcement is differentially experienced in ASD, it is likely to impact on decision-making processes and behavior. In addition to establishing differences, associations between learning and phenotypic correlates warrant further study in order to elucidate whether such differences necessarily manifest in impairments related to symptom severity.Several factors may have contributed to inconsistencies in the literature. First, previous studies have often studied single age groups or a broad age range within a small sample size. Evidence from both cognitive and neuroimaging studies attests to important developmental differences in reinforcement learning and flexible behavior in neurotypical individuals [26-28]. Young children often perseverate, taking longer than older children to learn new rules and switch their responses [8]. During adolescence, notable changes in goal-directed decision-making occur, often manifesting in risky decisions thought to be attributable to hypersensitivity to rewards [29-31]. In adulthood, there is evidence for the use of more sophisticated, “controlled” cognitive strategies [32, 33]. Hence, a developmental approach in ASD is needed to ascertain whether potential impairments reflect delayed development or atypical cognitive processes.Second, previous studies have also tended to use task performance measures that often aggregate error scores and do not directly characterize learning processes governing behavior. Computational models capture the dynamics of learning over time—emulating a participant’s experience—and delineate component processes underlying PRL by approximating mechanisms that may have led to task behavior. Estimating and comparing different reinforcement learning models allows for the evaluation of competing mechanisms by quantifying how likely each model is to have generated the observed behavior. Moreover, by approximating putative mechanisms, computational models enable better mapping between behavior and neurobiology, particularly important for understanding neurodevelopmental disorders [34].Studies of ASD using modeling have shown evidence of slower, faster, and equal rates of learning compared to neurotypical individuals. Optimal learning rates depend on the stability of the task environment. A changeable environment requires fast learning guided by recent feedback, whereas a stable environment requires slower learning over time (e.g., [35, 36]). Crucially, probabilistic feedback also requires learning to ignore “misleading” punishment. Previously, autistic adults were shown to have a slower learning rate than neurotypical adults when using higher-probability reward contingencies, but they performed comparably or outperformed neurotypical adults when the contingency was near chance [21, 22]. Perhaps, then, a key difficulty lies in learning regularities and ignoring irregularities, in addition to learning change per se [37]. This is consistent with previous findings of a tendency to “overlearn” volatility in ASD adults, resulting in reduced learning of probabilistic errors [38]. Whether these findings extend to children and adolescents (see [39] for differing findings) and which underlying processes are different in ASD remain to be seen.Here, we examined learning processes underlying flexible behavior in ASD and typical development (TD) across developmental stages using a PRL paradigm. Our secondary aim was to investigate possible relationships with symptomatology in ASD. To achieve this, we (1) tested a large sample of individuals with a wide age range that was sufficiently powered to compare children, adolescents, and adults and (2) used reinforcement learning models to compare quantitative mechanistic explanations of flexible behavior and identify the latent processes on which individuals may differ. We included measures of RRB subtypes as our focus, social-communication difficulties for comparison, and associated symptoms of ADHD and anxiety as frequently co-occurring features that may also relate to atypical learning and flexible behavior. Based on previous literature, we hypothesized that younger age groups would perform less well on the task than older age groups and that autistic individuals would perform less well than neurotypical individuals. Additionally, we hypothesized differences in dominant underlying cognitive processes across development. Finally, we predicted that reduced flexible behavior would be related to higher RRB symptom severity, in particular behavioral rigidity/insistence on sameness.
Methods
Ethics statement
The study was approved by the independent local ethics committees of the participating centers (London Queen Square Health Research, Authority Research Ethics Committee: 13/LO/56; Radboud University Medical Centre Institute Ensuring Quality and Safety Committee on Research Involving Human Subjects Arnhem-Nijmegen: 2013/455; UMM University Medical Mannheim, Medical Ethics Commission II: 2014–540 N-MA; University Campus Bio-Medical Ethics Committee of Rome: 18/14 PAR ComET CBM) and conducted according to the principles expressed in the Declaration of Helsinki. Written informed consent was obtained from all participants and/or their parent/guardian (when appropriate) prior to the study.
Participants
This study was part of the EU-AIMS Longitudinal European Autism Project (LEAP; [40, 41])—a multidisciplinary, multicenter study of children (6–11 years), adolescents (12–17 years), and adults (18–30 years) with and without ASD from six European sites. The current study included data from 321 individuals with an existing clinical diagnosis of ASD and 251 typically developing (TD) individuals, with full-scale IQ scores ranging from 74 to 148. Descriptive statistics for the sample are listed in Table 1. Full-scale IQ was measured using the Wechsler scales (see [41]). Although ASD individuals were additionally assessed using the Autism Diagnostic Observation Schedule [42, 43] and Autism Diagnostic Interview-Revised (ADI-R, [44]), reaching instrument cutoffs were not inclusion criteria, as clinical judgment has been found to consistently improve diagnostic stability [45]. However, task behavioral analyses were repeated in a subset of individuals who meet ADI-R criteria as specified by [46] (S1 Table). Although the full EU-AIMS LEAP sample includes individuals with mild intellectual disabilities (N = 83), initial analyses showed evidence of poor task learning in this group, and thus they were omitted from further analyses. Those with only partial data (N = 3) or who chose the same stimulus throughout the task (N = 1) were excluded from analysis (see S1 Text for further sample information).
Table 1
Participant characteristics (total N = 572): Mean (SD), [N] if missing data, unless otherwise stated.
Characteristic
Children
Adolescents
Adults
Total sample
ASD
TD
ASD
TD
ASD
TD
ASD
TD
N
81
64
114
90
126
97
321
251
Sex (Percentage male)
70.37
60.94
76.32
68.89
69.84
72.16
72.27
68.13
Age in years
9.59 (1.50)
9.52 (1.54)
14.94 (1.71)
15.39 (1.71)
22.80 (3.55)
23.25 (3.29)
16.67 (5.92)
16.93 (6.02)
Full-scale IQ
105.54 (14.35)
111.81 (12.50)
101.81 (15.92)
106.69 (13.32)
103.97 (15.21)
109.14 (12.29)
103.60 (15.28)
108.95 (12.82)
ADI-R RRB
4.46 (2.89)[79]
-
4.30 (2.68)[112]
-
4.07 (2.54)[116]
-
4.25 (2.68)[307]
-
RBS-R Stereotyped Behavior
3.83 (3.33)[71]
0.19 (0.68)[54]
3.64 (3.97)[96]
0.14 (0.62)[69]
1.86 (2.92)[91]
-
3.06 (0.16)[258]
0.16 (0.63)[129]
RBS-R Ritualistic-Sameness
7.48 (5.52)[71]
0.35 (0.91)[54]
7.39 (6.26)[96]
0.41 (1.31)[69]
4.79 (4.44)[91]
-
6.50 (5.59)[258]
0.36 (1.12)[129]
ADI-R Social Interaction
15.14 (6.8)[79]
-
17.46 (6.59)[112]
-
14.78 (6.80)[116]
-
15.85 (6.81)[307]
-
ADI-R Communication
13.32 (5.56)[79]
-
13.48 (5.56)[112]
-
11.82 (5.67)[116]
-
12.81 (5.64)[307]
-
SRS-2 SCI
73.44 (11.19)[73]
44.60 (5.10)[55]
74.67 (10.89)[93]
45.35 (6.05)[71]
64.32 (10.89)[87]
-
70.75 (11.90)[253]
44.97 (5.58)[132]
ADHD hyper/impulsive parent-report
4.33 (2.93)[72]
0.37 (1.17)[52]
2.77 (2.77)[96]
0.20 (0.84)[71]
1.33 (1.80)[94]
-
2.68 (2.77)[262]
0.25 (0.97)[130]
ADHD inattentive parent-report
5.25 (3.00)[72]
0.62 (1.60)[52]
4.77 (3.12)[96]
0.89 (1.81)[71]
3.23 (3.20)[94]
-
4.35 (3.22)[262]
0.76 (1.70)[130]
ADHD hyper/impulsive self-report
-
-
-
-
1.61 (1.99)[96]
0.61 (1.43)[72]
1.61 (1.99)[96]
0.61 (1.43)[72]
ADHD inattentive self-report
-
-
-
-
2.91 (2.38)[96]
0.81 (1.51)[72]
2.91 (2.38)[96]
0.81 (1.51)[72]
Anxiety (BAI/BYI-IIa)
14.62 (8.77)[72]
6.00 (4.97)[51]
14.13 (10.05)[61]
8.67 (7.05)[64]
14.97 (13.24)[97]
4.27 (5.1)[73]
-
-
Task behavior
Accuracy (overall)
0.65 (0.11)
0.68 (0.13)
0.67 (0.13)
0.76 (0.14)
0.73 (0.15)
0.77 (0.14)
0.69 (0.14)
0.74 (0.14)
PerErrors
0.28 (0.14)
0.26 (0.15)
0.30 (0.18)
0.23 (0.18)
0.27 (0.20)
0.21 (0.16)
0.28 (0.18)
0.23 (0.16)
Win-stay
0.69 (0.16)
0.70 (0.16)
0.72 (0.16)
0.81 (0.15)
0.80 (0.16)
0.84 (0.15)
0.75 (0.17)
0.79 (0.16)
Lose-shift
0.55 (0.11)
0.53 (0.14)
0.50 (0.15)
0.43 (0.17)
0.45 (0.18)
0.41 (0.19)
0.49 (0.16)
0.45 (0.18)
a Parent-report for children, self-report for adults and adolescents.
Abbreviations: ADHD, attention-deficit hyperactivity disorder; ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder; BAI, Beck Anxiety Inventory; BYI-II, Beck Youth Inventories–Second Edition; hyper/impulsive, hyperactivity/impulsivity; Lose-shift, changing the response following punishment as a proportion of total lose trials; PerErrors, perseverative errors, expressed as a proportion of reversal trials; RBS-R, Repetitive Behavior Scale-Revised; SD, standard deviation; SRS-2 SCI, Social Responsiveness Scale 2nd Edition Social Communication Index; TD, typical development; Win-stay, repeating the previous choice following reward expressed as a proportion of total win trials
a Parent-report for children, self-report for adults and adolescents.Abbreviations: ADHD, attention-deficit hyperactivity disorder; ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder; BAI, Beck Anxiety Inventory; BYI-II, Beck Youth Inventories–Second Edition; hyper/impulsive, hyperactivity/impulsivity; Lose-shift, changing the response following punishment as a proportion of total lose trials; PerErrors, perseverative errors, expressed as a proportion of reversal trials; RBS-R, Repetitive Behavior Scale-Revised; SD, standard deviation; SRS-2 SCI, Social Responsiveness Scale 2nd Edition Social Communication Index; TD, typical development; Win-stay, repeating the previous choice following reward expressed as a proportion of total win trials
Experimental paradigm
Participants completed a computerized PRL task whereby they were instructed to choose one of two colored shapes (vertical yellow bars or horizontal blue bars) presented in two of four possible locations with an 80:20 reward/punishment contingency (Fig 1A). Positive feedback consisted of green, smiling emoticons and negative feedback of red, frowning emoticons (i.e., reward/punishment) and accompanying sounds (bell chime/buzzer, respectively). The task employed a pseudorandom fixed sequence comprising 80 trials with a reversal midway. Participants’ first stimulus choice was considered correct in the acquisition phase; after the reversal, the initially incorrect stimulus became the usually rewarded stimulus and vice versa (Fig 1B and 1C). To reduce task demand and avoid potential floor effects in the younger age groups or clinical sample, the contingency ratio was higher than some previous studies (70:30; [10, 47]). Participants used arrow keys to respond and had unlimited response time per trial (see S1 Text for task instructions). This paradigm has previously been used in neurotypical individuals and other clinical groups [47, 48] and was specified by the European Medicines Agency in their letter of support for EU-AIMS LEAP [49].
Fig 1
Task presentation and pooled task behavior.
(A) An example of several consecutive trials—on each trial, participants have to choose between two stimuli, presented pseudorandomly in two of the four possible locations. Feedback is received in the form of a smiling green face (positive) or a sad red face (negative) and is probabilistic, meaning that some is “misleading” (e.g., trial 3). Win-stay trials are those in which individuals repeat their stimuli choice following positive feedback (e.g., trials 2 and 3), and lose-shift trials are those in which individual change their stimuli choice following negative feedback (e.g., trials 4 and 5). (B) The structure of the task—the first stimuli chosen by each participant is correct in the acquisition phase (trials 1–40; here: yellow). Feedback was given with an 80:20 reward/punishment ratio; green blocks indicate reward and red blocks indicate punishment. In the reversal phase (trials 41–80), the true correct stimulus is reversed (here: blue) as is the contingency schedule. (C) Overall trial-by-trial behavior—All participants’ data, sorted by performance, with average performance overlaid (black line) regardless of diagnosis or age group. Compare to (B) to see how task structure is experienced in practice (see S1 Data).
Task presentation and pooled task behavior.
(A) An example of several consecutive trials—on each trial, participants have to choose between two stimuli, presented pseudorandomly in two of the four possible locations. Feedback is received in the form of a smiling green face (positive) or a sad red face (negative) and is probabilistic, meaning that some is “misleading” (e.g., trial 3). Win-stay trials are those in which individuals repeat their stimuli choice following positive feedback (e.g., trials 2 and 3), and lose-shift trials are those in which individual change their stimuli choice following negative feedback (e.g., trials 4 and 5). (B) The structure of the task—the first stimuli chosen by each participant is correct in the acquisition phase (trials 1–40; here: yellow). Feedback was given with an 80:20 reward/punishment ratio; green blocks indicate reward and red blocks indicate punishment. In the reversal phase (trials 41–80), the true correct stimulus is reversed (here: blue) as is the contingency schedule. (C) Overall trial-by-trial behavior—All participants’ data, sorted by performance, with average performance overlaid (black line) regardless of diagnosis or age group. Compare to (B) to see how task structure is experienced in practice (see S1 Data).
Analysis of task behavior
Behavioral performance on the task was assessed using accuracy during acquisition and reversal phases, perseverative errors, and win/lose feedback sensitivity. Accuracy was quantified as the proportion of correct responses. Perseverative errors were defined as two or more consecutive errors during the reversal phase—i.e., trials in which the participant chose the previously rewarded stimulus, despite negative feedback—and are reported as a proportion of reversal phase trials. Win-stay and lose-shift behaviors index the effect of an outcome on the subsequent choice. They are defined, respectively, as repeating the previous choice following reward (as a proportion of total rewarded trials) and changing the response following punishment (as a proportion of total punished trials). As in previous studies using this task [10, 47, 48, 50, 51], reaction time is not examined here because it is unlikely to capture task-relevant processes, since no response speed instructions are given nor is there a time limit for responding (see S1 Fig for further discussion).
Reinforcement learning models
We compared three reinforcement learning models to examine different computational mechanisms driving information integration and the cognitive processes underlying learning and flexible adaptation. Each model extends the Rescorla-Wagner value update rule [52] but in different ways in terms of how information is integrated. The Rescorla-Wagner update rule assumes that individuals assign and update internal stimulus value signals based on the prediction error, i.e., the mismatch between outcome (received reward/punishment following choice of this stimulus) and prediction (expected value of choosing this stimulus). Below, we omit results from the original Rescorla-Wagner model, as all other models consistently outperformed it (see S1 Text and S2 Table).
(1) Counterfactual update model
Previous studies suggest individuals may use counterfactual updating in reversal learning tasks, as it captures the anti-correlatedness of the choice stimuli (i.e., where one is correct, the other is incorrect; [53, 54]). The counterfactual update (CU) model extends the standard Rescorla-Wagner algorithm by updating the value of both choice stimuli.
Here, the value V of both the chosen c and unchosen nc stimulus are updated with the actual prediction error and the counterfactual prediction error per trial t, respectively. O is the outcome received. The learning rate η evidences the magnitude of the value update affected by both prediction errors—put simply, the speed of learning. In this framework, reduced flexible behavior may be underpinned by too frequent response switches quantified by excessive value updating after punishment.
(2) Reward-punishment model
Alternatively, reduced flexible task behavior may result from reduced punishment learning. Reduced punishment learning would have a disproportionate effect during the reversal phase because punishments following choices of the previously rewarded stimulus would have a diminished influence on choice behavior due to a failure to devalue this stimulus. To assess whether this mechanism drives reduced flexible behavior, we use a different extension of the Rescorla-Wagner model, with separate learning rates for reward and punishment (reward-punishment model [R-P]; [47]). This allows for the capture of differential learning to feedback types.
Here, ηrew is the learning rate for rewards and ηpun is the learning rate for punishment; O is the outcome received. In this model, only the chosen stimulus value is updated.
(3) Experience-weighted attraction–dynamic learning rate model (EWA-DL)
Finally, reduced flexible behavior may result from a growing insensitivity to novel information. By this mechanism, a failure to update values based on new information (i.e., accumulating negative feedback denoting a true reversal) would cause perseveration of the previously rewarded response and delayed or even complete failure to switch. We examined this mechanism using the experience-weight parameter from a reduced version of the EWA model as presented in previous work [47], where we used the formulation of a nonstationary learning rate through updating of an experience weight. This dynamical learning rate allows for interpolation between different forms of updating (accumulating versus averaging rho shifts from 0 to 1). Note that we do not use the exact same model of the original EWA model [55], as we omit the feature of blending belief-based versus reinforcement learning. To make this distinction clear, we have labeled this model as EWA-DL (but note that it is the identical model to [47]). The EWA-DL model extends classic reinforcement learning with an experience-weight parameter that captures the attribution of significance to past experience over and above new information as an individual progresses through the task. This effectively reduces the learning rate over time. Thus, in this context, perseveration would arise from a slowness, after reversal, to update the value of the now usually rewarded stimuli due to an overreliance on preceding task experience. The growth of the experience weight n and update of the stimulus values V are defined as follows:
Here, n is the “experience weight” of the chosen stimulus on trial t, which is updated on every trial using the experience decay factor ρ. V is the value of choice c on trial t for outcome O received in response to that choice, and φ is the decay factor for the previous payoffs. In this model, φ is equivalent to the inverse of the learning rate in Rescorla-Wagner models (or alternatively, n = 1 –φ; see also [47]). For ρ > 0, the experience weights promote more sluggish updating with time. Previous work has shown the EWA-DL to be the winning model in neurotypical adults in the same PRL task [47].
Softmax action selection
For all models, a softmax choice function was used to compute the action probability given the action values. On each trial t, the action probability of choosing option A (over B) was defined as follows:
Here, β (0 < β < 5) is the inverse temperature parameter that governs the stochasticity of the choice, computed using inverse logit transfer. We set the upper bound to 5, as individual parameters are regularized by group-level parameters that prevent extreme parameter estimates (see parameter estimation section), and our data indeed showed that all β estimates are smaller than 5. We refer to β in this paper as value sensitivity, as it reflects sensitivity to the difference in stimulus values, that is, the degree to which a (perceived) difference in stimulus values determines choice (see S1 Text). Higher β values denote decisions driven by relative value whereas lower β values denote more choice stochasticity. Additionally, a small indifference point parameter α (−0.5 < α < 0.5) is introduced, which captures any selection bias in which both options are equally likely to be selected. Including this indifference point parameter systematically improved performance of all models. The action probability of options A and B by definition sum to 1: p(B) = 1 – p(A).
Parameter estimation and model selection/validation
Parameter estimation was performed with hierarchical Bayesian analysis (HBA) using Stan language in R (RStan; [56, 57]), adopted from the hBayesDM package [58]. Posterior inference was performed using Markov chain Monte Carlo (MCMC) sampling in RStan. The models were fit separately for each of six groups—diagnosis (ASD, TD) × developmental stage (children, adolescents, adults)—and compared within each group to assess how well they fit the data (goodness-of-fit) while accounting for model complexity. Comparison of model fit was assessed per group using Bayesian bootstrap and model averaging, whereby log-likelihoods for each model were evaluated at the posterior simulations and a weight obtained for each model. Model weights include a penalizing term for model complexity and a normalizing term according to the number of models being compared; thus, for each group, model weights sum to 1 [59]. Higher model weight indicates better model fit. We conducted model recovery analyses, and, for completeness, we also ran model fitting across age groups (see S1 Text). Finally, we established that the winning models could replicate the observed behavior using one-step-ahead prediction (e.g., [60]). Here, parameters are drawn from the joint posterior distribution and combined with the outcome sequence to predict future choices thereby quantifying absolute model fit. That is, we let the model take random draws from each participant’s joint posterior distribution to generate choices. We iterated this procedure as many times as the number of samples (i.e., 4,000) per trial per participant. We implemented two ways to assess posterior predictions. First, we computed the predictive accuracy using the number of correct predictions divided by the total number of iterations and tested if this accuracy was significantly better than chance level (i.e., 50%). Second, we analyzed the generated data in the same way as we analyzed the observed data and compared whether results from generated data captured the behavioral pattern in our behavioral analysis (for further details on model specification and validation, see S1 Text).
Optimal learning parameters
We identified the optimal learning parameters for each model using simulation. Taking the CU model as an example, we first took the learning rate from a grid with 1,000 steps from 0 to 1 and then simulated choice data for every learning rate. We computed how often the simulated choice data matched the correct option (i.e., the more rewarding option). We repeated this simulation 10,000 times and identified the optimal learning rate as the value that resulted in the highest choice accuracy. We used the same procedure to determine the optimal learning parameter(s) for the R-P model and the EWA-DL.
Clinical measures
ASD symptomatology
Two measures were used to assess RRB symptom severity in ASD: (1) The ADI-R [44] is a structured parent/caregiver interview comprising 93 questions assessing most severe/early developmental ASD symptoms, which yields an algorithm score for RRB based on 12 items; (2) The Repetitive Behavior Scale-Revised (RBS-R; [61]) is a 43-item parent-report questionnaire tapping current RRB, which typically yields a total score and five subscales [62]. Here, we use the Ritualistic-Sameness and Stereotyped Behavior subscales as the best indices of behavioral rigidity (see S3 Table for a comparison of all subscales). To examine whether relationships were specific to RRB, ADI-R domain scores for Communication and Reciprocal Social Interaction were included, as were T-scores for the Social Communication Index on the Social Responsiveness Scale 2nd Edition (SRS-2; [63])—a parent-report questionnaire assessing current social-communication difficulties. On all measures, higher scores indicate greater symptom severity.
Comorbid symptomatology
The DSM-5 rating scale of ADHD [64] and the Beck Anxiety Inventory (BAI; [65]) were used to assess associated symptoms. For ADHD symptoms, parents of all ASD participants completed the parent-report form, and in addition, ASD adults completed the self-report form. For anxiety, adult participants completed the BAI in self-report form, whereas adolescents completed the self-report version of the anxiety subscale of the Beck Youth Inventories (BYI-II; [66]). Parents/caregivers of children completed the same BYI-II subscale in parent-report form.
Statistical analysis
All analyses were conducted in R [67]. First, we characterized the cohort with respect to sex, age, and IQ differences. Second, to examine the effects of diagnosis and age group on the task performance measures, we employed linear mixed-effects models using the lme4 package in R [68]. The models included diagnosis and age group (and for accuracy, phase) as between-participant factors (including their interaction[s]) and site as a random factor. Including sex in the models did not improve model fit. Post hoc pairwise comparisons were computed from contrasts between factors using lsmeans package with Tukey adjustments [69]. Following the reinforcement learning model comparisons and validation using one-step-ahead predictions, we examined case-control differences on winning model parameters in each age group. Finally, we used correlational analyses to examine associations between task behavior, model parameters, and symptomatology. Symptomatology associations were conducted only in the ASD groups using Spearman’s correlations owing to non-normality in scores. Significance thresholds for correlational analyses are Bonferroni-corrected for multiple comparisons—children/adolescents (.05/11): p = .0045 and adults (.05/13): p = .0038. Effect sizes are reported as Cohen’s d.
Results
Sex, age, and IQ group differences
Diagnostic groups did not differ on sex or age, either overall or within each age group (all p > .1). However, all groups differed significantly on full-scale IQ, with TD groups scoring higher than ASD groups (p ranging .01–.005; d ranging 0.32–0.47). Therefore, for all further group comparisons, we assessed whether results changed with IQ as a confound regressor, and, in addition, we conducted analyses of task behavior in an IQ-matched subsample (S2 Text and S4 Table). Results were largely unchanged throughout (see S2 Text and S2 Fig).
Task behavior
Grouped trial-by-trial behavior is shown in Fig 2A and descriptive statistics in Table 1. All diagnostic and age groups performed above chance in both phases of the task, showing task comprehension (all p < 2.2 × 10−16; see S3 Text, S3 Fig and S5 Table). A repeated-measures analysis of accuracy showed significant main effects of phase (F[1,566] = 294.25, p < 2.2 × 10−16), diagnosis (F[1,566] = 21.96, p = 9.52 × 10−8), and age group (F[,566] = 16.64, p = 3.49 × 10−6) but no significant interactions (all p > .1). Post hoc analyses revealed accuracy was on average significantly higher (1) in the acquisition phase than in the reversal phase, reflecting the challenge of flexible adaptation (p < .0001, d = 0.82); (2) in TD individuals compared to ASD individuals (p < .0001, d = 0.29); and (3) in older age groups compared to younger age groups (adults-adolescents, p = .0113, d = 0.22; adults-children, p < .0001, d = 0.51; adolescents-children, p = .0062, d = 0.29; Fig 2B).
Fig 2
Task behavior.
(A) Trial-by-trial data for each age group with diagnostic group averages overlaid. More evidence of task understanding in adults, as indicated by more correct task behavior and steeper shifts at reversal in comparison to children. (B) Task accuracy was greater (1) in the acquisition phase compared to the reversal phase, (2) in older age groups compared to younger, and (3) in TD individuals compared to ASD individuals. (C-E) Linear mixed-effects models showed a main effect of diagnosis for all three task performance measures (perseverative errors, win-staying, lose-shifting) and a main effect of age for win-staying (D) and lose-shifting (E) but not perseverative errors (C). For win-staying, a diagnosis × age group interaction was also found. Post hoc tests revealed ASD adolescents showed significantly reduced win-staying compared with TD adolescents (D), ***p < .001 (see S1 Data). ASD, autism spectrum disorder; TD, typical development.
Task behavior.
(A) Trial-by-trial data for each age group with diagnostic group averages overlaid. More evidence of task understanding in adults, as indicated by more correct task behavior and steeper shifts at reversal in comparison to children. (B) Task accuracy was greater (1) in the acquisition phase compared to the reversal phase, (2) in older age groups compared to younger, and (3) in TD individuals compared to ASD individuals. (C-E) Linear mixed-effects models showed a main effect of diagnosis for all three task performance measures (perseverative errors, win-staying, lose-shifting) and a main effect of age for win-staying (D) and lose-shifting (E) but not perseverative errors (C). For win-staying, a diagnosis × age group interaction was also found. Post hoc tests revealed ASD adolescents showed significantly reduced win-staying compared with TD adolescents (D), ***p < .001 (see S1 Data). ASD, autism spectrum disorder; TD, typical development.Next, a significant main effect of diagnosis on perseverative errors was observed (F[1,565.42] = 11.07, p = .0009, d = 0.30; Fig 2C), such that ASD individuals made on average significantly more perseverative errors than TD individuals; however, there was no significant effect of age nor interaction between diagnosis and age group (p > .2). For both accuracy and perseverative errors, results were unchanged both in the IQ-matched subsample and with IQ as a confound regressor (S2 Text and S2 Fig).Regarding feedback sensitivity, ASD individuals showed on average significantly less win-stay and more lose-shift behavior relative to TD individuals, and for both there was a main effect of age (win-stay: diagnosis [F(1,563.28) = 12.06, p = .0006, d = 0.24], age group [F(2, 521.29) = 27.78, p = 3.4 × 10−12]; lose-shift: diagnosis [F(1, 564.28) = 9.86, p = .0018, d = 0.23], age group [F(2,390.88) = 19.50, p = 8.5 × 10−9]). Pairwise post hoc comparisons revealed win-staying increased and lose-shifting decreased with age (Fig 2D and 2E). For win-stay behavior, the predicted interaction between diagnosis and age group was approaching significance (p = .057). A between-diagnosis group analysis of each age group revealed ASD adolescents showed less win-staying than TD adolescents (p < .0008; Fig 2D, d = 0.54), which survived Bonferroni correction (correcting for task behavioral measures × age groups: p-value = .05/[3 × 3] = .0056). For lose-shift behavior, there was no significant interaction between diagnosis and age group (p = .3). Results were again consistent in the IQ-matched subsample and when IQ was entered as a confound regressor (S2 Text and S2 Fig).The pattern of results reported here is also replicated in the additional analyses conducted with the subset of ASD individuals who meet ADI-R criteria (S2 Text and S2 Fig).
Model comparison and validation
Model weightings are shown in Fig 3A, and all winning model’s parameters had independent contributions (S4 Fig). There were no between-diagnosis group differences in terms of model preference, only changes across development. Within both ASD and TD age groups, model weights showed that for children, the CU model provided the highest model evidence; for adolescents, the R-P model provided the highest model evidence; and for adults, the EWA-DL provided the highest model evidence. Results were unchanged when models were fitted with (z-scored) IQ as a covariate (see S6 Table). Model recovery results showed that all models’ identities can be well recovered (S5 Fig). Collapsing age groups, the R-P model provided the highest model evidence in both diagnostic groups (S7 Table). One-step-ahead predictions of each group’s winning model showed the models captured the key features of task behavior (e.g., the first response to negative feedback, the switch at reversal), with posterior predictive accuracy values of 0.61 and above. All models performed significantly better than chance level (p ≤ 1.23 × 10−11). Average simulated behavior closely resembled participants’ behavior (Fig 3B).
Fig 3
Model comparisons, validations, and parameters.
(A) Evidence (model weights) for models within each diagnostic and age group. Very similar patterns are observed for TD and ASD groups; winning models for children, adolescents, and adults are the CU, R-P, and EWA-DL, respectively. (B) One-step-ahead posterior predictions for each age and diagnostic group according to winning models. Colored lines indicate diagnostic-group-averaged trial-by-trial task behavior; shaded areas indicate 95% HDI of the one-step-ahead simulation using the entire posterior distribution. Compare with actual task data in Fig 2A. Posterior predictive accuracies are also indicated on each plot (ASD: red; TD: blue). (C) Model parameter comparisons. Within each winning model and thus age group, parameter estimates were compared between diagnostic groups: (1) ASD children showed a significantly higher learning rate (η) than TD children, in which simulations showed the optimal learning rate to be 0.18; (2) ASD adolescents showed a significantly lower reward learning rate than TD adolescents, but no difference between punishment learning rates was observed; (3) ASD adults showed significantly lower φ than TD adults, the optimal value was shown to be 0.85 in simulations, and ASD adults also showed significantly greater experience decay (ρ) than TD adults, suggesting great perseveration. (D) Learning rate simulations showing optimal learning rates for each model (Counterfactual update, compare to Fig 3C Children; Rew-Pun, compare to Fig 3C Adolescents—Learning rate; EWA, Experience-weighted attraction-dynamic learning rate, compare to Fig 3C Adults—Inverse learning rate). ***p < .001, **p < .01, *p < .05; Δ indicates group mean (see S1 Data). ASD, autism spectrum disorder; CU, counterfactual update; d, Cohen’s d model; EWA-DL, experience-weighted attraction–dynamic learning rate model; HDI, highest density interval; R-P, reward-punishment model; Rew-Pun, reward-punishment; RL, reinforcement learning; RW, Rescorla-Wagner; TD, typical development.
Model comparisons, validations, and parameters.
(A) Evidence (model weights) for models within each diagnostic and age group. Very similar patterns are observed for TD and ASD groups; winning models for children, adolescents, and adults are the CU, R-P, and EWA-DL, respectively. (B) One-step-ahead posterior predictions for each age and diagnostic group according to winning models. Colored lines indicate diagnostic-group-averaged trial-by-trial task behavior; shaded areas indicate 95% HDI of the one-step-ahead simulation using the entire posterior distribution. Compare with actual task data in Fig 2A. Posterior predictive accuracies are also indicated on each plot (ASD: red; TD: blue). (C) Model parameter comparisons. Within each winning model and thus age group, parameter estimates were compared between diagnostic groups: (1) ASD children showed a significantly higher learning rate (η) than TDchildren, in which simulations showed the optimal learning rate to be 0.18; (2) ASD adolescents showed a significantly lower reward learning rate than TD adolescents, but no difference between punishment learning rates was observed; (3) ASD adults showed significantly lower φ than TD adults, the optimal value was shown to be 0.85 in simulations, and ASD adults also showed significantly greater experience decay (ρ) than TD adults, suggesting great perseveration. (D) Learning rate simulations showing optimal learning rates for each model (Counterfactual update, compare to Fig 3C Children; Rew-Pun, compare to Fig 3C Adolescents—Learning rate; EWA, Experience-weighted attraction-dynamic learning rate, compare to Fig 3C Adults—Inverse learning rate). ***p < .001, **p < .01, *p < .05; Δ indicates group mean (see S1 Data). ASD, autism spectrum disorder; CU, counterfactual update; d, Cohen’s d model; EWA-DL, experience-weighted attraction–dynamic learning rate model; HDI, highest density interval; R-P, reward-punishment model; Rew-Pun, reward-punishment; RL, reinforcement learning; RW, Rescorla-Wagner; TD, typical development.
Within-model diagnostic group comparisons
We then investigated which computational mechanisms underpin poorer task performance in ASD for the different age groups. To this end, we compared diagnostic groups on parameter estimates from the winning model of each age group (Table 2; see also S4 Text).
Table 2
Model parameters for each age and diagnosis group’s winning model and within age-group comparisons.
Mean (SD)
Highest Density Interval (of MCMC)
d
p value
ASD
TD
ASD
TD
Children – Counterfactual update
η
0.258 (0.126)
0.193 (0.087)
[0.206, 0.311]
[0.150, 0.235]
0.600
0.0003
β
0.979 (0.783)
1.202 (0.892)
[0.886, 1.073]
[1.063, 1.340]
-0.266
0.117
α
-0.014 (0.319)
-0.042 (0.153)
[-0.092, 0.069]
[-0.116, 0.025]
0.114
0.482
Adolescents – Reward-punishment
ηrew
0.368 (0.169)
0.443 (0.223)
[0.268, 0.466]
[0.359, 0.536]
-0.382
0.0039
ηpun
0.336 (0.098)
0.311 (0.116)
[0.265, 0.402]
[0.264, 0.356]
0.231
0.671
β
1.494 (0.897)
2.535 (1.108)
[1.290, 1.745]
[2.209, 2.854]
-1.033
1.51×10−11
α
-0.032 (0.255)
-0.031 (0.161)
[-0.088, 0.016]
[-0.076, 0.010]
-0.003
0.985
Adults – Experience-weighted attraction
φ
0.521 (0.185)
0.587 (0.101)
[0.476, 0.571]
[0.546. 0.630]
-0.439
0.0009
ρ
0.379 (0.268)
0.308 (0.200)
[0.292, 0.465]
[0.208, 0.407]
0.298
0.026
β
1.231 (0.742)
1.290 (0.763)
[1.092, 1.378]
[1.131, 1.457]
-0.078
0.566
α
-0.052 (0.308)
0.040 (0.344)
[-0.120, 0.015]
[-0.030, 0.102]
-0.281
0.040
SD = standard deviation; MCMC = Markov Chain Monte Carlo sampling; d = Cohen’s d effect size
SD = standard deviation; MCMC = Markov Chain Monte Carlo sampling; d = Cohen’s d effect size
Children—CU model
ASD children showed a significantly higher learning rate than TDchildren (t[140.46] = 3.68, p < .001, d = 0.62; 95% confidence interval [CI] 0.26 to −0.93; Fig 3C). Simulations showed the optimal learning rate (i.e., leading to higher choice accuracy) for the CU model is 0.18 (Fig 3D, see also S1 Text), which is closer to the learning rate for TDchildren (MTD = 0.19) than the learning rate for ASD children (MASD = 0.26). A higher learning rate in our learning schedule reflects oversensitivity to feedback (including probabilistic punishment, which should be ignored). There were no differences on the other model parameters (β, α; p > .1). Results were unchanged with IQ as a confound regressor.
Adolescents—R-P model
A repeated-measures feedback type × diagnosis linear mixed-effect model with learning rates as dependent variables showed a significant main effect of feedback type (F[1,202] = 33.04, p = 3.20 × 10−8) and a significant interaction between feedback type and diagnosis (F[1,202] = 12.57, p = .0004), but no main effect of diagnosis (p = .1; Fig 3C). Reward learning rates were significantly larger than punishment learning rates (p < .0001, d = 0.43). Pairwise post hoc comparisons showed autistic adolescents’ reward learning rate was significantly lower than TD adolescents’ reward learning rate (p = .004, d = −0.39), but their punishment learning rates were not significantly different (p = .7). Additionally, TD adolescents’ reward learning rate was significantly higher than both their punishment learning rate (p < .001, d = 0.74) and ASD adolescents’ punishment learning rate (p < .001, d = 0.62).In the context of the R-P model (with two learning rates), simulations showed the optimal reward and punishment learning rates for choice accuracy are 0.96 and 0.60, respectively (Fig 3D and S6 Fig). This optimal pattern of a reward learning rate higher than the related punishment learning rate is also shown in TD adolescents’ learning rates, whereas autistic adolescents showed on average similar levels of reward and punishment learning and reduced learning from rewards compared to TD adolescents. In addition to reduced learning from rewards, autistic adolescents also showed significantly lower value sensitivity (β; t[169.27] = −7.24, p = 1.51 × 10−11, d = −1.05, 95% CI −1.32 to −0.73), reflecting more stochastic choice behavior. These results suggest that reduced reward learning and lower value sensitivity drive worse task performance in ASD adolescents. Results were unchanged with IQ as a confound regressor.
Adults—EWA-DL
Autistic adults showed on average a significantly lower inverse learning rate (φ; t[201.2] = −3.37, p = .0009, d = −0.46, 95% CI −0.71 to −0.17)—which is effectively comparable to a higher Rescorla-Wagner learning rate. Simulations show that in this model, the optimal value for φ is 0.85 (MASD = 0.52, MTD = 0.59; Fig 3D and S5 Fig). ASD adults also showed significantly higher experience-weight values (ρ) than TD adults (t[220.82] = 2.25, p = .021, d = 0.30; 95% CI 0.04 to −0.56), indicating a faster reliance on past (acquisition) experience, leading to inflexibility. When IQ was entered as a confound regressor, the difference in φ remained significant (p = .004), but the difference in experience decay (ρ) did not (p = .2).For associations between task behavior and model parameters, see S4 Text and S8 Table.
Symptomatology correlations in ASD
All correlations with symptomatology are listed in S9 Table and S10 Table. Here, we discuss only those that remained significant after Bonferroni correction for multiple comparisons.In the ASD children, perseverative errors were positively correlated with anxiety (Fig 4A; r72 = 0.34, p = .0040). However, no associations with model parameters survived multiple comparison corrections. For the adolescent group, neither associations with task behavioral measures nor model parameters survived Bonferroni correction. In the adult group, both perseverative errors and experience decay (ρ) were positively correlated with ADI-RRRB (perseverative errors–Fig 4B, r116
= 0.29, p = .0013; experience decay, ρ–Fig 4F, r116 = 0.28, p = .0022). Additionally, perseverative errors were positively associated with parent-reported ADHD hyperactivity/impulsivity (Fig 4C; r94 = 0.32, p = .0017), though this association would not survive Bonferroni correction when controlling for the RRB association (r89 = 0.26, p = .013). Win-stay behavior was negatively correlated with both ADI-RRRB and RBS-R Ritualistic-Sameness behavior (Fig 4D and 4E; ADI-RRRB r116 = −0.31, p = .0007; RBS-R Ritualistic-Sameness r91 = −0.30, p = .0004), and relatedly so was value sensitivity (β; Fig 4G and 4H; ADI-RRRB r116 = −0.29, p = .0019; RBS-R Ritualistic-Sameness r91 = −0.32, p = .0017). Value sensitivity was also negatively associated with parent-reported ADHD symptomatology in ASD adults (Fig 4I and 4J; ADHD hyperactivity/impulsivity r116 = −0.37, p = .0003; ADHD inattention r116 = −0.30, p = .0037).
Fig 4
Symptomatology correlations in ASD.
(A) In ASD children, perseverative errors were significantly correlated with anxiety (r72 = 0.34, p = .0040). In ASD adults, (B) perseverative errors were significantly correlated with ADI-R RRB (r116
= 0.29, p = .0013). (C) Perseverative errors were further significantly positively related to parent-reported ADHD Hyperactivity/Impulsivity (r94 = 0.32, p = .0017). Win-staying was significantly negatively related to (D) ADI-R RRB (r116 = −0.31, p = .0007) and (E) RBS-R Ritualistic-Sameness (r91 = −0.30, p = .0004). In ASD adults, experience decay (ρ) was significantly positively associated with (E) RRB (ADI-R RRB r116 = 0.28, p = .0022) as was (F, G) value sensitivity (β; ADI-R RRB r116 = −0.29, p = .0019; RBS-R r91 = −0.30, p = .0040). (H, I) Value sensitivity (β) was also significantly negatively correlated with parent-reported ADHD symptomatology (ADHD hyperactivity/impulsivity r116 = −0.37, p = .0003; ADHD inattention r116 = −0.30, p = .0037). ADHD, attention-deficit hyperactivity disorder; ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder; RBS-R, Repetitive Behavior Scale-Revised; RRB, restricted, repetitive behavior (see S1 Data).
Symptomatology correlations in ASD.
(A) In ASD children, perseverative errors were significantly correlated with anxiety (r72 = 0.34, p = .0040). In ASD adults, (B) perseverative errors were significantly correlated with ADI-RRRB (r116
= 0.29, p = .0013). (C) Perseverative errors were further significantly positively related to parent-reported ADHD Hyperactivity/Impulsivity (r94 = 0.32, p = .0017). Win-staying was significantly negatively related to (D) ADI-RRRB (r116 = −0.31, p = .0007) and (E) RBS-R Ritualistic-Sameness (r91 = −0.30, p = .0004). In ASD adults, experience decay (ρ) was significantly positively associated with (E) RRB (ADI-RRRB r116 = 0.28, p = .0022) as was (F, G) value sensitivity (β; ADI-RRRB r116 = −0.29, p = .0019; RBS-R r91 = −0.30, p = .0040). (H, I) Value sensitivity (β) was also significantly negatively correlated with parent-reported ADHD symptomatology (ADHD hyperactivity/impulsivity r116 = −0.37, p = .0003; ADHD inattention r116 = −0.30, p = .0037). ADHD, attention-deficit hyperactivity disorder; ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder; RBS-R, Repetitive Behavior Scale-Revised; RRB, restricted, repetitive behavior (see S1 Data).No correlations with learning rates (η, ηrew, ηpun, φ) nor lose-shift behavior survived Bonferroni correction in any age group. Of note, no significant associations between either task behavior or model parameters and social-communication difficulties were observed.
Discussion
In this study, we examined flexible behavior on a PRL task and used reinforcement learning models to investigate underlying learning mechanisms in autistic and neurotypical children, adolescents, and adults. Overall, we found evidence of on average reduced flexible behavior in autistic individuals, as indexed by poorer task performance across measures. Our results also show a developmental effect whereby older age groups outperformed younger age groups on the task. Using computational modeling of behavior, we showed that dominant learning mechanisms shift with developmental stage, but not diagnosis, and that poorer task performance in ASD is underpinned by atypical use of the age-related dominant learning mechanism in each age group. Furthermore, we found evidence for an association between perseveration and behavioral rigidity in ASD, but only in adults.These findings emphasize the importance of a developmental framework when examining mechanistic accounts of both intact and reduced flexible behavior. Although the role of development is well documented in the neurotypical literature, particularly with respect to key brain regions for cognitive flexibility, goal-directed decision-making, and feedback learning [9, 26, 70], age-related differences in ASD have been relatively understudied. Examining learning mechanisms across development, we found dominant differential integration of reward and punishment feedback in both adolescent groups, corresponding with literature that suggests neurotypical adolescents are hyperresponsive to rewards [29, 71]. In contrast, children’s behavior was best captured by a single learning rate, and adults showed evidence of increasingly weighting their accumulating experience to inform subsequent decisions and slow down new learning. This dominant experience-weight mechanism in adults is consistent with previous neurotypical research [47]; however, our study is the first to report the same dominant mechanism in ASD adults. These results therefore posit that cognitive and reinforcement-based processes are governed primarily by age, leading to the relative dominance of different learning mechanisms in different age groups. In this way, differential feedback learning may be developing in children and strengthened in adolescence, and experience weighting may similarly develop and then prevail in adulthood.Previous research suggests that reversal learning—and, more broadly, cognitive flexibility—is impaired in ASD (e.g., [1, 72]) and may be underpinned by the recruitment of different brain regions to TD [22]. Our findings provide support for the impairment hypothesis in that on average the ASD group was less accurate and more perseverative and showed reduced outcome sensitivity compared to the TD group. Furthermore, this pattern of results was consistent in both subsample analyses, showing robustness of findings in both an IQ-matched subsample and a subsample including only those ASD individuals who reach ADI-R criteria [46]. Notably, autistic adolescents showed reduced win-staying compared to TD adolescents, in line with previous studies that showed reduced win-staying in adults [21, 22]. However, in this study, we did not find reduced win-staying specifically in autistic adults compared to TD adults.Our computational modeling findings suggest that reduced flexible behavior in the ASD group is underpinned by significant differences in the efficient use of learning mechanisms within each age group on this task. Both the children and adult ASD groups showed faster learning rates compared to their TD counterparts. Here, faster learning rates are less optimal, as they result in reduced ability to ignore probabilistic feedback. These results are consistent with predictive coding and Bayesian accounts of ASD that suggest “overlearning” in response to feedback and difficulties ignoring noise, putatively due to precise or inflexible prediction errors [37, 38]. Indeed, studies using volatile task environments or near-chance reward contingencies have reported intact learning and updating or superior performance in ASD [22, 39]. In these contexts, fast learning rates are optimal, as changes are more frequent and therefore updating must be too.Thus, findings demonstrate that altered learning rates in ASD have different effects on behavior depending on the learning environment and, in tandem, that computational models characterize differences rather than solely deficits, shedding light on environments in which differences may be expressed as strengths rather than difficulties. The computational differences in ASD appear to manifest as pronounced difficulties when the environment is less volatile, and learning when to ignore probabilistic feedback is as important as tracking change. These difficulties may underpin the marked difficulties with minor (probabilistic) deviations in routines or unexpected changes in ASD that caregivers so frequently report [73]. In different environments, faster learning may manifest in strengths; these differences have important implications for intervention development.In ASD adolescents, reduced flexible behavior—and, particularly, reduced win-staying—was underpinned by reduced reward learning compared to TD adolescents. This finding is consistent with previous research showing impaired reward circuitry dysfunction in autistic adolescents [74]. Whereas neurotypical adolescents are thought to demonstrate increased risk due to high reward sensitivity, reduced reward learning in autistic adolescents may result in reduced risk-taking and serve as a protective effect [75]. Reduced reward learning could also have implications for behavioral interventions. If autistic adolescents do not learn from typical rewards in the same way that TD adolescents do, the type(s) of rewards used in behavioral interventions would require adapting [76]. For example, there is evidence to suggest autistic individuals assign specific reward value to their circumscribed interests such that they may be of value in intervention design [77-79].Reduced flexible behavior has previously been associated with RRB in ASD [1, 80–82], though results are not consistent despite a strong theoretical link. Here, we observed robust, moderately strong associations between perseveration and RRB in autistic adults. We also found no evidence of associations with social-communication difficulties, providing support for the specificity to RRB. On the RBS-R, these associations were specific to the Ritualistic-Sameness and Stereotyped Behavior subscales, capturing behavioral rigidities. Previous literature has also reported associations between flexibility impairments and RRB symptom severity in ASD adults [83] with mixed findings in children and adolescents [82, 84–86]. Moving forward, examining this association across developmental stages will continue to be important.To our knowledge, this study is the first to elucidate a potential learning mechanism by which behavioral rigidity manifests in autistic adults: perseveration as a result of a reluctance or inability to switch—“getting stuck”—because new information is devalued in favor of past experience, which in turn impedes updating choice behavior. Furthermore, as this mechanism has been associated with dopamine transporter differences in neurotypical adults [47], and abnormalities in the dopaminergic system have been implicated in ASD [87], this study highlights a potential mechanistic link between neurobiology and behavior worthy of further study.Beyond perseveration, RRB in autistic adults positively associated with reduced value sensitivity (i.e., more stochastic choice behavior). This mechanism was also associated with more ADHD symptoms in autistic adults. Reduced value sensitivity has previously been identified as a key factor in poor task performance in anhedonia [88]. Together, these findings suggest that value sensitivity may have transdiagnostic value in explaining aspects of reduced flexible behavior. As altered decision-making is prevalent across many neurodevelopmental and neuropsychiatric disorders, examining underlying processes in relation to symptom dimensions rather than purely diagnostic categories will likely be of greater value for understanding implicated brain circuitries [89].In autistic adolescents, we found no relationship between performance measures or learning mechanisms and clinical symptoms. In children with ASD, we observed a positive association between perseverative behavior and anxiety symptoms. Previous studies have demonstrated a relationship between anxiety and reduced flexible behavior in non-autistic adults [90, 91] and children and adolescents with anxiety disorders [92]. One plausible link between perseveration and anxiety may be the intolerance of uncertainty (IU) construct, as uncertainty is inherent in probabilistic tasks. IU is a core construct in anxiety disorders [93] and a possible transdiagnostic mechanism [94] shown to be relevant for anxiety in ASD [95]. Associations between anxiety and RRB in ASD have frequently been reported [96, 97]. Together, our findings broadly support the notion that reduced flexible behavior is of clinical relevance in ASD; however, the extent to which particular processes may be differentially linked to specific aspects of RRB versus commonly co-occurring features of anxiety or ADHD at different developmental stages will require further examination.
Limitations
This study has a number of limitations. Firstly, despite the large sample size and wide age range, the sample does not include children younger than 6 or adults above 30 years of age. Future research including very young children and older adults could allow for the assessment of any other age-related changes in dominant learning mechanisms. Secondly, it is important to note that each group’s winning model is only relative to the other models tested here—although we note that the models capture behavior well and perform far above chance. However, it is (always) possible that other models may perform even better and further models may be developed in the future. A full model with all parameters combined was not possible because of convergence issues, emphasizing the relative dominance of learning mechanisms rather than any suggestions of mutual exclusivity. We highlight, nevertheless, that the study is the first to compare reinforcement learning models in ASD across age groups. Thirdly, our approach necessitated that we implicitly treated each diagnostic and age group as relatively homogeneous. The increasing recognition of the considerable phenotypic and etiological diversity of ASD indicates potential individual differences in learning processes within or across these a priori defined subgroups. Estimating the learning strategy for each individual would allow for a “bottom-up” approach to identifying potential subgroups based on learning strategies. Fourth, our sample was limited to individuals with an ASD diagnosis and TD counterparts. Given that reduced flexible behavior and atypical reinforcement learning are implicated in many other areas of psychiatry, it would be informative to extend this study with a transdiagnostic sample, in the context of the research domain criteria framework (RDoC; [89]). Additionally, given the growing literature suggesting differential reward processing in ASD, future work could assess potential differences in learning and flexible behavior in the context of different reward modalities, i.e., use different types of feedback, such as monetary stimuli. Finally, it will be crucial to verify our results through replication. The current sample has been reassessed as part of a longitudinal project, thereby providing some opportunity for this.
Conclusions
Current results suggest group-level impairments in flexible behavior across developmental stages in ASD. We show evidence of developmental shifts in dominant computational mechanisms underlying PRL that are consistent across ASD and TD individuals. Within each age group, differences in model parameter estimates showed less optimal learning in ASD, underpinning poorer task performance. Additionally, we show that perseverative behavior—and, in adults, learning mechanisms—were related to behavioral rigidities or co-occurring symptoms of anxiety or ADHD. Findings emphasize the importance of understanding reduced flexible behavior in ASD within a developmental framework and underline the strength of computational approaches in ASD research.
Excel spreadsheet containing, in separate sheets, the underlying numerical data for figures and figure panels: 1C, 2A-2E, 3C, 3D, 4A-4J, S1, S2A-S2L, S3A-S3B, S4, and S7.
(XLSX)Click here for additional data file.
Supplementary methods.
(DOCX)Click here for additional data file.
Additional IQ and subsample analyses.
(DOCX)Click here for additional data file.
Evidence of learning.
(DOCX)Click here for additional data file.
Further results for comparisons of model parameter estimates.
(DOCX)Click here for additional data file.
z-RTs in the PRL task averaged across task trials; shaded area represents the standard deviation.
Notably, reaction times do not change at the point following reversal, illustrating that reaction times are unlikely to reflect task-relevant processes. PRL, probabilistic reversal learning; z-RT, reaction time (z-scored).(TIF)Click here for additional data file.
Box plots showing task behavior for (A-D) the full sample, (E-H) the IQ-matched subsample, and (I-L) Risi and colleagues’ ADI-R criteria ASD subsample.
The pattern of results remains largely unchanged across both subsample analyses. ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder.(TIF)Click here for additional data file.(A) Trial-by-trial average proportion of correct responses (here, yellow in acquisition phase, blue in reversal phase) plotted separately for the groups that passed and failed the learning criterion. The red lines indicate the mean for that task phase (acquisiton/reversal) and the orange lines indicate the 95% confidence intervals. Thus, both groups performed above chance in both task phases. (B) Diagnostic and age group average proportion of correct responses for each task phase, plotted separately for the pass/fail groups to confirm that perfgormance above chance was maintained even within diagnostic and age subgroups.(TIF)Click here for additional data file.
Independent contribution of model parameters.
Pair plots of each group’s winning model parameters for ASD (top panel) and TD (bottom panel). In each pair plot, diagonal plots show marginal distributions of each parameter; off-diagonal plots show pairwise scatters of parameters. ASD, autism spectrum disorder; CU, counterfactual update model; EWA, experience-weighted attraction–dynamic learning rate model; RP, reward-punishment model; TD, typical development.(TIF)Click here for additional data file.
Model recovery.
Data from 40 synthetic participants were simulated with each of our three main models. Color indicates model weights calculated with Bayesian model averaging using Bayesian bootstrap (higher model weight value indicates higher probability of the candidate model to have generated the observed data). CU, counterfactual update model; EWA, experience-weighted attraction–dynamic learning rate model; RP, reward-punishment model.(TIF)Click here for additional data file.
Simulation showing a larger value difference for a higher reward learning rate (TD) than a lower reward learning rate (ASD), when punishment learning rates are comparable.
ASD, autism spectrum disorder; TD, typical development.(TIF)Click here for additional data file.
Highly correlated factual and counterfactual learning rates.
(TIF)Click here for additional data file.
Participant numbers and ADI-R scores (mean, SD) for the full ASD sample and Risi and colleagues’ (2006) ADI-R criteria subsample.
ADI-R, Autism Diagnostic Interview-Revised; ASD, autism spectrum disorder; SD, standard deviation.(DOCX)Click here for additional data file.
Effective number of parameters for the RW and CU models.
CU, counterfactual update; RW, Rescorla-Wagner.(DOCX)Click here for additional data file.
Behavior and model parameter estimates correlations with all RBS-R subscales.
RBS-R, Repetitive Behavior Scale-Revised.(DOCX)Click here for additional data file.
Descriptive statistics (mean, SD—unless otherwise stated) for the full sample and the IQ-m, within age and diagnostic groups, with p-values for within-age group, between diagnostic group comparisons of age, sex, and IQ.
IQ-m, IQ-matched subsample; SD, standard deviation.(DOCX)Click here for additional data file.
Numbers, proportions, and chi-squared statistics for learning criterion attainment status (pass/fail) by diagnostic and age groups.
(DOCX)Click here for additional data file.
Model weights for model runs with IQ as a covariate.
(DOCX)Click here for additional data file.
Model weights for model runs with age groups collapsed.
(DOCX)Click here for additional data file.
Correlations between task behavior and model parameters.
(DOCX)Click here for additional data file.
Correlations between task behavior, age, IQ, and symptomatology.
(DOCX)Click here for additional data file.
Correlations between model parameters, age, IQ, and symptomatology.
(DOCX)Click here for additional data file.12 Dec 2019Dear Dr Crawley,Thank you for submitting your manuscript entitled "Modeling flexible behavior in children, adolescents and adults with autism spectrum disorder and typical development" for consideration as a Research Article by PLOS Biology.Your manuscript has now been evaluated by the PLOS Biology editorial staff, as well as by an Academic Editor with relevant expertise, and I am writing to let you know that we would like to send your submission out for external peer review.However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.Please re-submit your manuscript within two working days, i.e. by Dec 16 2019 11:59PM.Login to Editorial Manager here: https://www.editorialmanager.com/pbiologyDuring resubmission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF when you re-submit.Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. Once your manuscript has passed all checks it will be sent out for review.Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.Kind regards,Gabriel Gasque, Ph.D.,Senior EditorPLOS Biology24 Jan 2020Dear Dr Crawley,Thank you very much for submitting your manuscript "Modeling flexible behavior in children, adolescents and adults with autism spectrum disorder and typical development" for consideration as a Research Article at PLOS Biology. Your manuscript has been evaluated by the PLOS Biology editors, by an Academic Editor with relevant expertise, and by three independent reviewers. You will note that reviewer 2, Stefano Palminteri, has signed his comments.In light of the reviews (below), we will not be able to accept the current version of the manuscript, but we would welcome re-submission of a much-revised version that takes into account the reviewers' comments. We cannot make any decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is also likely to be sent for further evaluation by the reviewers.We expect to receive your revised manuscript within 2 months.Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension. At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may end consideration of the manuscript at PLOS Biology.**IMPORTANT - SUBMITTING YOUR REVISION**Your revisions should address the specific points made by each reviewer. Please pay special attention to the comments reviewers 1 and 3 raise regarding IQ matching. We think it is fundamental to address this point fully for a successful revision. We also think reviewer 2 raises sophisticated points about the models, which should be thoroughly addressed.Please submit the following files along with your revised manuscript:1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.*NOTE: In your point by point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.*Re-submission Checklist*When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_ChecklistTo submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.Please make sure to read the following important policies and guidelines while preparing your revision:*Published Peer Review*Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/*PLOS Data Policy*Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5*Blot and Gel Data Policy*We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements*Protocols deposition*To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methodsThank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.Sincerely,Gabriel Gasque, Ph.D.,Senior EditorPLOS Biology*****************************************************REVIEWS:Reviewer #1: This is a revised manuscript, I did not see the original. The authors describe results from a study of flexible behavior in a large sample of children, adolescents, and adults with autism spectrum disorder (ASD). The study uses computational modeling to quantify latent variables indexing perseveration and feedback sensitivity, and assess these variables in relation to diagnosis, developmental stage, and symptoms of ASD. In the participants completing the probabilistic reversal learning task, the authors observe that those with ASD showed more perseveration and less feedback sensitivity than typically developing (TD) peers, and older individuals in both groups showed greater feedback sensitivity than younger individuals. The computational modeling results suggest that learning mechanisms underlying flexible behavior differ across developmental stages, and in those with ASD reduced flexible behavior was driven by less optimal learning. Different aspects of task performance were related to symptoms in children vs. adults with ASD. This is an interesting, well-powered study on a timely topic. Some suggestions are below.It is interesting that the authors choose to use a probabilistic reversal learning (PRL) task to study flexibility in autism, rather than a traditional cognitive flexibility task. This is because the PRL task includes a reward outcome. Thus, both flexibility and reward processing are indexed by such a task, and reward processing may also be altered in ASD. Can the authors elaborate in the Introduction on their decision to use such a task, rather than some variant of the Wisconsin Card Sort Task or other task-switching or set-shifting paradigms that do not involve a reward component?Is there previous literature to indicate that the specific reward/punishment stimuli used in this task are appropriately sensitive for indexing reward/punishment in both ASD and TD groups across age?Why was reaction time not assessed in addition to accuracy? It would be helpful to include this information as well as statistics comparing groups.It is very surprising that the authors do not attempt to match the ASD and TD groups on full-scale IQ. It would be best to take a subset of the larger sample that is matched on full-scale IQ to see if the results obtained from the larger sample still hold. It is not sufficient to simply include IQ as a confounding regressor.The authors may wish to see the following relevant work:The paradox of cognitive flexibility in autism. Geurts HM, Corbett B, Solomon M. Trends Cogn Sci. 2009 Feb;13(2):74-82.Demystifying cognitive flexibility: Implications for clinical and developmental neuroscience. Dajani DR, Uddin LQ. Trends Neurosci. 2015 Sep;38(9):571-8.Reviewer #2, Stefano Palminteri: Daisy and colleagues presents a study of reinforcement learning across development, contrasting neurotypical and ASD subjects. The authors report that both factors (age and psychiatric scores) affect behavioral performance in this probabilistic reversal learning task (in particular metrics of efficiency and feedback sensitivity). They then apply computational modeling to this task and report that, across different age groups, choices are not explained by the same computational model and that some parameters differ across different age groups. There is much to like about this paper, such as the big sample size (not usual for this kind of research), coupling development and ASD research with computational modeling. The paper has the potential to become a mile-stone in the field. I do have however some important concerns about the model specifications, selection and statistical analyses.FP model specificationOne of the models is labelled 'fictitious play' (FP) and consists in a Rescorla-Wagner model where the value of the unchosen option is also updated (with the opposite prediction error). The first (but not the main) issue with this model is that the name if wrong and misleading. In behavioural game theory 'fictitious play' refers to a way of finding a best response to an opponent play by iteratively mentally simulating the other player response (see chapter 'Learning' of Camerer's book): a situation that clearly does not apply here, as FP learning concerns beliefs, not values. The more important issue concerning this model is that, in its current specification, the factual and counterfactual updates are governed by the same learning rate. This is problematic as the functional complexity (it has an additional equation compared to the Rescorla-Wagner) of the model is not quantified by an additional parameter. This is also problematic as it does not allow to quantify the counterfactual update separately from the factual one. It would be psychologically plausible to suppose that the two are differentially affected by ASD and age.EWA model specificationAnother model is the one labelled 'Experience-Weighted Attraction' model. I have also an issue concerning this labeling. Again, the EWA model has been developed for game theory not bandits; as this paper is addressed to psychologists and neuroscientists, rather than game theorists, I find the labeling unfortunate. Furthermore the key distinctive feature of the EWA model is counterfactual update (i.e., update of the value of the unchosen strategy, based on the opponent's choice), which is not relevant here. If anything the current model il closer to Erev and Roth's model (American Economic Review, 1995) than the EWA. Yet, labeling is not the main issue I have with this model. In this paper the EWA model embodies an interesting (and plausible) idea, that is, when "experience" increases (i.e., number of trials per option) new outcomes count less. The problem is that it does so in a formalism that is quite different compared to that used in the other two models. For instance, while for some sets of parameter the EWA model can approximate the RW model, for many other parameters values option values will converge to very different quantities. Furthermore, the fact that experience weights can increase unboundedly (when the number of trials increases), is at odds with neurobiologically plausible instantiations of RL (see any basal ganglia model, for instance). To obviate this issues and ensure commensurability with the other models, the authors should replace the EWA model with Miller's model (Psychol Rev. 2019), which suppose a parallel 'habit' learning (equivalent to the experience weight) with a formalise that is closer to that of the other models and parameters that are psychologically easy to interpret.Softmax specificationThe authors included a bias term in the softmax. I have issue with this parameter as I suspect it is capturing part of the effects induced by less ad hoc processes such as learning asymmetric, counterfactual update and increased habits. I suspect it's value is consistent with a bias toward the first rewarded stimulus. Am I correct? In any case, I would not include it, unless it is strongly justified by model comparison and model falsification.Model space specificationI think the model space should include the RW model + the full model that include all the features (counterfactual update learning rate; two learning rates for positive and negative prediction errors and the habit learning system of Miller) as two "extreme" benchmarks (hypo n vs. hyper-parameterisation).Model comparison and selectionI really liked the Bayesian model selection approach and the fact that authors showed also the simulation. However it would be interesting to see how the models perform on the other behavioral metrics (perseveration errors, win/stay etc.) to have a better idea of what behavioural feature is falsifying the "losing" models. It would also be important to show the model recovery on simulated datasets (i.e., the capacity to retrieve the correct model by model comparison in data where the ground truth is known). Finally, it would be interesting to know which is the "overall" winning model (across all ages).Statistics on the parametersI guess the parameters are correlated between them (to some extent, it is always the case). It would be interesting to verify the parameter recovery to check their correct estimability (similar to model recovery). Maybe a structural equation modeling (SEM)- approach would be useful to assess the effect of clinical score while controlling for the correlation between parameters and between scores. I also suspect that in the population, there are subjects whose behavior is random (i.e., the likelihood of any model model is not different from assuming random choices). It would be interesting to check the correlations (or SEM results on the non-random subjects (as parameters values of random subjects do not make sense). Finally, once the "full" model implemented, it would be interesting to perform the correlations (or SEM) on the parameters of the full model.Discussion beyond ASDI know that it is not central in this study, but, as I believe that authors should mention that fact that that while some studies (we found it: Lefebvre et al 2019; others too: Van Slooten et al, 2018, 2019) showed higher learning rates for positive compared to negative learning rates, some others (Niv et al 2011; Gershman, 2015) found the opposite. As the present study contributes to this debate, it would be worth mentioning the "controversy" in the discussion.Line 157Please avoid the term "poor" in describing adolescent decision-making.Reviewer #3:Thank you for the opportunity of reviewing the manuscript « Modeling flexible behavior in children, adolescents and adults with autism spectrum disorder and typical development » and apologies for the delay with which I am sending my review.Restricted and repetitive behaviours are key symptoms in ASD. These symptoms are plausibly linked to diminished flexibility, which can be measured and modeled using standard reverse learning tasks. In these tasks, participants must learn contingency rules and adjust to shifts in these rules in response to negative feedback. Some amount of perseveration is warranted (it would be maladaptive to change our minds every time we get a single negative feedback - 100% contingencies are rare in life) but too much perseveration is often the signature of poor cognitive flexibility. In this paper, the authors test large sample sizes (572 children, adolescents and 102 adults with ASD; N=321 and typical development; TD; N=251) using such a reversal learning task. Their results show that ASD participants perseverate more on average than TDparticipants, which is consistent with prior findings in the literature. They also find that feedback sensitivity is modulated by age, which is also consistent with prior findings in the literature. Finally, they find that the learning rule used by individuals with ASD is less optimal than the rule used by individuals in the control group, and that these differences relate to measurable symptoms (anxiety and perseveration).Does the paper address a question of sufficiently broad interest ?The paper addresses a question that is of broad interest: Restricted and repetitive behaviours are key symptoms in ASD, which is a common disorder affecting millions of people. But beyond ASD, understanding the development of learning is an important and broad question.Does the paper make a sufficient leap in the literature ?Much of what the paper shows has been argued for in the literature. But the paper makes a leap by testing a large sample, using empirical methods that are robust and solid computational analyses. Autism research has been fraught with papers using small sample sizes and poorly characterized samples, which makes a lot of the literature essentially uninterpretable. This paper, by contrast, is a strong, and much needed contribution. As the authors state in the introduction, the current literature is inconsistent and one of the main reason for this is poor methods, including small samples sizes.Is the analysis correct ?The task is sound and the analysis appears correct to me but I have strong reservations about sample characteristics, especially in the absence of a pre-registration plan. It is obviously too late to pre-register but given the value of this hard-to-acquire data, I think that this is a major mistake. It is great that a culture shift is happening towards increased sample sizes, but this should be combined with a culture shift towards better research practices overall, including pre-registration. I do not think that this is grounds for rejection (it would be punitive!) but I do think that it is a great shame that we have no way of knowing how much of the analysis was planned ahead of time, what was decided a priori, and what should be considered exploratory. This is all the more concerning that there are many degrees of freedom when dealing with such data (much clinical data was collected, many models were tested, many dependent variables can be compared).Here are some examples:1) I was surprised not to see the ADOS (and only the ADI-R) - was the ADOS collected ? I did not see information allowing me to know whether all participants in the sample were above overall ADIR cut-off for an ASD, or above SRS cut-off. If not, have the autors tried to restrict their analysis to the subsample that meets all diagnostic criteria?2) How much were the correlations between performance and symptom severity predicted ? Is it a problem that associations are found for some clinical outcomes but not others (even though some of them arguably tap the same construct).3) Based on the methods section, it isn't clear to me how the ASD and TD group were matched. Do they *happen* to not differ on age and sex or were they *chosen* to match on these variables ? Given that the groups differ on IQ, I was wondering why the authors had not decided to restrict their analysis to a well-matched sample.I do not think that the additional analyses provided in the SM address this concern appropriately. Covariance analyses assume statistical properties that are hard to meet and the analyses are hard to interpret when the groups differ substantially on the covariate, which they do. These analyses also require that the relationship between covariate and outcome be the same in both groups (ie regression slopes), which they are not. (unless I missed something)One option would be to test a larger control group (because it is comparatively harder to gather data from patients). This would allow the authors to regress performance in the RL task against IQ in the TD group and then check for each individual in the ASD group whether there is a discrepancy between observed and expeected performance given observed IQ.If that is not feasible in the current study, I would recommend random matching on age, sex and IQ (using an automated algorithm such as the matching package in R) and the subsequent analyses to be presented in the main text.There are indeed issues with matching but given the particular topic at stake here and given the statistical assumptions behind covariate analyses, this strategy appears sounder statistically. Specifically, matching may be better recommended when the variable used to match groups really controls for a constraint on experimental task performance that isn't central to the hypothesis. My understanding is that we are in a such a situation here IQ is a predictor of learning but the hypothesis is about perseveration, not intelligence.At the very least, given that the analyses were not pre-registered, the reader should have the option of seeing what happens when multivariate matching is done (on age sex and IQ). If the results are strikingly different, this should be discussed transparently.5 Jun 2020Submitted filename: PRL_Response to reviewers comments_DC_LZ.pdfClick here for additional data file.21 Aug 2020Dear Dr Crawley,Thank you for submitting your revised Research Article entitled "Modeling flexible behavior in children, adolescents and adults with autism spectrum disorder and typical development" for publication in PLOS Biology. I have now obtained advice from the original reviewers 1 and 2 and have discussed their comments with the Academic Editor, who also assessed the way you answered to the concerns originally raised by reviewer 3.Based on the reviews, we will probably accept this manuscript for publication, assuming that you will modify the manuscript to address the remaining points raised by reviewer 2. Please also make sure to address the data and other policy-related requests noted at the end of this email.We expect to receive your revised manuscript within two weeks.***IMPORTANT:Your revisions should address the specific points made by reviewer 2. In addition, we would like you to consider changing your title, which we think is too descriptive. We recommend one that conveys the central biological message and suggest the following. However, we are open to discuss alternatives:“Modeling flexible behavior deficits in Autism Spectrum Disorder shows age-dependency and less optimal learning within each age group.”Please submit the following files along with your revised manuscript:1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.*NOTE: In your point by point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Related" file type.In addition to the remaining revisions and before we will be able to formally accept your manuscript and consider it "in press", we also need to ensure that your article conforms to our guidelines. A member of our team will be in touch shortly with a set of requests. As we can't proceed until these requirements are met, your swift response will help prevent delays to publication.*Copyediting*Upon acceptance of your article, your final files will be copyedited and typeset into the final PDF. While you will have an opportunity to review these files as proofs, PLOS will only permit corrections to spelling or significant scientific errors. Therefore, please take this final revision time to assess and make any remaining major changes to your manuscript.NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:https://journals.plos.org/plosbiology/s/supporting-information*Published Peer Review History*Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/*Early Version*Please note that an uncorrected proof of your manuscript will be published online ahead of the final version, unless you opted out when submitting your manuscript. If, for any reason, you do not want an earlier version of your manuscript published online, uncheck the box. Should you, your institution's press office or the journal office choose to press release your paper, you will automatically be opted out of early publication. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.*Protocols deposition*To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosbiology/s/submission-guidelines#loc-materials-and-methods*Submitting Your Revision*To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include a cover letter, a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable), and a track-changes file indicating any changes that you have made to the manuscript.Please do not hesitate to contact me should you have any questions.Sincerely,Gabriel Gasque, Ph.D.,Senior Editor,ggasque@plos.org,PLOS Biology------------------------------------------------------------------------ETHICS STATEMENT:-- Please include the full name of the IACUC/ethics committee that reviewed and approved the animal care and use protocol/permit/project license. Please also include an approval number.-- Please include the specific national or international regulations/guidelines to which your animal care and use protocol adhered. Please note that institutional or accreditation organization guidelines (such as AAALAC) do not meet this requirement.-- Please include information about the form of consent (written/oral) given for research involving humanparticipants. All research involving humanparticipants must have been approved by the authors' Institutional Review Board (IRB) or an equivalent committee, and all clinical investigation must have been conducted according to the principles expressed in the Declaration of Helsinki.------------------------------------------------------------------------DATA POLICY:We note that you have stated in the online submission system that readers can access the raw data by contacting the EU-AIMS LEAP group. However, we request that you provide the underlying numerical values that underlie the summary data displayed in the following figure panels: Figures 1C, 2ABCDE, 3CD, 4A-J, S1, S2A-L, S3AB, S4, and S7.The numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5These data can be made available in one of the following forms:1) Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information' and are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).2) Deposition in a publicly available repository. Please also provide the accession code or a reviewer link so that we may view your data before publication.Please also ensure that each figure legend in your manuscript include information on where the underlying data can be found, and ensure your supplemental data file/s has a legend.Please ensure that your Data Statement in the submission system accurately describes where your data can be found.------------------------------------------------------------------------Reviewer remarks:Reviewer #1: The authors have done a very nice job conducting new analyses to address reviewer comments.Reviewer #2: Overall, I think that the authors did a good job in addressing my concerns. The additional clarifications and analyses are very welcome. As I mentioned in my previous review, I think this paper would become an important milestone in the computational psychiatry of ASD. I do, however, have few remaining questions.1/ In response to my point R2.1 (counterfactual learning rate issue) the authors mention that (quote):"we would like to point out that our method of model comparison does indeed account for this increased functional complexity in the counterfactual update model, unlike simpler methods based on e.g. AIC/BIC comparison. This is one of the motivations from our perspective to conduct all computational modeling analyses in the Bayesian framework, where inferences are drawn from joint posterior distributions, rather than point estimates. Under the Bayesian framework, the joint parameter spaces of the CU and the simple RL model differ; thus, the effective number of parameters of these two models is different. It is the latter (i.e., effective number of parameters) that we use in the penalizing term in computing the model evidence"However, it is still unclear to me how it is possible that functional complexity (i.e., adding one additional equation - or complexifying it) is taken into account for penalization in their Bayesian approach. As this is a crucial issue (and their claim is counterintuitive), I believe that the authors should unpack and explain this point better in the revised manuscript.2/ In response to my point R2.2 the authors decided to maintenant the name EWA for their model, based on the fact that they used the same name in a previous paper (2013), while they acknowledge that their model lacks the distinctive features of the EWA model. I still think that the label is not appropriate and, in a sense, "historically unfair". I strongly encourage the authors to take a look at Table 6.2 and table 6.3 of the book "Behavioral Game Theory" (by Colin Camerer), they will see their model will be classified as "Reinforcement Learning" (Roth and Erev) by Camerer himself.Let's say a first group invent a model with the process X inside and call it 'John'. A second group invent a model with the processes X and Y inside and call it 'Paul'. A third group uses a model with the process X inside and call it 'Paul with no Y'. The first group would have all the rights to be annoyed, right? Of course this is not such a big deal, but I would love to know what are the authors' thoughts on this.3/ in the SI 'Additional methods' I believe the authors wanted to cite Lefebvre 2017, rather than 2018.18 Sep 2020Submitted filename: PRL_Response to reviewers comments_2b_DC_LZ.pdfClick here for additional data file.22 Sep 2020Dear Dr Crawley,On behalf of my colleagues and the Academic Editor, Franck Ramus, I am pleased to inform you that we will be delighted to publish your Research Article in PLOS Biology.The files will now enter our production system. You will receive a copyedited version of the manuscript, along with your figures for a final review. You will be given two business days to review and approve the copyedit. Then, within a week, you will receive a PDF proof of your typeset article. You will have two days to review the PDF and make any final corrections. If there is a chance that you'll be unavailable during the copy editing/proof review period, please provide us with contact details of one of the other authors whom you nominate to handle these stages on your behalf. This will ensure that any requested corrections reach the production department in time for publication.Early VersionThe version of your manuscript submitted at the copyedit stage will be posted online ahead of the final proof version, unless you have already opted out of the process. The date of the early version will be your article's publication date. The final article will be published to the same URL, and all versions of the paper will be accessible to readers.PRESSWe frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have not yet opted out of the early version process, we ask that you notify us immediately of any press plans so that we may do so on your behalf.We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.Thank you again for submitting your manuscript to PLOS Biology and for your support of Open Access publishing. Please do not hesitate to contact me if I can provide any assistance during the production process.Kind regards,Alice MussonPublishing Editor,PLOS Biologyon behalf ofGabriel Gasque,Senior EditorPLOS Biology
Authors: Thomas Insel; Bruce Cuthbert; Marjorie Garvey; Robert Heinssen; Daniel S Pine; Kevin Quinn; Charles Sanislow; Philip Wang Journal: Am J Psychiatry Date: 2010-07 Impact factor: 18.112
Authors: Eva Loth; Will Spooren; Lindsay M Ham; Maria B Isaac; Caroline Auriche-Benichou; Tobias Banaschewski; Simon Baron-Cohen; Karl Broich; Sven Bölte; Thomas Bourgeron; Tony Charman; David Collier; Fernando de Andres-Trelles; Sarah Durston; Christine Ecker; Andre Elferink; Marion Haberkamp; Robert Hemmings; Mark H Johnson; Emily J H Jones; Omar S Khwaja; Sabine Lenton; Luke Mason; Valentina Mantua; Andreas Meyer-Lindenberg; Michael V Lombardo; Laurence O'Dwyer; Koichi Okamoto; Gahan J Pandina; Luca Pani; Antonio M Persico; Emily Simonoff; Sitra Tauscher-Wisniewski; Jordi Llinares-Garcia; Spiros Vamvakas; Steve Williams; Jan K Buitelaar; Declan G M Murphy Journal: Nat Rev Drug Discov Date: 2016-01 Impact factor: 84.694
Authors: Karli K Watson; Stephanie Miller; Eleanor Hannah; Megan Kovac; Cara R Damiano; Antoinette Sabatino-DiCrisco; Lauren Turner-Brown; Noah J Sasson; Michael L Platt; Gabriel S Dichter Journal: Front Psychol Date: 2015-07-22
Authors: Rebecca B Hughes; Jayde Whittingham-Dowd; Steven J Clapcote; Susan J Broughton; Neil Dawson Journal: Autism Res Date: 2022-02-10 Impact factor: 4.633
Authors: Andrea M F Reiter; Michael Moutoussis; Lucy Vanes; Rogier Kievit; Edward T Bullmore; Ian M Goodyer; Peter Fonagy; Peter B Jones; Raymond J Dolan Journal: Nat Commun Date: 2021-06-22 Impact factor: 14.919