| Literature DB >> 32694654 |
Woo-Young Ahn1,2, Hairong Gu3, Yitong Shen4, Nathaniel Haines3, Hunter A Hahn3, Julie E Teater5, Jay I Myung3, Mark A Pitt3.
Abstract
Machine learning has the potential to facilitate the development of computational methods that improve the measurement of cognitive and mental functioning. In three populations (college students, patients with a substance use disorder, and Amazon Mechanical Turk workers), we evaluated one such method, Bayesian adaptive design optimization (ADO), in the area of delay discounting by comparing its test-retest reliability, precision, and efficiency with that of a conventional staircase method. In all three populations tested, the results showed that ADO led to 0.95 or higher test-retest reliability of the discounting rate within 10-20 trials (under 1-2 min of testing), captured approximately 10% more variance in test-retest reliability, was 3-5 times more precise, and was 3-8 times more efficient than the staircase method. The ADO methodology provides efficient and precise protocols for measuring individual differences in delay discounting.Entities:
Mesh:
Year: 2020 PMID: 32694654 PMCID: PMC7374100 DOI: 10.1038/s41598-020-68587-x
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Schematic illustration of adaptive design optimization (ADO) in the area of delay discounting. Unlike the traditional experimental method, ADO aims to find the optimal design that extracts the maximum information about a participant’s model parameters on each trial. In other words, ADO identifies the most informative or optimal design (d) using the participant’s previous choices (y), the mathematical model of choice behavior, and the participant’s model parameters (). In our delay discounting experiment with ADO, y would be 0 (choosing smaller and sooner reward) or 1 (larger and later reward), the mathematical model would be the hyperbolic function (see “Methods”), would be k (discounting rate) and (inverse temperature), and d* would be the experimental design (a later delay and a sooner reward, which are underlined in the figure) that maximizes the integral of the local utility function, , which is based on the mutual information between model parameters () and outcome random variable conditional upon design (y|d). For more mathematical details of the ADO method, see[24,25].
Figure 2Comparison of ADO and Staircase (SC) within-visit test–retest reliability of temporal discounting rates when assessed cumulatively in each trial (ADO) or every third trial (SC) (Experiment 1, college students) at each of the two visits. Two visits were separated by approximately one month. In each visit, a participant completed two ADO sessions and two SC sessions (within-visit test–retest reliability). Test–retest reliability was assessed cumulatively in each trial (See “Methods” for the procedure). Shaded regions represent the 95% frequentist confidence interval of the concordance correlation coefficient (CCC).
Figure 3Reliability and efficiency of the ADO method in Experiments 2 and 3 (A) Comparison of ADO and Staircase (SC) test–retest reliability of temporal discounting rates when assessed cumulatively in each trial (ADO) or every third trial (SC) (Experiment 2, patients with SUDs) (B) Test efficiency as measured by the cumulative test–retest reliability across trials (Experiment 3, Amazon MTurk workers). Dashed line = 0.9 test–retest reliability. Unlike Experiments 1 and 2, only ADO sessions were administered and each session consisted of 20 trials in Experiment 3. Shaded regions represent the 95% frequentist confidence interval of the concordance correlation coefficient (CCC).
Comparison of ADO and Staircase (SC) methods in their reliability, precision, and efficiency (see “Methods” for their definitions) of estimating temporal discounting rates (log(k)).
| Measures | ADO | Staircase (SC) | |
|---|---|---|---|
| Reliability: Maximum test–retest reliability (TRR) | Experiment 1 (College students), Visit 1 | 0.961 | 0.903 |
| Experiment 1 (College students), Visit 2 | 0.982 | 0.946 | |
| Experiment 2 (Patients w/ SUDs) | 0.973 | 0.892 | |
| Experiment 3 (Amazon Mturk)a | 0.965 | N/A | |
| Precision: Within-subject variability (SD of individual parameters) | Experiment 1 (College students), Visit 1 | 0.122 (0.105) | 0.413 (0.252) |
| Experiment 1 (College students), Visit 2 | 0.098 (0.070) | 0.537 (0.409) | |
| Experiment 2 (Patients w/ SUDs) | 0.073 (0.063) | 0.371 (0.180) | |
| Experiment 3 (Amazon Mturk)a | 0.339 (0.262) | N/A | |
| Efficiency: Trials required to reach 0.9 test–retest reliability | Experiment 1 (College students), Visit 1 | 7 | Failed to reach 0.9 even after 42 trials |
| Experiment 1 (College students), Visit 2 | 6 | 39 | |
| Experiment 2 (Patients w/ SUDs) | 11 | 27 | |
| Experiment 3 (Amazon Mturk)a | 11 | N/A | |
aExcept for Experiment 3 (Amazon Mturk participants), all experiments used 42 trials per session. In the Amazon Mturk experiment, there were 20 trials per session.