| Literature DB >> 23847528 |
Russell A Poldrack1, Deanna M Barch, Jason P Mitchell, Tor D Wager, Anthony D Wagner, Joseph T Devlin, Chad Cumba, Oluwasanmi Koyejo, Michael P Milham.
Abstract
The large-scale sharing of task-based functional neuroimaging data has the potential to allow novel insights into the organization of mental function in the brain, but the field of neuroimaging has lagged behind other areas of bioscience in the development of data sharing resources. This paper describes the OpenFMRI project (accessible online at http://www.openfmri.org), which aims to provide the neuroimaging community with a resource to support open sharing of task-based fMRI studies. We describe the motivation behind the project, focusing particularly on how this project addresses some of the well-known challenges to sharing of task-based fMRI data. Results from a preliminary analysis of the current database are presented, which demonstrate the ability to classify between task contrasts with high generalization accuracy across subjects, and the ability to identify individual subjects from their activation maps with moderately high accuracy. Clustering analyses show that the similarity relations between statistical maps have a somewhat orderly relation to the mental functions engaged by the relevant tasks. These results highlight the potential of the project to support large-scale multivariate analyses of the relation between mental processes and brain function.Entities:
Keywords: classification; data sharing; informatics; metadata; multivariate
Year: 2013 PMID: 23847528 PMCID: PMC3703526 DOI: 10.3389/fninf.2013.00012
Source DB: PubMed Journal: Front Neuroinform ISSN: 1662-5196 Impact factor: 4.081
Figure 1An overview of the draft data organization scheme for the OpenFMRI project. A schematic of the directory tree for a dataset is presented, with each subdirectory shown on a separate branch. This structure allows specification of an arbitrary set of tasks, runs, and statistical models. The key files included in the base directory for the dataset specify features that are consistent across all of the data (such as demographics, task naming, and scan ordering), while key files in subdirectories specify details that may change across models or runs.
List of processing steps applied to data, and tools used for each operation.
| Motion correction | mcflirt (FSL) |
| Brain extraction (highres) | Freesurfer |
| Brain extraction (BOLD) | bet (FSL) |
| Quality assurance and generation of confound files | fmriqa (custom) |
| Creation of design files | custom code |
| First-level (within-run) statistical modeling | feat (FSL) |
| Second-level statistical modeling (for multi-run datasets) | feat (FSL) |
| Group statistical modeling | feat (FSL) |
| Cortical surface generation and parcellation (highres) | Freesurfer |
List of datasets used in the preliminary analyses below.
| 1 | ds001 | 1 | Balloon analog risk task: Parametric pump effect vs. control | Schonberg et al., |
| 2 | ds002 | 1 | Classification learning task: Task vs. baseline | Aron et al., |
| 3 | ds002 | 2 | Classification learning task: Feedback vs. baseline | Aron et al., |
| 4 | ds002 | 3 | Classification decision: Task vs. baseline | Aron et al., |
| 5 | ds003 | 1 | Rhyme judgment: Task vs. baseline | Xue and Poldrack, |
| 6 | ds005 | 1 | Mixed-gambles task: Parametric gain response | Tom et al., |
| 7 | ds006A | 1 | Mirror reading task: Mirror-reversed vs. plain items | Jimura et al., in preparation |
| 8 | ds007 | 1 | Stop signal task: Letter classification vs. baseline | Xue et al., |
| 9 | ds007 | 2 | Stop signal task: Letter naming vs. baseline | Xue et al., |
| 10 | ds007 | 3 | Stop signal task: Pseudoword naming vs. baseline | Xue et al., |
| 11 | ds008 | 1 | Stop signal task: Successful stop vs. baseline | Aron et al., |
| 12 | ds008 | 2 | Conditional stop signal task: Successful stop vs. baseline | Aron et al., |
| 13 | ds011 | 1 | Tone-counting task: Task vs. baseline | Foerde et al., |
| 14 | ds011 | 2 | Single-task classification learning: Task vs. baseline | Foerde et al., |
| 15 | ds011 | 3 | Dual-task classification learning: Task vs. baseline | Foerde et al., |
| 16 | ds011 | 4 | Classification decision: Task vs. baseline | Foerde et al., |
| 17 | ds017A | 2 | Conditional stop signal task: Go-critical vs baseline | Rizk-Jackson et al., unpublished |
| 18 | ds051 | 1 | Abstract-concrete task: novel vs. repeated words | Alvarez and Poldrack, unpublished |
| 19 | ds052 | 1 | Classification learning task: Positive feedback vs. baseline | Poldrack et al., |
| 20 | ds052 | 2 | Classification reversal learning task: Negative feedback vs. baseline | Poldrack et al., |
| 21 | ds101 | 1 | Simon task: incorrect vs. correct | Kelley and Milham, unpublished |
| 22 | ds102 | 1 | Flanker task: incongruent vs. congruent | Kelly et al., |
| 23 | ds107 | 1 | One-back task: words vs. consonants | Duncan et al., |
| 24 | ds108 | 1 | Emotion regulation task: Regulate-negative vs. Look-negative | Wager et al., |
| 25 | ds109 | 1 | False belief task: False belief story vs. false picture story | Moran et al., |
| 26 | ds110 | 1 | Incidental memory encoding task with cueing: Valid cue high confidence hits vs. misses | Uncapher et al., |
Accession and task numbers refer to the specific descriptors used in the database.
Figure 2Rendered maps of the voxels with significant loadings on the 20 ICA components identified statistical images for the datasets listed in Table . Each column displays the loading for a single component; voxels shown in red had positive loading for that component.
Figure 3Classification accuracy using reduced-dimension data, as a function of the number of ICA components in the dimensionality reduction step. Dimensionality reduction was first performed on an independent set of data (from run 2 for each subject), and the data from run 1 were then projected onto those components. Reported accuracy (thick lines) reflects average accuracy of task classification across all data points, from a total of 25 possible labels. The thin lines at the bottom reflect the empirically derived 95% cutoff for the null hypothesis of chance accuracy, obtained by performing the same classification 100 times with randomized labels, and taking the 95th largest value. RBF, radial basis function; SVM, support vector machine.
Figure 4Hierarchical clustering of statistical maps across tasks, after projection into the 20-dimensional ICA space depicted in Figure .