| Literature DB >> 24409112 |
Matthew D Turner1, Chayan Chakrabarti2, Thomas B Jones2, Jiawei F Xu2, Peter T Fox3, George F Luger2, Angela R Laird4, Jessica A Turner5.
Abstract
Identifying the experimental methods in human neuroimaging papers is important for grouping meaningfully similar experiments for meta-analyses. Currently, this can only be done by human readers. We present the performance of common machine learning (text mining) methods applied to the problem of automatically classifying or labeling this literature. Labeling terms are from the Cognitive Paradigm Ontology (CogPO), the text corpora are abstracts of published functional neuroimaging papers, and the methods use the performance of a human expert as training data. We aim to replicate the expert's annotation of multiple labels per abstract identifying the experimental stimuli, cognitive paradigms, response types, and other relevant dimensions of the experiments. We use several standard machine learning methods: naive Bayes (NB), k-nearest neighbor, and support vector machines (specifically SMO or sequential minimal optimization). Exact match performance ranged from only 15% in the worst cases to 78% in the best cases. NB methods combined with binary relevance transformations performed strongly and were robust to overfitting. This collection of results demonstrates what can be achieved with off-the-shelf software components and little to no pre-processing of raw text.Entities:
Keywords: CogPO; annotations; bioinformatics; data mining; multi-label classification; neuroimaging; text mining
Year: 2013 PMID: 24409112 PMCID: PMC3864256 DOI: 10.3389/fnins.2013.00240
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Characteristics of the data by dimension of the CogPO ontology and label sets.
| # | ||||||
|---|---|---|---|---|---|---|
| Behavioral domain | 40 | 1.846 | 8 | 0.429 | 0.413 | 9 |
| Cognitive paradigm class | 48 | 1.291 | 4 | 0.336 | 0.761 | 8 |
| Instruction type | 14 | 1.648 | 6 | 0.251 | 0.510 | 17 |
| Response modality | 5 | 1.308 | 3 | 0.036 | 0.700 | 21 |
| Response type | 9 | 1.324 | 4 | 0.069 | 0.696 | 10 |
| Stimulus modality | 5 | 1.150 | 3 | 0.036 | 0.858 | 25 |
| Stimulus type | 17 | 1.494 | 4 | 0.247 | 0.587 | 8 |
LCavg, LCmax = average and maximum number of labels per instance, respectively; PUNIQ = ratio of unique label combinations/sample size (247); Pmin = proportion with the minimum number of labels (always 1, in this dataset); k = value set for the kNN algorithm, see section 2.3.
Performance of SMO, NB, and kNN under the two problem transformation methods, label powerset (LP) and binary relevance (BR).
| Behavioral domain | 0.374 25.0% | 0.285 14.6% | 0.437 24.1% | 0.350 08.5% | ||
| Cognitive paradigm class | 0.404 37.5% | 0.187 17.0% | 0.416 28.3% | 0.262 11.7% | ||
| Instruction type | 0.475 36.5% | 0.390 26.8% | 0.494 25.9% | 0.488 20.2% | ||
| Response modality | 0.733 51.0% | 0.636 48.2% | 0.740 47.4% | 0.698 41.7% | ||
| Response type | 0.689 51.8% | 0.619 41.6% | 0.702 44.5% | 0.656 33.2% | ||
| Stimulus modality | 0.838 78.1% | 0.741 68.1% | 0.814 72.4% | 0.768 65.2% | ||
| Stimulus type | 0.439 30.7% | 0.317 16.9% | 0.387 21.0% | 0.368 16.5% | ||
All results are based on the abstract alone corpus. Decimals are F-micro scores and percentages are exact matches. The strict winner for each transformation-label dimension combination is highlighted. See text for details.
Cross-corpora comparison experiment.
| Behavioral domain | 0.534 | 0.501 | 0.440 | 0.448 | |
| Cognitive paradigm class | 0.464 | 0.464 | 0.420 | 0.394 | |
| Instruction type | 0.534 | 0.498 | 0.488 | 0.456 | |
| Response modality | 0.744 | 0.731 | 0.710 | 0.694 | |
| Response type | 0.706 | 0.699 | 0.660 | 0.662 | |
| Stimulus modality | 0.814 | 0.794 | 0.770 | 0.805 | |
| Stimulus type | 0.478 | 0.470 | 0.410 | 0.430 |
Table presents F1-micro values (see text) for naive Bayes under the binary relevance transformation across the five corpora that vary the feature space: words from (1) abstracts, titles, and MeSH keywords; (2) words from abstract text alone; (3) words from both titles and MeSH keywords; (4) title words alone; (5) MeSH keywords alone. Highest F1-micro highlighted in boldface; this does not indicate statistical significance. See text for details.
Characteristics of several multi-label data sets compared with ours.
| StimModAbs | 4,449,705 | 247 | 5 | 3603 | 14.59 | 1.15 | 0.036 | 0.008 |
| CogParaAll | 46,451,808 | 247 | 48 | 3919 | 15.87 | 1.13 | 0.336 | 0.004 |
| Medical | 63,770,490 | 978 | 45 | 1449 | 1.48 | 1.25 | 0.096 | 0.158 |
| Slashdot | 89,777,116 | 3782 | 22 | 1079 | 0.29 | 1.18 | 0.041 | 0.139 |
| Enron | 90,296,206 | 1702 | 53 | 1001 | 0.59 | 3.38 | 0.442 | 0.096 |
Values taken from Read et al. (2011); see there for details and sources. For notation, see section 2.1.2 and 2.2. Included are the values for the least and most complex data sets included in this paper.
Abstract alone corpus; stimulus modality labels.
Abstract, title, and keyword corpus; cognitive paradigm class labels.