| Literature DB >> 29335643 |
Pasi Saari1, Iballa Burunat2, Elvira Brattico3, Petri Toiviainen2.
Abstract
Pattern recognition on neural activations from naturalistic music listening has been successful at predicting neural responses of listeners from musical features, and vice versa. Inter-subject differences in the decoding accuracies have arisen partly from musical training that has widely recognized structural and functional effects on the brain. We propose and evaluate a decoding approach aimed at predicting the musicianship class of an individual listener from dynamic neural processing of musical features. Whole brain functional magnetic resonance imaging (fMRI) data was acquired from musicians and nonmusicians during listening of three musical pieces from different genres. Six musical features, representing low-level (timbre) and high-level (rhythm and tonality) aspects of music perception, were computed from the acoustic signals, and classification into musicians and nonmusicians was performed on the musical feature and parcellated fMRI time series. Cross-validated classification accuracy reached 77% with nine regions, comprising frontal and temporal cortical regions, caudate nucleus, and cingulate gyrus. The processing of high-level musical features at right superior temporal gyrus was most influenced by listeners' musical training. The study demonstrates the feasibility to decode musicianship from how individual brains listen to music, attaining accuracy comparable to current results from automated clinical diagnosis of neurological and psychological disorders.Entities:
Mesh:
Year: 2018 PMID: 29335643 PMCID: PMC5768727 DOI: 10.1038/s41598-018-19177-5
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Accuracy (%; mean and 68% (dark color), 95% (light color) confidence intervals across the cross-validation runs) with different number of regions.
Z-values and p-values for the log-likelihood ratios averaged across the cross-validation runs (p < 0.05 in boldface).
| Region | Z-value | p-value |
|---|---|---|
| Caudate nucleus L | −0.122 | 0.549 |
| Caudate nucleus R | 0.428 | 0.334 |
| Middle frontal gyrus R | 0.008 | 0.497 |
| Middle frontal gyrus, orbital part R | 1.000 | 0.159 |
| Anterior cingulate and paracingul. gyrus L | 1.901 |
|
| Anterior cingulate and paracingul. gyrus R | 2.229 |
|
| Inferior frontal gyrus, opercular part R | 2.225 |
|
| Inferior frontal gyrus, triangular part R | 0.299 | 0.382 |
| Superior temporal gyrus R | 1.654 |
|
Figure 2Musicianship discrimination in the nine discriminative regions–Z-values for the group differences of log-likelihood ratios averaged across cross-validation runs.
The average decoding accuracy (%) using the regions from the left (LH), right (RH), and both hemispheres. The statistical significant differences to the chance classification rate are typed in boldface.
| Region | LH | LH + RH | RH |
|---|---|---|---|
| ACG | 57.78 | 61.39 |
|
| IFGoper | 48.06 | 56.67 |
|
| STG | 50.83 | 57.78 | 56.94 |
Figure 3Distributions of feature beta coefficients for region/feature combinations yielding the most significant differences between the group means as shown by the two sample t-tests. Significance from one sample t-tests for the groups are marked with *(p < 0.05) and **(p < 0.01) after the group labels.
Participant demographics (Mus = musicians, NMus = nonmusicians).
| group | N | age | gender | hand | soc-eco status | WAIS-III PSI | music listening (h/week) | musical training (years) | instrument playing (years) |
|---|---|---|---|---|---|---|---|---|---|
| Mus | 18 | 28.2±7.8 | 9F | 18R | 43.6 | 116.3 | 18.2±11.2 | 15±4.7 | 21.2±6.2 |
| NMus | 18 | 29.2±10.7 | 10F | 17R | 35.4 | 115.7 | 12.4±6.7 | 1.6±2.2 (n = 8) | 2.1±3.0 (n = 9) |
Figure 4Different stages of the decoder training: participant-specific region time series encoding with linear regression, statistical musicianship group modeling with multivariate normal distributions, feature extraction with log-likelihood ratios between groups; and classification. The first three stages are employed for multiple regions, and classicifation is done based on the obtained features related to these regions. The region selection stage is excluded in the visualization. An example participant held out from the decoder training, shown in grey, is classified as a musician.