| Literature DB >> 29921829 |
Abstract
Statistical learning (SL) is a method of learning based on the transitional probabilities embedded in sequential phenomena such as music and language. It has been considered an implicit and domain-general mechanism that is innate in the human brain and that functions independently of intention to learn and awareness of what has been learned. SL is an interdisciplinary notion that incorporates information technology, artificial intelligence, musicology, and linguistics, as well as psychology and neuroscience. A body of recent study has suggested that SL can be reflected in neurophysiological responses based on the framework of information theory. This paper reviews a range of work on SL in adults and children that suggests overlapping and independent neural correlations in music and language, and that indicates disability of SL. Furthermore, this article discusses the relationships between the order of transitional probabilities (TPs) (i.e., hierarchy of local statistics) and entropy (i.e., global statistics) regarding SL strategies in human's brains; claims importance of information-theoretical approaches to understand domain-general, higher-order, and global SL covering both real-world music and language; and proposes promising approaches for the application of therapy and pedagogy from various perspectives of psychology, neuroscience, computational studies, musicology, and linguistics.Entities:
Keywords: Markov model; domain generality; entropy; implicit learning; information theory; n-gram; order; statistical learning; uncertainty; word segmentation
Year: 2018 PMID: 29921829 PMCID: PMC6025354 DOI: 10.3390/brainsci8060114
Source DB: PubMed Journal: Brain Sci ISSN: 2076-3425
Figure 1Example of n-gram and Markov models in statistical learning (SL) of language (a) and music (b) based on information theory. The top are examples of sequences, and the others explain how to calculate TPs (P(e+1|e)) based on zero- to second-order Markov models. They are based on the conditional probability of an event e+1, given the preceding n events based on Bayes’ theorem. For instance, in language ((a), This is a sentence), the second-order Markov model represents that the “a” can be predicted based on the last subsequent two words of “This” and “is”. In music ((b), C4, D4, E4, F4), second-order Markov model represents that the “E” can be predicted based on the last subsequent two tones of “C” and “D”.
Figure 2SL models and the sequences used in neural studies. All of the models and paradigms in sequences based on concatenation of words (a), Markov model of tone (b) and word (c), and concatenation of words with different TPs of the last stimuli in words (d) are simplified so that the characteristics of paradigms can be compared. In the example of word-segmentation paradigm (a), the same words do not successively appear. TP—transitional probability.
Figure 3The entropy (uncertainty) of predictability in the framework of SL. The uncertainties depend on (a) TP ratios in a first-order Markov model (i.e., bigram model) and (b) orders of models in the TP ratio of 10% vs. 90%.
Figure 4Representative equivalent current dipole (ECD) locations (dots) and orientations (bars) for the N100 m responses superimposed on the magnetic resonance images (a) (Daikoku et al., 2014 [32]; and the SL effects (b) (Daikoku et al., 2015 [10]) (NS = not significant). When the brain encodes the TP in a sequence, it expects a probable future stimulus with a high TP and inhibits the neural response to predictable stimuli. In the end, the SL effects manifest as a difference in amplitudes of neural responses to stimuli with lower and higher TPs (b).
Overview of neurophysiological correlations with auditory statistical learning. TP—transitional probability; ABR—auditory brainstem response; MMN—mismatch negativity; STS—superior temporal sulcus; STG—superior temporal gyrus; IFG—inferior frontal gyrus; PMC—premotor cortex; PTC—posterior temporal cortex.
| Paradigms | Order of TP | Neural Correlates | References |
|---|---|---|---|
| Word segmentation | First-order | ABR | Skoe et al., 2015 [ |
| P50 | Paraskevopoulos et al., 2012 [ | ||
| N100 | Sanders et al., 2002 [ | ||
| MMN | Koelsch et al., 2016 [ | ||
| P200 | De Diego Balaguer et al., 2007 [ | ||
| N200–250 | Mandikal Vasuki et al., 2017 [ | ||
| P300 | Batterink et al., 2015 [ | ||
| N400 | Cunillera et al., 2009 [ | ||
| STS, STG | Farthouat et al., 2017 [ | ||
| Left IFG | Abla and Okanoya, 2008 [ | ||
| PMC | Cunillera et al., 2009 [ | ||
| Hippocampus | Schapiro et al., 2014 [ | ||
| Markov model | First-order | P50 | Daikoku et al., 2016 [ |
| Wernicke’s area | Bischoff-Grethe et al., 2000 [ | ||
| Hippocampus | Harrison et al., 2006 [ | ||
| Higher-order | P50 | Daikoku et al., 2017 [ | |
| N100 | Furl et al., 2011 [ | ||
| P200 | Furl et al., 2011 [ | ||
| Right PTC | Furl et al., 2011 [ |