Literature DB >> 30452442

Recruitment of the motor system during music listening: An ALE meta-analysis of fMRI data.

Chelsea L Gordon1, Patrice R Cobb2, Ramesh Balasubramaniam1.   

Abstract

Several neuroimaging studies have shown that listening to music activates brain regions that reside in the motor system, even when there is no overt movement. However, many of these studies report the activation of varying motor system areas that include the primary motor cortex, supplementary motor area, dorsal and ventral pre-motor areas and parietal regions. In order to examine what specific roles are played by various motor regions during music perception, we used activation likelihood estimation (ALE) to conduct a meta-analysis of neuroimaging literature on passive music listening. After extensive search of the literature, 42 studies were analyzed resulting in a total of 386 unique subjects contributing 694 activation foci in total. As suspected, auditory activations were found in the bilateral superior temporal gyrus, transverse temporal gyrus, insula, pyramis, bilateral precentral gyrus, and bilateral medial frontal gyrus. We also saw the widespread activation of motor networks including left and right lateral premotor cortex, right primary motor cortex, and the left cerebellum. These results suggest a central role of the motor system in music and rhythm perception. We discuss these findings in the context of the Action Simulation for Auditory Prediction (ASAP) model and other predictive coding accounts of brain function.

Entities:  

Mesh:

Year:  2018        PMID: 30452442      PMCID: PMC6242316          DOI: 10.1371/journal.pone.0207213

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

In the case of (most) music, we do not merely passively receive temporal patterns, but actively engage with the sound stream by discerning an underlying periodicity. This profound shaping of temporal perception is central to understanding and participation in music, dance and even speech/conversation. In recent years, neuroimaging studies have shown that passively listening to music activates brain regions that reside in the motor system proper. The same neural correlates underlying the creation of music and moving to music appear to be involved even when one is only listening to a musical piece [1-7]. The motor system has received increasing attention in non-purely-motor domains [8-12]. Activity in motor regions during perception of human actions and language is ubiquitous. In early theories of cognitive processing, motor processes and perceptual processes were understood as entirely separate and encapsulated mechanisms [13]. As evidence accumulates that shows this is not the case and there is substantial overlap among the domains, theories of action production and action perception must be informed accordingly. More recent proposals argue for common coding of perceptual and motor information [14, 15] that arises primarily due to the co-activation of perceptual and motor components of a given action. Anderson’s [16] theory of neural reuse additionally suggests that we should expect newly evolved functions such as language to make use of previously instantiated neural mechanisms whose computational functionality can be co-adapted for new purposes. As such, it is likely that motor activation observed during speech perception, for instance, corresponds to a sharing of computational or functional resources for perception and production of a given speech sound. We can expect a similar sharing of resources for music production and music perception as well. Patel and Iversen [17] advanced a theory of motor activation during music perception called the ASAP (Action Simulation for Auditory Prediction) hypothesis. The theory suggests that the same neural underpinnings involved in the simulation of body movements are utilized by the motor planning system to entrain neural activation with musical beat. This entrainment allows simulations to be used online during music listening as a predictive sensory signal for the upcoming music beat. The simulation is not tied to a particular effector-based movement, but a simulation of a timed, rhythmic motion. Patel and Iversen suggest the dorsal auditory stream as a potential underlying neural pathway for this process [18]. Rauschecker [19] has also proposed that a unified function of the dorsal stream may be anticipatory control of sensorimotor events. In particular, he suggests the posterior superior temporal (ST) regions, along with the inferior parietal lobe (IPL), interface extensively with PMC, linking the motor and auditory regions to established sensorimotor networks for audiomotor processes, such as speech and music. This network is established through similar mechanisms to those used in motor control theories [20, 21], where a feed-forward projection carrying an efference copy of a motor command is used as a prediction of the upcoming sensory consequences, which can then be compared with the actual sensory outcome of the motor act. Rauschecker proposes that the projection from inferior frontal gyrus (IFG) to vPMC is a likely candidate for carrying an efference copy, while IPL to posterior ST would carry an “afference” copy of the predicted motor signal, thus allowing a continuous audiomotor predictive loop underlying smooth perceptuomotor abilities. Another candidate theory, suggested by Schubotz [11], is of active sensory prediction of events using the motor system. Schubotz extends the idea of emulators from motor control theory to encompass inanimate event perception in addition to human action prediction. Schubotz suggests the following: when we repeatedly hear a melody, the lateral PMC builds up sensorimotor representations using input from association areas of the cortex. These sensorimotor representations are only audiomotor, lacking the proprioceptive-motor representations that are involved in an actual performed movement. The lateral PMC eventually establishes an internal model of this melody which can be used for perceptual prediction. This internal model is similar to that involved in motor control, but with the components for movement and sensory feedback removed. Schubotz proposes what she calls the HAPEM (Habitual Pragmatic Event Map) framework, which states that “the prediction of an event that is structured with regard to a property P engages the area of the lateral premotor cortex that is best adapted to specify its motor output in terms of property P”. What this means is that the perception of events with different properties recruits particular somatotopic regions of vPMC, selected based on similarity to the underlying properties of that area of vPMC. For instance, the regions of vPMC that correspond to executing and observing mouth movements are recruited for the perception of rhythmic events, due to the underlying rhythmic nature of the vocal system. The above theories all posit that cortical motor areas play a role in music listening. Another emerging theme is that of the motor system having a predictive role in perceptual processes. These accounts are primarily in agreement in terms of which sub-areas in the motor system are involved. Schubotz’s framework directly proposes involvement of both lateral PMC and pre-SMA/SMA, while ASAP and Rauschecker’s theory both set the dorsal auditory stream (which includes dPMC) as the primary substrate. However, activated regions within the motor system measured by neuroimaging methods tend to vary between research studies. For instance, numerous music listening experiments report motor activity in both supplemental motor area (SMA) and dorsal premotor cortex (dPMC) [1– 6]. Among these studies, a few show neural activations in cerebellum [2, 4, 5] or primary motor cortex (M1) [7] during a music listening task. Said differently, most music-listening studies do not show activation in every region of the motor system, nor do they show uniform activation in any one part of the motor system. In order to gain insight into the functional contribution of the motor system to passive music perception, one necessary step is to determine which motor regions are consistently contributing across music listening instances. There are many factors likely to contribute to differences across studies, as each individual experiment has its own musical stimuli that vary in terms of particular characteristics, such as rhythmicity, familiarity, and valence of the music, for instance. Stimuli consisting of highly regular rhythmic structure might engage brain regions important for timing and sequential structure (i.e., supplementary and pre-supplementary motor areas and the cerebellum), while others might not. Experiments also vary in terms of what a participant is directed to focus on in these paradigms, ranging from complete passive listening (not attending) to judging beat or other characteristics of the stimuli. Such task demands are also likely to influence which regions are active, as directing attention to a stimulus may encourage focusing on particular aspects of the music, such as its beat or rhythmicity. In the present study, we are interested in discovering what motor regions are engaged during all music perception—those activated during passive listening. We define passive listening as attentive listening while remaining still (i.e., not tapping along to the music). Identifying which regions are active consistently across all music listening tasks would help gain insight into the underlying processes and hone existing theories. Many theories outlining the functional contribution of individual motor areas exist, which can be used to determine what particular function is being carried out in a task utilizing that motor region. If one critical component is the dorsal auditory stream, which has a proposed role in motor planning and mapping auditory information onto potential motor acts, we should observe observation in dPMC [22, 23]. If activation is found in vPMC, the underlying mechanism might be similar to that proposed in the action observation network, which is responsible for mirror system activity for observed and produced actions [24, 25, 26]. Many studies that involve music with beat manipulation report activity in SMA and pre-SMA regions, which are presumed to be important for sequential processing of action-related stimuli and for inhibition of movements, respectively [27, 28]. Thus, SMA activity might indicate processing of sequential aspects of the music, and pre-SMA the inhibition of the natural tendency to move or sway to the music. We also might observe activation of structures in the basal ganglia, which appear to be involved in beat perception [29]. The basal ganglia are important for movement timing and sequential movement execution [30, 31]. M1 activation corresponds to particular motor commands that are carried out by specific muscle groups [32] and has also been found active during observation of actions [33, 34, 35]. Finally, the cerebellum is known for its crucial role in motor timing and coordination. Research on sensorimotor adaptation has long focused on the role of the cerebellum in predicting sensory consequences of movement and adapting to errors in these predictions (for a more interactive view see [36, 37, 38, 39]). Furthermore, cerebellar activation in conjunction with hippocampal activity is thought to underlie spatiotemporal prediction of movements [40]. This implication in predictive processes of motor control might extend to imagined and simulated motor computations, e.g. the cerebellum might be active in musical prediction even when no direct motor control is required. In order to determine which of these regions show reliable and consistent activation during music perception, we employed a meta-analysis of all neuroimaging experiments consisting of music listening using an activation likelihood estimation (ALE) [41]. We predict that from this meta-analysis will emerge a pattern of activation that will enlighten and instruct future theories aiming to explain the motor-specific contributions to passive music perception. Activation of any of the motor regions will provide conclusive evidence for the involvement of the regions of the brain typically considered “action areas”, in the perceptual domain of passive music listening. This will inform theories about what roles are played by the traditional motor system.

Methods

Meta-analyses provide a formal, statistical integration to combine the results of several studies that address a set of related research questions. There are several methods available for the meta-analysis of neuroimaging data and careful consideration was given as to which was most appropriate for this study. First, our study aims were to synthesize neuroimaging data of studies comparing rest and passive listening. More specifically, we wanted to identify regions of consistent activation across studies. Activation likelihood estimation (ALE) meta-analysis [41] addresses this by treating the spatial relationship between within study foci as fixed effects and between study relationships as random effects. Secondly, we considered the characteristics of our dataset. Unlike some other methods (e.g., KDA and MKDA), ALE uses a Gaussian kernel. When several distinct foci are located within the same general area, the Gaussian kernel is most likely to recover the separate foci. And, in general, if the spatial error on peak locations is approximately Gaussian (a reasonable assumption), then the Gaussian kernel will likely yield the most sensitive results. To investigate our research questions, we conducted ALE meta-analysis. Imaging studies commonly report brain locations of task-induced activations as coordinates in 3D space (x,y, and z). ALE meta-analysis techniques can be used to identify reliable activation patterns in 3D space across studies. ALE is a coordinate-based approach to a meta-analysis, allowing researchers to integrate imaging data. Studies are collected, coded and interpreted using analytical methods to assess likelihood of activation through agreement or overlap in activation patterns. To perform the ALE meta-analysis, we began by first locating relevant studies. Relevant studies were those that utilized functional brain imaging of healthy subjects listening tasks. We conducted literature searches in Medline and the BrainMap database [42] using a combination of the following: (1) a functional brain imaging modality, including positron emission tomography (PET) and functional magnetic resonance imaging (fMRI) and (2) relevant adjectives related to auditory stimuli. For example, a single search consisted of “Imaging” AND “passive listening” OR “fMRI or functional magnetic resonance imaging” AND “auditory”. The literature search of Medline was performed February 2016 and returned 132,294 papers. The literature search of BrainMap was performed September 14, 2016 and returned 244 studies. To ensure our ability to investigate the specified research questions a subsequent study selection process was done by applying the following inclusion criteria to the studies: (1) subjects were healthy adult participants; (2) The analyses include contrasts against rest or a suitable low-level control condition; (3) peak coordinates of group-level activations were reported; (4) foci activation were available in the form of standardized stereotaxic coordinates in either Talairach or Montreal Neurological Institute (MNI) space; (5) that results from the entire scanned volume were reported; and (6) data were available as of September 2016. An effort was made to obtain unreported coordinates from selected studies meeting all other criteria, however, this effort did not return any results. The subsequent review process was performed in two phases. First, an automated review of study titles was done using the R environment (R Development Core Team, 2008) to remove studies that were not in healthy human subject populations. The automated review removed 8144 papers from the database. Next, reviewers read the abstract and/or methods sections of remaining studies to assess appropriateness using the above inclusion criteria. Fig 1 illustrates the full review process for the meta-analysis. The process yielded 42 experiments that met the criteria for inclusion. A full list of experiments included can be found in Table 1. Experiments included a total of 386 unique subjects, approximately 195 male and 171 female.
Fig 1

Flow diagram of study review.

Table 1

A list of the studies and experiments that were part of our meta-analysis.

ExperimentSubjFociComparisonInstructionsMusic TypeMusical TrainingAgeMaleHandedness
Alluri et al., 2012 [56]11215Listen vs. restRemain still and to relax while listening to the musical stimulus and to maintain their gaze on the screenModern tango (Adios Nonino by Astor Piazzolla)Mean years of music training 16.1 ± 6 SDMean age 23.2 ± 3.7 SD6 (55%)Unknown
Baumgartner et al., 2006 [57]927Listen and look at picture vs. fixation baselineInstructed the subjects to place themselves into the same mood as expressed by the presented emotional stimuliEmotional classical orchestra musicUnknownMean age 24.8; range 21–300 (0%)Right
Blood et al., 1999 [58]104Listen vs. baseline (acoustically matched noise bursts)Instructed to listen carefully. After the scan, subjects used a bipolar rating scale to rate emotional valence and intensity of stimuli.Novel emotional music with varying dissonanceNo more than amateur trainingUnknown5 (50%)Right
Blood et al., 1999 [58]108
Brown et al., 2004 [60]1021Listen vs. restInstructed to listenattentively to the music with their eyes closed withoutmaking any movement or response.Wordless, instrumental rembetika style songs (unfamiliar to participants)NonmusiciansMean age 33.8; range 21–515 (50%)Right
Brown et al., 2007 [59]1157Listen vs. rest; Listen and discrimination task vs. control (button press)Melody listening: listen with eyes closed. Discrimination task: Listen and button press.Piano melodies and harmonies, primarily adapted for this work.University music education majors with a mean of 5.0 years of formal music instruction in voice or instrument. Having had an average of 12.3 years of involvement in musical production.Mean age 24.6; range 19–465 (45%)Right
Caria et al., 2011 [61]1420Listen vs. silent controlInstructed to passively attend to music.Instrumental piecesNonmusiciansMean age 24.3 ± 3.02 SD6 (43%)Unknown
Chen et al., 2008; Exp 1 [5]1218Listen with anticipation vs. silent baselineListened attentivelyRhythmic musicNonmusiciansMean age 23.83; range 20–326 (50%)Right
Chen et al., 2008; Exp 2 (A) [5]12*17Listen with anticipation vs. silent baselinePassively listenRhythmic musicNonmusiciansMean age 24; range 19–346 (50%)Right
Chen et al., 2008; Exp 2 (B) [5]12*9Passive Listen vs. silent baselinePassively listenRhythmic musicNonmusiciansMean age 24; range 19–346 (50%)Right
Demorest et al., 2010 [62]1615Listen vs. restListen, followed by memory test.Three music examples from the Western classical tradition, three examples from the Turkish classical tradition and three examples from the Chinese classical tradition<1 year of private music lessons and <3 years of ensemble (e.g., choir and orchestra) participationMean age 28.6 years with a range of 20.1–45.1 years8 (50%)Right
Dobek et al., 2014 [63]1233Listen vs. baselineAdministered pain (thermal stimulation)Self-selected by participantsNon-musiciansRange 18–40 years0 (0%)Unknown
Flores-Gutierrez et al., 2007 [64]197**Music–noiseInstructed to remain attentively focused on theauditory stimuli as their only taskComplex emotional musical piecesNo formal musical trainingMean age 25 (SD = 3.05)11 (58%)Right
Grahn et al., 2007 [6]2712Music—restInstructed not to move any part of their body duringpresentation of the rhythms, followed by response given by button press to rhythm discrimination taskRhythmic sequencesFourteen out of 27 had musical training, defined as over 5 years of formal musical training and current regular musical activity and 13 had no musical training (reported no formal musical training or musical activities).Mean age 24.5; range 19–3819 (70%)Right
Habermeyer et al., 2009 (A) [65]16*8Listen vs. silent baselineWatch silent movie without paying attention to the presented soundsDeviant melodic patterns8 trained lifelong musicians; 8 nonmusiciansMean age 44.5 ± 9.9 years14 (88%)
Habermeyer et al., 2009 (B) [65]16*3Listen vs. silent baselineWatch silent movie without paying attention to the presented soundsStandard melodic patterns8 trained lifelong musicians; 8 nonmusiciansMean age 44.5 ± 9.9 years14 (88%)Unknown
Heine et al., 2015 [66]819Music vs. baseline soundsInstructed to keep their eyes closed, stay awake, avoid any structured thoughts, and listen attentively to the musicDynamic musical excerpts chosen by loved ones from a listUnknownMean age 26, SD ± 34 (50%)Unknown
Hugdahl et. al., 1999 [67]125Musical instruments–simple tonesButton press at target soundExcerpts from musical instrumentsUnknownRange 20–3012 (100%)Right
Langheim et al., 2002 [68]64Passive listening vs. RestPassive listeningClassical music (Vivaldi’s Concerto in G minor, Bach's Suite in C major, part 2, Partita 2 and Partita 3)At least 15 years of musical experience (two violinists, one pianist and three cellists); mean length of study 19.6 years, range 15–26 yearsMean age 27; range 22–322 (33%)Right
Leaver et al., 2009; (A) [69]109Familiar and unfamiliar musicSubjects were instructedto attend to the stimulus being presented and to imagine, but notvocalize, the subsequent melodyShort piano melodies constructed for this experimentAt least 2 years musical experience (mean = 6.5, sd = 4.17)UnknownUnknownUnknown
Leaver et al., 2009 (B) [69]93Familiar and unfamiliar musicSubjects were instructedto attend to the stimulus being presented and to imagine, but notvocalize, the subsequent melodyShort piano melodies constructed for this experimentNonmusiciansUnknown6 (67%)Unknown
Mirz et al., 1999 [70]57Music–baselineSubjects were asked to listen to thepresented sounds without performing any semantic,phonological, temporal, intensity, orpitch analysisClassical music (W.A. Mozart, Piano Concerto No. 21, 65 dB SPL)UnknownMean age 34; range 24–502 (40%)Right
Morrison et al., 2003 (A) [71]6*3Music vs. restFollowing the scan subjects completed a poststudyrecognition test3 Baroque Western examplesTrained professional violinists and violistsmean age 38.3 years2 (33%)2 left handed, 4 right handed
Morrison et al., 2003 (B) [71]6*3Music vs. restFollowing the scan subjects completed a poststudyrecognition test3 Chinese examplesTrained professional violinists and violistsmean age 38.3 years2 (33%)2 left handed, 4 right handed
Morrison et al., 2003 (C) [71]6*2Music vs. restFollowing the scan subjects completed a poststudyrecognition test3 Baroque Western examplesNon-musiciansmean age 34.2 years2 (33%)Right
Morrison et al., 2003 (D) [71]6*2Music vs. restFollowing the scan subjects completed a poststudyrecognition test3 Chinese examplesNon-musiciansmean age 34.2 years2 (33%)Right
Ohnishi et. al., 2001 (A) [72]145Music vs. restInstructed to passively listen to musicItalian concert BMV 989 by J.S. Bach>12 years of 4–8 h of training per day) with AP (n = 10) or relative pitch (n = 4)Range 20–272 (14%)Right
Ohnishi et. al., 2001 (B) [72]144Music vs. restInstructed to passively listen to musicItalian concert BMV 989 by J.S. BachNonmusicians (no formal education musical and never played an instrument)Range 21–272 (14%)Right
Rogalsky et. al., 2011 [73]205Melodies vs. restPassive listeningSimple novel piano melodiesTwelve participants had some formal musical training (mean years of training = 3.5, range 0–8)Mean age 22.6 years; range 18–319 (45%)Right
Satoh et al., 2001 [75]9*8Music (alto) vs. baselineSubjects were asked to listening to and concentrate on the tone of the alto part of the harmony, and make a sign when they heard the tonic tone3 fairly unknown motets; musical pieces of harmonious style with four vocal parts, composed by Anton Bruckner.Musicians (music students)Mean age 21.8 years; range 21–289 (100%)Right
Satoh et al., 2001 [75]9*10Music (harmony) vs. baselineSubjects were asked to listen to the melody as a whole, and make a sign upon hearing the minor chord3 fairly unknown motets; musical pieces of harmonious style with four vocal parts, composed by Anton Bruckner.Musicians (music students)Mean age 21.8 years; range 21–289 (100%)Right
Satoh et. al., 2003 [76]11*7Music (soprano) vs. baselineSubjects were asked to listen to the soprano part of the harmony, and make a sign when they regarded a tonal sequence as one phraseThree new musical pieces of harmonious style with three vocal partsNonmusicians (no formal musical education or training)Mean age 21.2 years; range 20–3011 (100%)Right
Satoh et. al., 2003 [76]11*10Music (harmony) vs. baselineSubjects were asked to listen to the melody as a whole, and make a sign upon hearing a dissonant chordThree new musical pieces of harmonious style with three vocal partsNonmusicians (no formal musical education or training)Mean age 21.2 years; range 20–3011 (100%)Right
Satoh et. al., 2006 [74]10*16Music (familiarity) vs. baselineSubjects were asked to listen to the melodies and then judge whether the melody was familiar33 melodies (27 melodies were well-known old Japanese nursery songs)Nonmusicians (no formal musical education or training)Mean age 21.6; range 20–2810 (100%)Right
Satoh et. al., 2006 [74]10*13Music (alteration-detecting task) vs. baselineSubjects were asked to listen to the same melodies and detect the altered notes by making a sign33 melodies (27 melodies were well-known old Japanese nursery songs)Nonmusicians (no formal musical education or training)Mean age 21.6; range 20–2810 (100%)Right
Schmithorst, 2005 [77]1530Melodies—random tonesPassive listening30 s of an unharmonized popular melody, followed by 30 s of tones of random frequency and duration, followed by 30 s of the previous melody, harmonized using triads an octave below7 out of 15 received prior formal musical training, receiving formal instruction, continuously from early childhood (8 years old) throughout adolescenceMean age 37.8 ± 15.2 SD11 (73%)Unknown
Toiviainen et al., 2014 [78]1538Comprised the B-side of the album Abbey Road by The Beatles (1969).UnknownMean age 25.7 ± 5.2 SD10 (67%)Right
Trost et al., 2011 [79]1520Music vs. random tonesSubjects were asked to listen closely and provided a rating of emotional feeling following the music pieceEmotional classical musicNo professional music experienceMean age 28.8 +- 9.98 (53%)Right
Tsai et al., 2010 [49]12*7Music—baselineSubjects were asked to passively listen to unlearned percussion musicSichuan opera percussion music, Beijing opera percussion music, syllable representation of Beijing opera percussion music, and Taiwanese opera tunes played by the erhuMusic training for more than 4 yearsRange: 20–262 (17%)Right
Tsai et al., 2010 [49]12*7Music—baseline noiseSubjects were asked to listen and hum covertly along to learned percussion musicSichuan opera percussion music, Beijing opera percussion music, syllable representation of Beijing opera percussion music, and Taiwanese opera tunes played by the erhuMusic training for more than 4 yearsRange: 20–262 (17%)Right
Tsai et al., 2010 [49]12*7Music—baseline noiseSubjects were asked to listen and hum covertly along to the verbalized syllable representation of learned percussion musicSichuan opera percussion music, Beijing opera percussion music, syllable representation of Beijing opera percussion music, and Taiwanese opera tunes played by the erhuMusic training for more than 4 yearsRange: 20–262 (17%)Right
Tsai et al., 2010 [49]12*7Music—baseline noiseSubjects were asked to listen and hum covertly along to the verbalized syllable representation of learned melodic musicSichuan opera percussion music, Beijing opera percussion music, syllable representation of Beijing opera percussion music, and Taiwanese opera tunes played by the erhuMusic training for more than 4 yearsRange: 20–262 (17%)Right

* The same pool of participants was used for separate analysis/study protocols. These were considered separate experiments for the purposes of this meta-analysis because analyses were performed separately and/or the dependent variable was altered between conditions.

** The published data was missing one z coordinate. An attempt was made to contact the authors, however, we were unable to obtain the missing information.

* The same pool of participants was used for separate analysis/study protocols. These were considered separate experiments for the purposes of this meta-analysis because analyses were performed separately and/or the dependent variable was altered between conditions. ** The published data was missing one z coordinate. An attempt was made to contact the authors, however, we were unable to obtain the missing information. Coordinates (X, Y, Z) for selected studies were recorded and, where necessary, transformed to Talairach space. Coordinates from individual studies were transferred to a text file formatted for analysis in GingerALE 2.3.6 (http://www.brainmap.org/ale/; Research Imaging Center, University of Texas, San Antonio, TX). These were transferred either using brainmap’s Sleuth software (if the studies were located in the brainmap database), which outputs coordinates in the correct format for GingerALE, or were transferred individually by hand. The ALE meta-analysis was carried out in GingerALE. The ALE procedure was as follows: (1) model of single-study activation foci as peaks of three-dimensional Gaussian probability densities with subject-based full-width at half-maximum values [43]; (2) summation of probability densities to produce a statistical map estimating the likelihood of activation at each voxel; (3) thresholding of this ALE map based on the null hypothesis of a uniform distribution of foci; (4) correcting for multiple comparisons by family-wise error thresholding. Resulting statistical maps show clusters where convergence between foci is greater than would be expected by chance. Statistical maps were thresholded using cluster-level family-wise error correction P<0.05 (cluster-forming threshold voxel-level P<0.001). We split the data into separate studies that used either musicians only or nonmusicians only, with the intention of performing a contrast analysis between the two groups. Unfortunately, there were too few studies in these groups individually (14 experiments in each group), so we were unable to complete this contrast.

Results

Fig 2 shows the activations during passive listening, demonstrating the common brain network underlying music perception. Talairach coordinates for these ALE foci are presented in Table 2. Activations were seen in the bilateral superior temporal gyrus, transverse temporal gyrus, insula, pyramis, bilateral precentral gyrus, and bilateral medial frontal gyrus. As shown in Fig 2, there was activation in the left and right premotor cortex (BA 6), right primary motor cortex (BA 4), and the left cerebellum.
Fig 2

Significant clusters from meta-analysis of passive listening tasks in healthy volunteers (family-wise error correction (P<0.05)).

The 3D brain is shown to indicate slice levels.

Table 2

Talairach coordinates for voxel clusters.

AreaBAConjunction
xyzALE
Temporal Lobe
Superior Temporal GyrusR2252-1660.061329
R2252-6-40.054407
L41-52-1860.069281
L41-42-34120.033931
Frontal Lobe
Precentral GyrusR450-4460.052819
Anterior Lobe
Cerebellum (Culmen)L-28-60-260.045433

The Talairach coordinates of the significant ALE clusters are presented for the conjunction of passive listening (p < 0.05, FWE). The ALE values for the conjunction represent the minimum ALE value from the passive listening ALE maps. The ALE values shown are the true values times 10^-3. BA, Brodmann area.

Significant clusters from meta-analysis of passive listening tasks in healthy volunteers (family-wise error correction (P<0.05)).

The 3D brain is shown to indicate slice levels. The Talairach coordinates of the significant ALE clusters are presented for the conjunction of passive listening (p < 0.05, FWE). The ALE values for the conjunction represent the minimum ALE value from the passive listening ALE maps. The ALE values shown are the true values times 10^-3. BA, Brodmann area. An inspection of Table 3 reveals that Cluster 1 is centered over the right primary auditory cortex, and spans from BA 22 and BA 41/42 (primary and secondary auditory cortices) in the right hemisphere to BA 6 (right premotor cortex). Likewise, in the left hemisphere, cluster 2 is centered over the left primary auditory cortex, and spans from BA 22 and BA 41/42 (primary and secondary auditory cortices) in the left hemisphere to BA 6 (left premotor cortex). Cluster 3 reveals motor system activation in the right hemisphere, centered in right premotor cortex and spanning from premotor to primary motor cortex. Finally, cluster 4 is located in the left cerebellum. Fig 2 depicts the activation patterns seen bilaterally for a range of z values.
Table 3

Contributing foci and study making up each voxel cluster.

ClusterLabelTotal # of FociCluster SizeBAStudies Contributing to Cluster
1Right Superior Temporal Gyrus756336 mm32219 foci from Alluri, 20123 foci from Baumgartner, 20065 foci from Brown, 20045 foci from Brown, 20072 foci from Chen, 2008: Experiment 12 foci from Chen, 2008: Experiment 2 A1 foci from Chen, 2008: Experiment 2 B1 foci from Dobek, 20141 foci from Flores-Gutierrez, 20071 foci from Habermeyer, 20091 foci from Habermeyer, 20091 foci from Heine, 20151 foci from Hugdahl, 19991 foci from Langheim, 20021 foci from Leaver, 20091 foci from Leaver, 20091 foci from Mirz, 19991 foci from Morrison, 2003 A1 foci from Morrison, 2003 B1 foci from Morrison, 2003 C1 foci from Morrison, 2003 D2 foci from Ohnishi, 2001 A2 foci from Ohnishi, 2001 B2 foci from Rogalsky, 2011 B1 foci from Satoh, 20061 foci from Schmithorst, 20054 foci from Toiviainen, 20142 foci from Trost, 20113 foci from Tsai, 2010 A2 foci from Tsai, 2010 B2 foci from Tsai, 2010 C3 foci from Tsai, 2010 D
2Left Superior Temporal Gyrus625248 mm34115 foci from Alluri, 20121 foci from Baumgartner, 20061 foci from Blood, 19992 foci from Blood, 19991 foci from Brown, 20047 foci from Brown, 20071 foci from Chen, 2008: Experiment 11 foci from Chen, 2008: Experiment 2 A3 foci from Chen, 2008: Experiment 2 B2 foci from Demorest, 20101 foci from Flores-Gutierrez, 20071 foci from Grahn, 20071 foci from Habermeyer, 20091 foci from Habermeyer, 20091 foci from Heine, 20151 foci from Hugdahl, 19991 foci from Langheim, 20021 foci from Leaver, 20091 foci from Leaver, 20091 foci from Mirz, 19991 foci from Morrison, 2003 A1 foci from Morrison, 2003 B1 foci from Morrison, 2003 C1 foci from Morrison, 2003 D1 foci from Ohnishi, 2001 A1 foci from Ohnishi, 2001 B2 foci from Rogalsky, 2011 B3 foci from Toiviainen, 20142 foci from Tsai, 2010 A1 foci from Tsai, 2010 B2 foci from Tsai, 2010 C2 foci from Tsai, 2010 D
3Right Precentral Gyrus12824 mm341 foci from Alluri, 20121 foci from Baumgartner, 20061 foci from Brown, 20072 foci from Caria, 20111 foci from Chen, 2008: Experiment 11 foci from Chen, 2008: Experiment 2 A1 foci from Chen, 2008: Experiment 2 B1 foci from Grahn, 20071 foci from Tsai, 2010 B1 foci from Tsai, 2010 C1 foci from Tsai, 2010 D
4Left Anterior Lobe15760 mm38 foci from Alluri, 20121 foci from Brown S, 20071 foci from Caria, 20111 foci from Chen J L, 2008: Experiment 11 foci from Chen J L, 2008: Experiment 2 B1 foci from Grahn J A, 20071 foci from Tsai, 2010 C1 foci from Tsai, 2010 D

Discussion

We found evidence for consistent activation of various regions of the brain during passive music listening. As expected, our results showed activation in the primary and secondary auditory areas bilaterally. This is consistent with the existing literature showing that these areas are the critical regions of cortex for processing incoming auditory information [44]. Other activated areas included both right primary motor cortex, right and left lateral premotor cortex, and left cerebellum. We discuss in turn the implications for each of these findings below.

Activation of premotor cortex

We were unable to pinpoint activation to any further subregions of lateral PMC (i.e., dorsal or ventral), as the activation pattern could be consistent with either dorsal or ventral PMC. The average coordinates for these regions overlap in such a way that neither can be ruled out. This means that premotor involvement could be via dorsal, ventral, or both. The activation of PMC in the present analysis is consistent with both ASAP and the HAPEM framework. However, because we do not know whether this activation is localized to ventral or dorsal PMC, it is unclear if this activity reflects involvement of the dorsal stream, or potentially the action observation network that recruits vPMC for action simulation. Also, given that these clusters only represent aggregate BOLD activation, we do not have insight into the temporal dynamics of this activity, which will be crucial for inference about its origin.

Activation of primary motor cortex

M1 activity could reflect either an excitatory or inhibitory contribution, as the BOLD signal does not differentiate the two. Vigneswaran et al. [45] report that while many M1 neurons are active during action observation and thus classified as mirror neurons [46], M1 neurons directly connecting to spinal circuitry and thus contributing to observed action are suppressed during the observation of action, in order to prevent explicit action. Simulative properties of mirror neurons have also been confirmed in response to auditory sounds [47]. Thus, either excitatory, inhibitory, or both could give rise to activation of M1 during passive music listening. Some theories of mirror neuron activity [48] additionally claim that the mirror neuron network uses active inference during the perception of observed actions using predictive mechanisms. Further examination of the studies that contributed to the right premotor/primary motor activation cluster reveals that a number of them used tasks that were not passive listening in the same way as passive background listening. For example, Chen et al. [5] required in some experimental conditions for participants to anticipate later tapping to the beat in subsequent trials, which may recruit motor planning regions during the listening task. Grahn and Brett [6] asked their participants to determine whether the rhythms of two stimuli were the same or different, which may recruit motor areas to assist with the detection task. Finally, Tsai et al. [49] asked participants in some of their tasks to covertly hum along with the music that they were hearing, which may recruit motor areas for subvocalization. Therefore, more work should be done to decisively conclude whether motor areas are recruiting during background listening, in addition to passively listening for properties of the music while remaining still.

Activation of cerebellum

Many studies contributed to the cluster indicating left cerebellum. The activation of PMC and cerebellum during music listening supports predictive theories of the motor system, such that the cerebellum might provide the predictive component in a forward model of the upcoming sensory consequences. The cerebellum may be providing an inverse model for mapping sensory input to the simulated movement that would give rise to that sensation. An investigation of the temporal dynamics of communication among these regions can again provide further insight into the mechanism.

Lack of activation in SMA/pre-SMA and basal ganglia structures

While quite a few studies did report activation of SMA and pre-SMA, we did not find corresponding activation in our meta-analysis. We also failed to find evidence of basal ganglia activation. One potential reason for this discrepancy may be that another process on top of passive listening is needed to engage these regions. The generally agreed upon roles of both SMA and the basal ganglia are of sequential learning and timing [31, 50, 51]. Because these are properties of the majority of music, this region is likely to be recruited in many musical contexts. However, without directly listening for these properties of the music, it appears that automatic SMA and basal ganglia activation is not prevalent. Looking more closely at those experiments that explicitly report SMA/pre-SMA activation, they do appear to have a musical beat component to them, which relies on the underlying sequential and timing properties of the music. Bengtsson et al [4] encouraged participants to focus on rhythmic properties of the music, as did Chen, Penhume and Zatorre [5] and Grahn and Brett [6]. Baumann et al. [1] required subjects to do a counting task during passive listening as a distractor, which may have resulted in SMA activation. Experiments showing basal ganglia activation also appear to involve a beat detection task [29]. It also may be the case that activation of SMA/pre-SMA is only prevalent in musically-trained individuals, who are more likely to attend to and perceive the structural aspects of the music due to their background training. Baumann et al. [1] report increased activation of both pre-SMA and SMA in musicians compared to nonmusicians, as did Bangert [3] for SMA activation. Participants in Meister et at [2] showing SMA activation were all musicians. Thus, it appears SMA activation is likely due to either a trained musical background and/or a focus on the rhythmic properties of music. These results show that the passive perception of music engages a large and complex network of brain regions. This includes activation of areas in the motor system proper. Activation of the cerebellum and primary and premotor cortices suggests that perceived music is partly processed in areas typically considered as important for action-relevant information only. Recruitment of premotor areas during music listening supports many theories of motor involvement during perceptual tasks [11, 17, 19]. The idea of shared neural resources for tasks with underlying computational similarities has gained recent theoretical and grounded neurobiological support [16]. Most current theories suggest that perceptual processing involves the same kinds of temporal prediction involved in action, making a shared circuit useful for action-based and perception-based processing. An alternative (or potentially compatible) hypothesis is that involvement of PMC reflects the process of simulation [35], where the motor system underlies simulation of the actions required to create the observed sensory information. Our findings are consistent with both of these theoretical frameworks, though it does not provide any insight for distinguishing which theory best fits the data, as this meta-analysis only tells us which areas are active at some point in the process. This work supports the currently merging conceptualizations of action and perception [14, 48]. One limitation of the present meta-analysis was that we were unable to obtain data from contacted authors for studies that did not report all of the observed brain activations. It is possible that unpublished or unreported activations may have biased our results toward reporting motor and auditory areas, as studies that do not find activation in these areas of interest are less likely to be reported. This inability to obtain unavailable data also likely contributed to our inability to obtain enough studies for the musician/nonmusician contrast. Further exploration of differential activation in musicians relative to nonmusicians is important for advancing this work. Musicians exhibit plasticity-induced changes perceptual and motor abilities, as well as changes in structural and functional neuronal connectivity [52-55]. In particular, we believe that musicians passively listening to music should also recruit supplementary motor cortex, and might show greater activation of the cerebellum, which has a larger volume in musicians [54]. Another interesting avenue to pursue is to run more studies that directly compare different types of music listening tasks. For instance, we might compare background listening to listening in anticipation of some movement to listening for particular musical features, such as rhythm or grooviness. This will incorporate context-dependent music listening, which may reveal that different (but likely highly overlapping) networks are recruited in separate contexts. An additional limitation of this approach is that while we can identify which brain areas are active at some point during the music listening task, the BOLD signal cannot tell us anything about the temporal dynamics of the process. Complementary methods, such as EEG, should be used along with imaging data to investigate the functional connectivity among these music listening networks. This will also allow us to determine which of the existing theories fit best with the data. In summary, this study adds support to the idea that motor planning activity serves not only to help us move but is recruited for music perception even in the absence of movement. Further exploration will elucidate the functional purpose of this recruitment, as well as why and how different music listening contexts seem to engage slightly different networks. An understanding of the auditory-motor interactions underlying music perception could explain a growing number of findings suggesting an important link between music perception and the action systems of the brain.

Final dataset as a NIFTI file.

(NII) Click here for additional data file.

Completed PRISMA checklist for meta-analysis.

(DOC) Click here for additional data file.
  74 in total

Review 1.  Neural underpinnings of temporal processing: a review of focal lesion, pharmacological, and functional imaging research.

Authors:  D L Harrington; K Y Haaland
Journal:  Rev Neurosci       Date:  1999       Impact factor: 4.353

2.  Meta-analysis of the functional neuroanatomy of single-word reading: method and validation.

Authors:  Peter E Turkeltaub; Guinevere F Eden; Karen M Jones; Thomas A Zeffiro
Journal:  Neuroimage       Date:  2002-07       Impact factor: 6.556

Review 3.  Dorsal and ventral streams: a framework for understanding aspects of the functional anatomy of language.

Authors:  Gregory Hickok; David Poeppel
Journal:  Cognition       Date:  2004 May-Jun

4.  The Brain's concepts: the role of the Sensory-motor system in conceptual knowledge.

Authors:  Vittorio Gallese; George Lakoff
Journal:  Cogn Neuropsychol       Date:  2005-05       Impact factor: 2.468

5.  Metabolic and electric brain patterns during pleasant and unpleasant emotions induced by music masterpieces.

Authors:  Enrique O Flores-Gutiérrez; José-Luís Díaz; Fernando A Barrios; Rafael Favila-Humara; Miguel Angel Guevara; Yolanda del Río-Portilla; María Corsi-Cabrera
Journal:  Int J Psychophysiol       Date:  2007-03-14       Impact factor: 2.997

6.  Hippocampal-cerebellar interaction during spatio-temporal prediction.

Authors:  Yoshiyuki Onuki; Eus J W Van Someren; Chris I De Zeeuw; Ysbrand D Van der Werf
Journal:  Cereb Cortex       Date:  2013-08-22       Impact factor: 5.357

7.  Throwing while looking through prisms. I. Focal olivocerebellar lesions impair adaptation.

Authors:  T A Martin; J G Keating; H P Goodkin; A J Bastian; W T Thach
Journal:  Brain       Date:  1996-08       Impact factor: 13.501

8.  Brain activation during anticipation of sound sequences.

Authors:  Amber M Leaver; Jennifer Van Lare; Brandon Zielinski; Andrea R Halpern; Josef P Rauschecker
Journal:  J Neurosci       Date:  2009-02-25       Impact factor: 6.167

9.  Music modulation of pain perception and pain-related activity in the brain, brain stem, and spinal cord: a functional magnetic resonance imaging study.

Authors:  Christine E Dobek; Michaela E Beynon; Rachael L Bosma; Patrick W Stroman
Journal:  J Pain       Date:  2014-07-28       Impact factor: 5.820

10.  Mapping aesthetic musical emotions in the brain.

Authors:  Wiebke Trost; Thomas Ethofer; Marcel Zentner; Patrik Vuilleumier
Journal:  Cereb Cortex       Date:  2011-12-15       Impact factor: 5.357

View more
  24 in total

1.  Neural Encoding and Representation of Time for Sensorimotor Control and Learning.

Authors:  Ramesh Balasubramaniam; Saskia Haegens; Mehrdad Jazayeri; Hugo Merchant; Dagmar Sternad; Joo-Hyun Song
Journal:  J Neurosci       Date:  2020-12-30       Impact factor: 6.167

2.  Preferential activation for emotional Western classical music versus emotional environmental sounds in motor, interoceptive, and language brain areas.

Authors:  Rebecca J Lepping; Jared M Bruce; Kathleen M Gustafson; Jinxiang Hu; Laura E Martin; Cary R Savage; Ruth Ann Atchley
Journal:  Brain Cogn       Date:  2019-08-09       Impact factor: 2.310

3.  Differential motor system entrainment to auditory and visual rhythms.

Authors:  Daniel C Comstock; Ramesh Balasubramaniam
Journal:  J Neurophysiol       Date:  2022-06-29       Impact factor: 2.974

Review 4.  Time Perception for Musical Rhythms: Sensorimotor Perspectives on Entrainment, Simulation, and Prediction.

Authors:  Jessica M Ross; Ramesh Balasubramaniam
Journal:  Front Integr Neurosci       Date:  2022-07-05

Review 5.  Identifying a brain network for musical rhythm: A functional neuroimaging meta-analysis and systematic review.

Authors:  Anna V Kasdan; Andrea N Burgess; Fabrizio Pizzagalli; Alyssa Scartozzi; Alexander Chern; Sonja A Kotz; Stephen M Wilson; Reyna L Gordon
Journal:  Neurosci Biobehav Rev       Date:  2022-03-05       Impact factor: 9.052

6.  The rediscovered motor-related area 55b emerges as a core hub of music perception.

Authors:  Netanell Avisdris; Ofir Shany; Tali Siman-Tov; Carlos R Gordon; Avigail Lerner; Omer Shuster; Roni Y Granot; Talma Hendler
Journal:  Commun Biol       Date:  2022-10-18

7.  The Influence of Sound-Based Interventions on Motor Behavior After Stroke: A Systematic Review.

Authors:  Tamaya Van Criekinge; Kristiaan D'Août; Jonathon O'Brien; Eduardo Coutinho
Journal:  Front Neurol       Date:  2019-11-01       Impact factor: 4.003

8.  Neurophysiological Changes Induced by Music-Supported Therapy for Recovering Upper Extremity Function after Stroke: A Case Series.

Authors:  Shashank Ghai; Fabien Dal Maso; Tatiana Ogourtsova; Alba-Xifra Porxas; Myriam Villeneuve; Virginia Penhune; Marie-Hélène Boudrias; Sylvain Baillet; Anouk Lamontagne
Journal:  Brain Sci       Date:  2021-05-20

9.  Audiovisual Modulation in Music Perception for Musicians and Non-musicians.

Authors:  Marzieh Sorati; Dawn Marie Behne
Journal:  Front Psychol       Date:  2020-05-29

10.  Home-Based Gait Training Using Rhythmic Auditory Cues in Alzheimer's Disease: Feasibility and Outcomes.

Authors:  Joanne E Wittwer; Margaret Winbolt; Meg E Morris
Journal:  Front Med (Lausanne)       Date:  2020-01-31
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.