Literature DB >> 34675231

An ALE meta-analytic review of top-down and bottom-up processing of music in the brain.

Victor Pando-Naude1, Agata Patyczek2, Leonardo Bonetti3, Peter Vuust3.   

Abstract

A remarkable feature of the human brain is its ability to integrate information from the environment with internally generated content. The integration of top-down and bottom-up processes during complex multi-modal human activities, however, is yet to be fully understood. Music provides an excellent model for understanding this since music listening leads to the urge to move, and music making entails both playing and listening at the same time (i.e., audio-motor coupling). Here, we conducted activation likelihood estimation (ALE) meta-analyses of 130 neuroimaging studies of music perception, production and imagery, with 2660 foci, 139 experiments, and 2516 participants. We found that music perception and production rely on auditory cortices and sensorimotor cortices, while music imagery recruits distinct parietal regions. This indicates that the brain requires different structures to process similar information which is made available either by an interaction with the environment (i.e., bottom-up) or by internally generated content (i.e., top-down).
© 2021. The Author(s).

Entities:  

Mesh:

Year:  2021        PMID: 34675231      PMCID: PMC8531391          DOI: 10.1038/s41598-021-00139-3

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Music is a highly multifaceted aspect of the human experience and therefore a privileged tool to gain insights into the key features of the human brain, its structure, and supporting neural processes. One of such key features is the ability to constantly integrate information and coordinate a variety of different tasks. Indeed, a successful existence depends on the brain’s capacity to combine external stimuli with internal representations to ultimately select the best response to the ever-changing environment[1]. The processes which regulate such needs have often been framed in the notorious “top-down/bottom-up” dichotomy. This simplistic approach has been redefined by the predictive coding theory (PC) which intends to set a framework that explains the complexity of the flow of information in the brain. Within this framework, music is an ideal example of the constant integration of multi-modal top-down and bottom-up information, as it includes different processes such as perception, imagery, and production/creation[2]. Despite decades of research in the field of music cognition, there is no comprehensive study investigating brain activation associated to these three distinct modalities of music processing. Throughout the years, key examples of both top-down and bottom-up processes have been revealed. On the one hand, it has been shown that environmental stimuli activate different brain cortices, coherently with the chosen sensory modality[3] (e.g., visual inputs activate visual brain regions, while auditory inputs activate auditory cortices). According to the PC, bottom-up processing is represented as a series of loops formed by ascending and descending connections between neighbouring levels in a hierarchically organized system (feedforward-feedback). This system is reliable and fast, with a semi-hardwired architecture sufficiently stable to allow for rapid processing of stimuli, and that can be dynamically re-shaped. On the other hand, top-down mechanisms have been described in diverse manners including anatomical (connections between levels of hierarchy), cognitive (hypothesis-driven), gestaltist (modulation of bottom-up), and dynamic (entrainment of neuronal populations by oscillatory activity)[4]; all suggesting that the system allows individuals to predict future events and stimuli. Notably, the term top-down should always be challenged when used[5]. According to the PC, top-down processes can be seen as influences that allow the semi-hardwired network to be flexible and ensure efficient and reliable sensory processing. In other words, the system takes as much new information as possible (perception) to construct predictions about the environment to rapidly apply it in a behaviour (action). Taking this in consideration, it is clear that bottom-up and top-down mechanisms are not opposites, and that ascending and descending connections are involved in both processes. In this context, music is an ideal domain to study the constant integration of multi-modal mechanisms[6], as music engages audio-motor coupling and other cognitive and emotional mechanisms such as pleasure[7]. Thus, how the brain processes music may represent a privileged tool to gain insights into such central topics. Music is built upon several concurrent bottom-up processes such as perception of sounds and rhythms. Moreover, it also comprises a number of top-down mechanisms such as music production and music imagery, the latter being the abstract mental manipulation of musical sounds[8]. On the neural level, previous studies investigating such processes have provided a wide array of underlying active brain areas, spanning from auditory and sensorimotor cortices[9] to cerebellum[2], cingulum[10], basal ganglia[11], hippocampus[12], and amygdala[13]. In light of this, music may offer an ideal opportunity to investigate similarities and differences between top-down and bottom-up processes in complex multi-modal brain functioning. A large number of high-quality studies on music neuroscience already exists. However, individual studies do carry the limitation of small sample size and therefore consequent low reliability[14]. Recent advancements in statistical methods and meta-analyses allow us to overcome restrictions related to data protection regulation[15], gaining insights from a large population size and embrace “big data” human imaging. Currently, there are a number of coordinate-based algorithms for meta-analysis of neuroimaging studies such as the Activation Likelihood Estimation (ALE)[16] and kernel density analysis (KDA)[17]. The ALE method reports location probabilities associated with each foci (set of x, y, and z coordinates), whereas KDA shows the number of foci surrounding a specific voxel. In the field of music cognition, the ALE methodology has been used to investigate several independent features of music cognition including recruitment of motor areas[2], auditory-motor entrainment[18], hierarchical auditory organisation[19], and music-evoked emotions[20]. Further, projects following open science principles such as the BrainMap[21] platform and its GingerAle[14] meta-analysis algorithm enable sharing of data and so accelerate progress in human brain mapping[22]. These factors converge to make the BrainMap ALE meta-analysis a convenient and appropriate method to compare brain activation across studies of music cognition. Thus, for the first time, we have performed coordinate-based meta-analyses (CBMA) of a wide range of functional magnetic resonance imaging (fMRI) studies using state-of-the-art methods such as ALE, aiming to assess the neural mechanisms underlying music cognition. As a secondary aim, we sought to assess the broad nature of musical processing as a model of cognition itself. Specifically, appraising perception, production, and imagery in light of top-down and bottom-up processes and the recent theory of predictive coding. In this way, we explore the common and distinct patterns of brain activation during music perception, production, and imagery; and discuss how the brain activity associated with the three modalities may reflect the top-down and bottom-up nature of the brain.

Results

A total of 1707 articles were identified through database searching, and after removing 497 duplicates, 1210 articles were initially screened by title and abstract. After excluding 907 records, 303 articles were assessed for eligibility in the full-text screening stage. From these, 130 fulfilled criteria and were included in both qualitative and quantitative synthesis (ALE meta-analysis) (Supplementary Figure 1).

Characteristics of studies

The characteristics of all studies included in the final qualitative and quantitative synthesis of the meta-analysis are shown in Table 1. In music perception, a total of 105 studies and 2035 participants (122 female; age = 26.6 ± 6.7 years) was identified. Musicians were reported in 12% of the included studies, non-musicians in 69% and both in 13%. Neuroimaging method comprised of both fMRI (96%) and PET (4%). Musical features varied across studies including emotion (24%), melody (11%), harmony (9%), timbre (4%), tonality (4%), memory (3%), pitch (3%), structure (3%), rhythm (2%), tension (2%), creativity (2%), while 15% of studies did not specified the musical feature. Most of the auditory stimuli included unfamiliar stimuli (76%).
Table 1

Characteristics of the studies included in meta-analysis.

AuthorYearDomainnMethodFeatureFamiliarityMalesFemalesTotalMusiciansAgeEducation (years)Onset (years)Training (hr/week)
Mean± SDMean± SDMean± SDMean± SD
1Agustus2018Perception20fMRIMelodyY101020Y66.6
2Alluri2012Perception11fMRITimbre, key, rhythmN6511Y23.23.716.169.13.42.51.2
3Alluri2013Perception11fMRINon-specificY7411N26.45
4Alonso2016Perception22fMRIMelodyN111122N24.095.45
5Alttenmüller2014Production18fMRIEmotion + N14418N28.78.78.45.7
6Angulo-Perkins2014Perception53fMRINon-specificN292453B288
7Armony2015Imagery47fMRINon-specificN272047B26.4
8Bangert2006Imagery14fMRINon-specificN6814Y
9Barrett2016Perception12fMRIEmotion + Y4812N22.4
10Barrett2018Perception25fMRITonalityYN
11Barrett2020Perception1fMRICreativityN011Y44
12Bastepe-Gray2020Perception7fMRINon-specificY707Y22.82.16.31.5102.5
13Baumann2007Imagery7fMRINon-specificN167B25.73.200
14Bengtsson2006Imagery11fMRIRhythm and melodyY11011Y335.71.4
15Bengtsson2007Perception11fMRIStructure/creativityY11011Y32.665.71.4
16Bengtsson2009Perception17fMRIRhythm + beat + metreN14317N23.6
17Bianco2016Perception29fMRIHarmonyN121729Y24.72.917.24.87.33.08
18Bishop2013Perception12fMRIRhythm + beat + metreY6612N21.23
19Blood1999Perception10PETEmotion + N5510N
20Bodner2001Perception7fMRIRhythm + beat + metreY7N
21Bogert2016Production56fMRIEmotion − non-specificNN28.28.2152
22Brattico2016Production29fMRIEmotion + Y141529B23.93.1
23Bravo2017aPerception12fMRIHarmonyN5712N295.16
24Bravo2017bImagery + Production14fMRIHarmonyN7714N30.453.05
25Bravo2020Perception16fMRIHarmonyN8816N295.16
26Brown2004Perception10PETNon-specificN5510N33.8
27Brown2007Perception11fMRIMelodyN5611N24.612.3
28Chapin2010Perception14fMRIEmotion − non-specificN5914N
29Chen2008aPerception12fMRIRhythm + beat + metreN12N24
30Chen2008bPerception24fMRIRhythm + beat + metreN121224N23.83
31Chiang2018Perception20fMRIStructure − normalN12820Y
32Danielsen2014Perception19fMRIRhythm + beat + metreN11819N35.26.2
33Demorest2009Production16fMRIMemoryN8816N28.6
34Donnay2014Perception11fMRIHarmonyN11011Y38.811
35Engel2011Perception22fMRICreativityN2200Y12.86.87.45.8
36Escoffier2013Perception16fMRIEmotion − non-specificN9716N21.71.9
37Fedorenko2012Production12fMRIStructure − normalN6612N
38Flores-Gutiérrez2007Perception19fMRIEmotion + and − N11819N253.05
39Fujisawa2011Perception12fMRIHarmonyY12012N
40González-García2016Perception11fMRIHarmonyY01111Y
41Grahn2007Perception27fMRIRhythm + beat + metreN81927B24.5
42Grahn2009aPerception35fMRIRhythm + beat + metreN231235N29.97.2
43Grahn2009bPerception36fMRIRhythm + beat + metreN211536N29
44Grahn2013Perception24fMRIRhythm + beat + metreN111324B27
45Green2008Perception21fMRIHarmonyNN27.3
46Green2012Perception21fMRIFamiliarityN91221N27.3
47Green2018Perception22fMRIFamiliarityN17522N
48Halpern1999Perception8fMRIMelodyN358B
49Halpern2004Perception10fMRITimbreN5510Y5
50Herdener2014Perception22fMRIRhythm + beat + metreN22B5
51Herholtz2012Perception10fMRINon-specificY5510N27
52Huang2016Perception18fMRINon-specificY18018N
53Janata2002aPerception8fMRITonalityY8N
54Janata2002bPerception12fMRIMelodyN5712Y28.58.33.5
55Janata2009Production13fMRIMemoryN21113Y20
56Jeong2011Perception15fMRIEmotion + and − N51015N22.83.4
57Jungblut2012Perception30fMRIRhythm + beat + metreN171330N26.3
58Khalfa2005Perception13fMRIEmotion non-specificN8513N28
59Kleber2007Perception16fMRINon-specificY51116Y31.068.2714.067.5925.737.26
60Kleider-Offutt2019Perception28fMRINon-specificN111728N21.4
61Koelsch2003Perception11fMRIEmotion + and − N5611N24.6
62Koelsch2014Perception20fMRIEmotion + N101020N25.554.8
63Koelsch2018Imagery24fMRIEmotion + and − Y121224N23.393.3
64Kornysheva2010Perception18fMRIRhythm + beat + metreN171118N25.1
65Langheim2002Perception6fMRINon-specificY246Y27
66Leaver2009Production20fMRINon-specificN101020N24.7
67Lee2011Production + perception12fMRIMelodyN7512N20.45.7
68Lehne2013Perception25fMRITension, loudnessN221325N23.93.66.43.6
69Levitin2005Perception13fMRIStructureN6713Y
70Levitin2016Production + perception1fMRINon-specificY101Y55
71Li2019Production31fMRIRhythm + beat + metreN151631N7
72Limb2006Imagery24fMRIRhythm + beat + metreN18624B
73Limb2008Production6fMRICreativityN606B34.210.4
74Liu2018Production12fMRICreativityN12012B30.39.84.3
75Matthews2020Perception51fMRIRhythm + beat + metreN302151B
76Meister2004Imagery + Perception12fMRILearntY21012Y26.6
77Merrill2012Imagery21fMRINon-specificN14721N24.22.4
78Mizuno2007Imagery18fMRITonalityN10810N23.12.812.45.8
79Montag2011Perception33fMRIEmotion + Y62733N23.555.04
80Morrison2003Perception6fMRINon-specificN246B38.3
81Mueller2015Perception23fMRIEmotion + N101323N25.92.9
82Ohnishi2001Perception28fMRINon-specificN141428B
83Park2013Perception12fMRIEmotion + N5712Y20.332.1413.23.6
84Park2014Production24fMRIEmotion − N201424B
85Parsons2005Perception8PETLearntY358Y
86Pereira2011Imagery27fMRIFamiliarityN91827N32
87Peretz2009Imagery9fMRIFamiliarityY099N
88Petrini2011Perception16fMRIEmotion non-specificN8816N21.75
89Pfordresher2014Perception20fRMIMelodyY41620Y2413.4
90Ragert2014Perception17fMRIMelodyN9817Y26.124.216.445.92
91Reiterer2008Perception17fMRITimbreN8917N25.2
92Rogalsky2011Imagery20fMRIMelodyY91120N22.63.5
93Sammler2010Imagery12fMRIFamiliarityY6612N29
94Schmithorst2005Perception15fMRIMelodyY41115N37.815.2
95Schön2010Perception10fMRINon-specificN2810N243
96Schwenzer2011Perception16fMRIPitchN8816N
97Shany2019Perception40fMRIEmotion + N182240B25.53.66
98Sikka2015Perception40fMRIFamiliarityN04040N
99Singer2016Perception40fMRIEmotion non-specificN182240B25.53.612.34.7
100Skouras2014Production32fMRIEmotion non-specificY141832N22.932.753.625.04
101Spada2014Production20fMRIMelody + harmonyN61420Y30.297.02216.8612
102Tabei2015Perception17fMRIEmotion non-specificN71017N21.42
103Tachibana2010Imagery + Production12fMRIPitchN7512N249.15.4
104Taruffi2017Perception24fMRIEmotion + and − N121224N25.3
105Tervaniemi2000Perception30PETHarmonyN30030N26.2
106Tervaniemi2006Perception17fMRITimbreN9817N25.1
107Thaut2008Perception12fMRIRhythm + beat + metreN9312Y26.11.87
108Tillmann2003Perception15fMRITonalityY6915Y19.48.22.54
109Tillmann2006Perception20fMRITimbre/familiarityY91120Y25.14.55.3
110Toiviainen2014Perception15fMRIVariousN10515N25.75.2
111Trost2012Perception15fMRIEmotion + and − N8715N28.89.9
112Trost2014Production18fMRIHarmonyY71118N25.87.55--
113Tsai2010Perception12fMRIRhythm + beat + metreY21012N
114Tsai2012Perception15fMRIRhythm + timbreN12315Y226.43.1
115Tsai2018Perception14fMRIPitchN31114N
116Tsai2019Production16fMRIPitch/harmonyY51116N22.441.58
117Uhlig2013Perception17fMRIMelodyY9817Y26.124.216.445.92
118Villarreal2013Perception24fMRIRhythm + creativityN91524N21.91.6
119Vuust2011Perception18fMRIRhythmN14418Y29187
120Wallmark2018Perception15fMRIEmotionN7815N19.10.72
121Watanabe2008Perception18fMRIMemoryN12618N22.40.4
122Whitehead2018Perception24fMRIMelodyN131124N25.54.24.73
123Wilson2010Perception26fMRIMelodyY111526B31.712.4
124Yoo2001Perception12fMRIPitchN8412N29.95.8
125Zarate2008Perception24fMRIPitchY121224B103.39
126Zarate2010Perception9fMRIPitchY369Y233.93114.28
127Zatorre1994Perception12PET/fMRIMelodyN6612N
128Zatorre1996Perception12PETMelodyN6612N22
129Zatorre2010Perception12fMRIMelodyY5712Y2316
130Zvyagintsev2013Perception15fMRIMelodyN8715N25.15.7

Y yes, N no, B both.

Characteristics of the studies included in meta-analysis. Y yes, N no, B both. In music production, a total of 19 studies and 292 participants (122 female; age = 29.8 ± 6.1 years) was identified. Musicians were reported in 63% of the included studies, non-musicians in 16%, and both in 21%. Neuroimaging method comprised of both fMRI (95%) and PET (5%). Musical features varied across studies including harmony (11%), melody (5%), pitch (10%) and rhythm, metre and/or beat (11%). Overall, most of the studied focused on learnt abstracts (68%) compared to creative improvisation (47%). Most of the tasks included unfamiliar content (53%). In music imagery, a total of 15 studies and 198 participants (108 female; age = 25.5 ± 3 years) was identified. Musicians were reported in 40% of the included studies, non-musicians in 47%, and both in 13%. Neuroimaging method comprised of both fMRI (93%) and PET (7%). Musical features varied across studies including melody (27%), pitch (13%), timbre (7%) and non-specific (53%) of which most included unfamiliar imagery (60%).

MRI quality

MRI quality of the included studies in the meta-analysis was assessed by a set of guidelines for the standardized reporting of MRI studies[23,24]. All studies reported their MRI design, software package and image acquisition, processing and analyses. Overall, good MRI practices were performed in the included studies (Supplementary Table 1 and Supplementary Table 2). Neuroimaging data was acquired in either 1.5 T (27%), 3 T (66%), while 7% of the studies did not specified the magnetic field strength. Most of the studies used the Siemens MRI scanner (50%), others include General Electric (28%), Phillips Electronics (4%), Hitachi (1%), Brucker (4%), Magnex Eclipse (1%), CTI (1%), and few failed to report the scanner (2%). Most of the structural images were acquired using T1-weighted sequence (42%), magnetization-prepared rapid acquisition with gradient echo sequence (MPRAGE) (27%), spoiled gradient recalled acquisition in steady state (SPGR) (9%) with 1mm3-voxel size in 39% of the studies. T2- functional images were acquired mainly using echo-planar imaging (EPI) sequence (86%). Statistical analysis was conducted in either SPM (60%), AFNI (6%), FSL (7%), LIPSIA (5%) BrainVoyager (6%), MATLAB (2%), while 13% of studies did not specify the analysis software.

Primary outcomes: ALE meta-analyses of music perception, production, and imagery

Music perception

The music perception ALE meta-analysis included 1898 foci, 105 experiments and 2035 subjects. Significant peak clusters resulted in the following areas: (1) right superior temporal gyrus (x = 52, y = − 20, z = 4) extending over insula and inferior frontal gyrus; (2) left superior frontal gyrus (x = − 54, y = − 16, z = 2) extending over transverse temporal gyrus, the insula and precentral gyrus; (3) left medial frontal gyrus (x = − 2, y = − 2, z = 66) including the superior frontal gyrus and the right medial frontal gyrus; (4) right lentiform nucleus (putamen) (x = 22, y = 8, z = 6) and the caudate body; (5) left lentiform nucleus (putamen) (x = − 22, y = 4, z = 6) and the caudate body; (6) left cerebellum (lobule III) (x = − 28, y = − 64, z = − 26); (7) left insula (x = − 32, y = 18, z = 10) including the inferior frontal gyrus and (8) right precentral gyrus (x = 54, y = 0, z = 46). Such results show consistency of cortical and subcortical areas presumably organized in a hierarchical manner processing both bottom-up and top-down musical information (Fig. 1, Table 2).
Figure 1

Anatomic likelihood estimation meta-analytic results for studies of music perception, production and imagery, at cluster level inference p < 0.05 (FWE). The primary outcomes included ALE meta-analyses of music perception, music production, and music imagery, independently. ROIs: CRBL cerebellum, INS insula, MedFG medial frontal gyrus, PreCG precentral gyrus (primary motor cortex or M1), PUT putamen, SPL superior parietal lobule, STG superior temporal gyrus (primary auditory cortex), THA thalamus, L left, R right, Z peak Z-value. Figure created with Mango (http://rii.uthscsa.edu/mango//userguide.html).

Table 2

Anatomic likelihood estimation meta-analytic results for music perception, production and imagery at cluster level inference p < 0.05 (FWE).

Cluster numberVolume (mm3)MNI coordinatesALEPZLabel (Side region BA)
xyz
1. Music perception: 1898 foci, 105 experiments, 2035 subjects
124,29652− 2041E− 017E−2710.7R Superior Temporal Gyrus BA13
54− 4− 48E−023E−178.4R Superior Temporal Gyrus BA22
64− 30105E−022E−106.2R Superior Temporal Gyrus BA42
4814− 44E−022E−075.1R Insula BA13
421263E−028E−064.3R Insula BA13
422623E−021E−054.3R Inferior Frontal Gyrus BA13
4810263E−028E−053.8R Inferior Frontal Gyrus BA9
223,864− 54− 1621E−011E−2811.0L Superior Temporal Gyrus BA22
− 40− 30127E−024E−147.5L Transverse Temporal Gyrus BA41
− 48− 34126E−027E−126.8L Superior Temporal Gyrus BA41
− 5014− 24E−023E−075.0L Insula BA13
− 546123E−027E−053.8L Precentral Gyrus BA6
36136− 2− 2665E−027E−095.7L Medial Frontal Gyrus BA6
− 212545E−027E−095.7L Superior Frontal Gyrus BA6
4− 2685E−022E−085.5R Medial Frontal Gyrus BA6
216485E−022E−085.5L Medial Frontal Gyrus BA32
4332822866E−026E−116.4R Putamen
10644E−028E−074.8R Caudate Body
52304− 22465E−021E−096.0L Putamen
− 81004E−027E−074.8L Caudate Head
61920− 28− 64− 265E−021E−085.6L Cerebellum
71816− 3218104E−023E−075.0L Insula BA13
− 302403E−026E−064.4L Insula BA13
− 422643E−024E−053.9L Inferior Frontal Gyrus BA13
81728540466E−024E−126.8R Precentral Gyrus BA4
2. Music production: 499 foci, 19 experiments, 292 subjects
1351266− 24102E−027E−074.8R Superior Temporal Gyrus BA42
50− 2062E−027E−074.8R Superior Temporal Gyrus BA13
64− 2642E−027E−074.8R Superior Temporal Gyrus BA22
62− 2222E−024E−064.5R Superior Temporal Gyrus BA41
50− 1042E−022E−054.1R Insula BA13
64− 28142E−022E−054.1R Superior Temporal Gyrus BA42
68− 1682E−024E−053.9R Transverse Temporal Gyrus BA42
22920− 480422E−028E−074.8L Precentral Gyrus BA6
− 40− 40462E−025E−064.4L Inferior Parietal Lobule BA40
− 36− 46482E−027E−064.3L Inferior Parietal Lobule BA40
− 40− 26462E−021E−054.2L Inferior Parietal Lobule BA40
− 44− 22502E−025E−053.9L Postcentral Gyrus BA2
− 44− 14442E−021E−043.6L Precentral Gyrus BA4
32080− 10− 10522E−021E−064.7L Medial Frontal Gyrus BA6
0− 4582E−029E−064.3L Medial Frontal Gyrus BA6
− 44542E−022E−054.1L Medial Frontal Gyrus BA6
− 22462E−025E−053.9L Cingulate Gyrus BA24
41912− 42− 2863E−023E−095.8L Superior Temporal Gyrus BA41
51072544322E−023E−075.0R Precentral Gyrus BA6
5210242E−028E−053.8R Inferior Frontal Gyrus BA9
3. Music imagery: 263 foci, 15 experiments, 189 subjects
1325606582E−026E−095.7L Medial Frontal Gyrus BA6
− 20622E−021E−075.2L Medial Frontal Gyrus BA6
23128− 34− 58562E−021E−075.1L Superior Parietal Lobule BA7
− 32− 54422E−022E−075.1L Angular Gyrus BA39
− 40− 64522E−028E−064.3L Inferior Parietal Lobule BA7
31384− 14− 1482E−022E−075.1L Thalamus
− 18− 462E−022E−054.2L Thalamus
4944540502E−021E−054.2L Precentral Gyrus BA6
5210441E−027E−053.8L Middle Frontal Gyrus BA6

ALE anatomic likelihood estimation, BA Brodmann area, P p-value, Z peak z-value, R right, L left.

Anatomic likelihood estimation meta-analytic results for studies of music perception, production and imagery, at cluster level inference p < 0.05 (FWE). The primary outcomes included ALE meta-analyses of music perception, music production, and music imagery, independently. ROIs: CRBL cerebellum, INS insula, MedFG medial frontal gyrus, PreCG precentral gyrus (primary motor cortex or M1), PUT putamen, SPL superior parietal lobule, STG superior temporal gyrus (primary auditory cortex), THA thalamus, L left, R right, Z peak Z-value. Figure created with Mango (http://rii.uthscsa.edu/mango//userguide.html). Anatomic likelihood estimation meta-analytic results for music perception, production and imagery at cluster level inference p < 0.05 (FWE). ALE anatomic likelihood estimation, BA Brodmann area, P p-value, Z peak z-value, R right, L left.

Music production

The music production ALE meta-analysis included 499 foci, 19 experiments and 292 subjects. Significant peak clusters resulted in the following areas: (1) right superior temporal gyrus (x = 66, y = 24, z = 10) extending over Wernicke’s area, anterior and posterior transverse temporal area; (2) left precentral gyrus (x = − 48, y = 0, z = 42) extending over precentral gyrus, inferior parietal lobule and post central gyrus; (3) left medial frontal gyrus (x = − 10, y = − 10, z = 52) including cingulate gyrus; (4) left superior temporal gyrus (x = − 42, y = 28, z = 6); (5) right precentral gyrus (x = 54, y = 4, z = 32) extending over the inferior frontal gyrus. These results evidence not only the obvious involvement of the somatomotor system, but also limbic and executive cognitive functions while playing music. Interestingly, similar areas were found in music perception, further evidencing the integration of incoming information with high-order functions and motor behaviour (Fig. 1, Table 2).

Music imagery

The music imagery ALE meta-analysis included 263 foci, 15 experiments and 189 subjects. Significant peak clusters resulted in the following areas: (1) left medial frontal gyrus (x = 0, y = 6, z = 58); (2) left superior parietal lobule (x = − 34, y = − 58, z = 56) extending over angular gyrus and inferior parietal lobule; (3) left thalamus (x = − 14, y = − 14, z = 8); (5) left precentral gyrus (x = 52, y = 10, z = 4) encompassing the middle frontal gyrus. These results show that music imagination recruits motor areas involved in the generation of movements such as the premotor cortex and supplementary motor area; areas of the basal ganglia which are also involved in facilitation of movements; and parietal areas involved in perceptual-motor coordination and theory-of-mind (Fig. 1, Table 2).

Contrast analyses

A subsequent contrast ALE meta-analysis was conducted to compare the primary outcomes, by means of conjunction and subtraction. The purpose of this analysis was to identify common and distinct brain activity between perception, production and imagery modalities of music cognition (Supplementary Table 3).

Music perception vs music production

Clusters identified as likely to be active during both music perception and production (conjunction) include the right superior temporal gyrus (BA42), bilateral superior temporal gyrus (BA13), right superior temporal gyrus (BA22), left superior temporal gyrus (BA41), insula, right transverse temporal gyrus (BA42), left medial frontal gyrus (BA6) and inferior frontal gyrus (BA9). The contrast analysis (subtraction) yielded areas specific to perception (perception–production) and production (production–perception). Areas more likely to be active in music perception are the superior temporal gyrus (BA21) and superior temporal gyrus (BA22). Conversely, areas more likely to be active in music production include bilateral post central gyrus (BA2, BA40), bilateral precentral gyrus (BA4, BA6), bilateral superior, middle and medial frontal gyrus (BA6), left cingulate gyrus (BA24, BA32), left superior temporal gyrus (BA41, BA13, 22), left inferior parietal lobule (BA40), left insula (BA13), cerebellum (culmen) and the lentiform nucleus.

Music production vs music imagery

Clusters that are presumably active during both music production and music imagery (conjunction) includes the left medial frontal gyrus (BA6). The contrast analysis (subtraction) resulted in areas either specific to music production (production-imagery) or to music imagery (imagery-production). Areas more likely to be active in music production include the left superior temporal gyrus (BA13, BA41, BA22), left transverse temporal gyrus (BA41), left postcentral gyrus (BA40) and the culmen. Areas more likely to be active while imagining music include the left superior parietal lobule (BA7), angular gyrus (BA39), medial frontal gyrus (BA6) and the precentral gyrus (BA6).

Music imagery vs music perception

Clusters likely active during both music imagery and music perception (conjunction) include the left medial frontal gyrus (BA6), right precentral gyrus (BA6) and the globus pallidus. The contrast analysis (subtraction) resulted in areas either specific to music imagery (imagery–perception) or to music perception (perception–imagery). Areas more related to music imagery include left inferior parietal lobule (BA7, BA40), right middle and left medial frontal gyrus (BA6), right precentral gyrus (BA4) and the ventral lateral and ventral posterior lateral thalamus. Finally, areas more likely to be active during music perception include the bilateral superior temporal gyrus (BA22), left transverse temporal gyrus (BA41), right claustrum, left insula (BA13) and left caudate head.

Meta-analytic connectivity modelling (MACM)

MACM was performed to functionally segregate the behavioural contribution and the patterns of co-activation of each music-related region-of-interest (ROI) resulted from the primary outcomes (n = 17). The ROIs were imported into the BrainMap database separately, to identify studies reporting activation within each ROI boundary (Supplementary Table 4). The functional characterization of each ROI is detailed in Supplementary Table 5 and include the behavioural domains of action, perception, emotion, cognition and interoception.

Music perception MACM

The right superior temporal gyrus ROI (1a) showed co-activation with left superior temporal gyrus, right claustrum, left medial frontal gyrus, right precentral gyrus, left insula, and left cerebellum. Relevant behavioural domains within its boundaries include execution, language, music, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, and oddball, phonological, pitch, semantic, and tone discrimination. The left superior temporal gyrus ROI (1b) showed co-activation with right superior temporal gyrus and right insula. Relevant behavioural domains within its boundaries include execution, language, music, positive emotion, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, and oddball, phonological, pitch, semantic, and tone discrimination. The left medial frontal gyrus ROI (1c) showed co-activation with left precentral gyrus, bilateral thalamus, bilateral cerebellum, bilateral superior temporal gyrus, and left superior parietal lobule. Relevant behavioural domains within its boundaries include execution, attention, music, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, and pitch, semantic, tactile, and tone discrimination. The right putamen ROI (1d) showed co-activation with left putamen, left medial frontal gyrus, and right inferior parietal lobule. Relevant behavioural domains within its boundaries include execution, attention, language, working memory, music, reasoning, reward, and auditory perception; and experimental paradigms including encoding, finger tapping, music comprehension, music production, reward, and pitch, semantic, tactile, and tone discrimination. The left putamen ROI (1e) showed co-activation with right putamen, left medial frontal gyrus, bilateral superior temporal gyrus, and left cerebellum. Relevant behavioural domains within its boundaries include execution, attention, language, working memory, music, reward, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, and phonological, pitch, semantic, and tone discrimination. The left cerebellum ROI (1f) showed co-activation with right cerebellum, bilateral insula, left medial frontal gyrus, bilateral superior temporal gyrus, left precentral gyrus, and right thalamus. Relevant behavioural domains within its boundaries include execution, attention, language, music, and auditory perception; and experimental paradigms including emotion induction, music comprehension, music production, and pitch, semantic, tactile, and tone discrimination. The left insula ROI (1g) showed co-activation with right insula, left medial frontal gyrus, right inferior frontal gyrus, left inferior parietal lobule, left precentral gyrus, and right middle frontal gyrus. Relevant behavioural domains within its boundaries include attention, language, working memory, music, reasoning, reward, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, reward, and oddball, phonological, semantic, and tone discrimination. The right precentral gyrus ROI (1h) showed co-activation with left precentral gyrus, left medial frontal gyrus, bilateral superior temporal gyrus, bilateral putamen, and right thalamus. Relevant behavioural domains within its boundaries include execution, attention, language, music, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, reward, and pitch, semantic, and tone discrimination.

Music production MACM

The right superior temporal gyrus ROI (2a) showed co-activation with left superior temporal gyrus, left medial frontal gyrus, left insula, left thalamus, and right precentral gyrus. Relevant behavioural domains within its boundaries include execution, attention, language, music, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, reward, and oddball, phonological, pitch, semantic, and tone discrimination. The left precentral gyrus ROI (2b) showed co-activation with right precentral gyrus, left medial frontal gyrus, left inferior parietal lobule, left middle temporal gyrus, left fusiform gyrus, right superior parietal lobule, and right cerebellum. Relevant behavioural domains within its boundaries include execution, attention, language, working memory, music, reasoning, and visual motor perception; and experimental paradigms including finger tapping, music comprehension, music production, and phonological, pitch, semantic, and tactile discrimination. The left medial frontal gyrus ROI (2c) showed co-activation with left precentral gyrus, right cerebellum, left postcentral gyrus, and left claustrum. Relevant behavioural domains within its boundaries include execution, attention, and reward; and experimental paradigms including finger tapping, music comprehension, music production, reward, semantic discrimination, and tactile discrimination. The left superior temporal gyrus ROI (2d) showed co-activation with right superior temporal gyrus, and left putamen. Relevant behavioural domains within its boundaries include execution, attention, language, music, social cognition, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, and oddball, phonological, pitch, semantic, and tone discrimination. The right precentral gyrus ROI (2e) showed co-activation with left precentral gyrus, anterior cingulate cortex, left cerebellum, bilateral superior parietal lobule, bilateral inferior parietal lobule, right insula, and right thalamus. Relevant behavioural domains within its boundaries include execution, language, reasoning, music, and social cognition; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, phonological discrimination, and pitch discrimination.

Music imagery MACM

The left medial frontal gyrus ROI (3a) showed co-activation with bilateral insula, left superior parietal lobule, right inferior parietal lobule, and bilateral cerebellum. Relevant behavioural domains within its boundaries include execution, attention, language, memory, music, reasoning, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, reward, and phonological, pitch, semantic, syntactic, tactile, and tone discrimination. The superior parietal lobule ROI (3b) showed co-activation with right superior parietal lobule, left anterior cingulate cortex, right inferior frontal gyrus, and bilateral insula. Relevant behavioural domains within its boundaries include execution, attention, language, working memory, music, reasoning, reward, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, reward, and pitch, semantic, and tactile discrimination. The left thalamus ROI (3c) showed co-activation with right thalamus, anterior cingulate cortex, left precentral gyrus, left inferior frontal gyrus, right insula, and right cerebellum. Relevant behavioural domains within its boundaries include execution, attention, working memory, music, reward, auditory perception; and experimental paradigms including emotion induction, finger tapping, motor learning, music comprehension, music production, reward, and oddball, phonological, tactile, and tone discrimination. d. The right precentral gyrus ROI (3d) showed co-activation with left precentral gyrus, left medial frontal gyrus, right insula, bilateral putamen, and bilateral superior temporal gyrus. Relevant behavioural domains within its boundaries include execution, music, negative emotion, and auditory perception; and experimental paradigms including emotion induction, finger tapping, music comprehension, music production, and phonological, pitch, semantic, tactile, and tone discrimination.

Discussion

In the present study, we have conducted ALE meta-analyses of neuroimaging studies with a total of 2660 foci, 139 experiments, and 2516 subjects, to realise a comprehensive picture of the top-down and bottom-up mechanisms required by music processing. Our main results, whose robustness is guaranteed by the large numbers of studies included in the analyses, provide a complex and appetising picture of the brain activity underlying music processing. We show that music perception and music production rely on similar brain activation involving auditory cortices and sensorimotor cortices. In turn, music imagery presents a particular and unique activation of parietal and motor regions. Finally, our primary outcomes were complemented with contrast analyses and meta-analytic connectivity modelling, describing in full the complexity of the brain areas underlying music processing.

Characteristics of included studies

The publications included in this comprehensive systematic review and meta-analysis reported a clear research question, inclusion and exclusion criteria for participants, description of methods and explicit results. In line with reproducibility efforts, the included studies used state-of-the-art techniques and computational MRI tools important for the support of standardization of neuroimaging studies. However, some works lacked important demographic data such as the years of education, age of musical training onset, and current time of musical practice, which may have an effect on behavioural tasks and neuroimaging data.

Music perception

Temporal lobe areas

As expected, the main result concerning music perception involved temporal regions and especially auditory cortices. Indeed, we found convergence in various temporal lobe areas assumed to be related to auditory feature processing as well as semantic and conceptual aspects of music. Namely, primary auditory regions of Heschl's gyrus (BA41, BA42), as well as secondary auditory area, namely Wernicke’s area (BA22). Mirrored by previous studies, these areas have been seen crucial for processing of auditory information including pattern recognition[25] and syntax processing[26]. Hence, activation of both primary and secondary areas may reflect a hierarchical computation of auditory processing including lower-level extraction followed by alignment to preconstructed predictions in association areas. Our MACM results show that the superior temporal gyrus where the primary auditory cortex lies, co-activates with motor and pre-motor areas, insula, and cerebellum, supporting the idea that a large network is established while listening to music involving audio-motor coupling and limbic and paralimbic processing of emotional content.

Audio-motor coupling

We also found convergence in motor-associated areas, assumingly related to temporal processing, inherent audio-motor coupling and entrainment. These include areas within the basal ganglia (caudate, putamen, globus pallidus and thalamus), primary motor areas (precentral gyrus BA4), supplementary motor areas (premotor cortex, supplementary motor area BA6), and the cerebellum (lobule III). As music unfolds over time and does so at a specific periodicity (rhythm, beat, metre), it is not surprising to see brain activation in areas related to temporal processing in the motor domain[27]. Furthermore, previous studies have proposed significant connectivity within the fronto-temporal-cerebellar network during the perceptual analysis of music listening[28,29]. In concordance with neuroimaging results, electroencephalography (EEG) studies have found synchronised neural activity in the gamma-band within this network, in response to rhythmic expectancy[30,31]. Additionally, specific neuronal firing patterns were found in those areas with respect to time comprehension[32]. These include ramping spike activity, which refers to a consistent increase or decrease in neuronal firing rate, spectral responses and network-produced ensemble patterns[33]. In relation to the lentiform nucleus, the putamen has been proposed to be important for internal beat perception[34] and for generation of internalised groupings or chunking of action representations[35]. MACM revealed an extensive network that can represent audio-motor coupling with areas such as primary auditory and motor cortices, premotor cortex, basal ganglia, insula, and cerebellum.

Broad activation of the insular cortex

Further, our ALE analysis outlined an arguably broad activation of the insular cortex (BA13) extending under the superior frontal gyrus, sub-lobar insular and superior temporal gyrus. The insula cortex, folded deep within the lateral sulcus and covered by the frontoparietal and temporal operculum[36], has been primarily divided into two parts: the anterior and posterior insula[37]. However, more than a dozen subdivisions have been identified[38]. With rich connectivity with both cortical and subcortical structures, the insular cortex is seen as an anatomical integration hub with rich connectivity and heavy cross-modal input[39]. Within our MACM analysis, this is particularly evident with the functionally connectivity of the insula spreading across both cortical and subcortical structures, likely reflecting the breadth of its function as a multimodal integration center. As categorised by Kurth and colleagues[40] in a meta-analysis of close to 1800 experiments, the insular cortex serves a number of categorically distinct functions including sensorimotor, olfacto-gustatory, socio-emotional and cognitive functions. With respect to music perception, such rich functional diversity aligns with previous findings presenting the role of the insular cortex in auditory processing including allocation of auditory attention[41], time perception and interceptive salience[42] as well as music evoked emotion and changes in physiological arousal[43]. With these in mind, a broad activation of the insular cortex in our results appraises the complexity and multiplicity of music as a stimulus to the human ear.

Music production

Motor areas

Playing an instrument or engaging in music production requires extremely intricate motor control functions as well as the ability to internalise and portray emotional meaning. Accordingly, and unsurprisingly, with regards to our music production analysis, we found convergence in areas related to motor planning and execution. Namely, areas within the precentral gyrus, (BA4), middle frontal gyrus (BA6) and the supramarginal gyrus (BA40). Aside from their obvious motor function in relation to instrument playing, these areas are also likely to be involved in various sensory-motor coupling. Previous studies have repeatedly outlined the PMC and the planum temporale to be the crucial players in various sensorimotor transformations[44]. Activation within the somatosensory association area such as the supramarginal gyrus has been previously linked to tactile interpretation and limb location[45] crucial for use of an instrument. In relation to music, the rostral dPMC has been recruited during increasingly more complex rhythms[34] and saliency[46]. This means that the PMC could be responsible for modifying motor sequences to reflect auditory sequential patterns in a predictive and anticipatory manner.

Limbic areas

Aside from motor dexterity crucial for the technical component of instrument playing, the performance must engage the listener with loaded emotional content. Although with reduced breadth of activation compared to motor areas, our results point to a number of regions involved in the processing and conveyance of emotional material including the anterior cingulate gyrus (BA24) and insula (BA13). Previous findings of neuroanatomical studies have outlined the temporal lobe having strong connectivity with the anterior cingulate gyrus[47]. Further, for an appropriate emotional response the anterior cingulate has shown to require specific sensory information[44]. Thus, the anterior cingulate gyrus may be crucial for top-down control of emotional meaning and bottom-up processing of emotional content. In complement, numerous neuroimaging studies have outlined the activation of cingulate gyrus during musical performance[48,49]. Interestingly, an emerging body of research is suggesting a role of the cerebellum in cognition and emotion with reciprocal connections to various limbic system regions. The fastigial nucleus in particular has been shown to have extensive topographic connections with non-motor systems including the limbic system encompassing areas such as the anterior cingulate gyrus[50].

Music imagery

Frontal regions

While imagining music, our results point to a convergence in motor-associated areas within the medial frontal gyrus (supplementary motor area or SMA), precentral gyrus, putamen and thalamus. As previously mentioned, neuroimaging studies on rhythm, metre and beat have outlined the medial frontal gyrus, superior frontal gyrus along with the basal ganglia in the context of temporal feature processing and inherent audio-motor coupling[51]. In relation to imagery, activation within the SMA is particularly interesting, being repeatedly found in various imagery domains including visual imagery[52]. Furthermore, this region is also found in relation to imagination in hallucination studies[53]. Within our results, the SMA was specific to imagery suggesting its distinct role in mental manipulation and generation of internalised representations also during music.

Parietal regions

The main result from our ALE meta-analysis show music imagery to recruit parietal regions including the superior and inferior parietal lobule as well as the angular gyrus. With regards to the parietal lobules (BA7), this result is far from surprising. These areas have been previously discussed in the context of auditory attention[54] and working memory[55]. Specific to imagery, previous research on motor imagery has reported the parietal lobule as a key region in several documentations using rTMS[56] and fMRI[57]. Furthermore, activation within the angular gyrus (BA39) compliments the aforementioned parietal activation. Previous findings have suggested its role in memory retrieval[58], working memory[59], and perspective taking[60]. Further, our MACAM analysis specific to the parietal areas have revealed a distinct connection pattern with a sequence of areas including frontoparietal and motor regions. Together, their joint role could be interpreted as key in initiating and formulating the experience of imagery involving appropriate cognitive abilities. Importantly, what must be noted is that despite several authors describing imagery to recruit secondary auditory areas, we found no temporal area activation as it did not survive the correction for multiple comparisons. However, such temporal areas were only found in the uncorrected analysis together with other limbic and paralimbic areas. This may be due to variability of designs in the imagery studies, individual differences in imagery vivacity or a low number of studies.

Comparison between music perception, production, and imagery

A direct comparison of music perception, production and imagery for each of the respective pairs revealed several indications of music processing specificity likely resulting from a bidirectional organisation of the brain. Music imagery showed a unique activation of left superior and inferior parietal lobule (BA7). Previously, studies outlined these areas to be key for top-down or goal directed attentional orienting[61], and specific to imagery, the superior parietal lobule has been named a potential candidate for top-down generation and maintenance of domain-unspecific mental images[62]. Crucially, both music perception and production, when contrasted with music imagery, showed similar broad activation of temporal and frontal areas associated with auditory processing and motor planning and execution. This is particularly interesting as to show that both perception and production must involve both top-down and bottom-up processes, thus, supporting the embodied and predictive approach which appreciates the human motor system and its actions as a reciprocal part of perception and cognition itself[63].

Limitations and future perspectives

Although our meta-analyses comprise a large number of studies and thus arguably provide reliable results, some limitations should be mentioned. First, by definition meta-analyses rely on the combination of the data from previous studies and therefore are highly dependent on the quality of such studies. In our case, we utilised studies that possessed high-quality standards, making feasible data extraction and conduction of statistical analysis. However, we should highlight that a thorough description of the methods and an even stronger focus on the description of the experimental settings would be highly beneficial for the field and thus we encourage future studies to further increase their already high-level quality. Second, we relied on the assumptions of the GingerALE method that was used in our study. Therefore, the accuracy of our findings necessarily relies on the statistical estimation of coordinate-based anatomic foci (input) which were treated as spatial probability distributions centred at the given coordinates. As conceivable, we could not control the heterogeneity of the methods used in the studies included in the meta-analysis (ranging from pre-processing software, statistical thresholds, smoothing, and participants’ characteristics) and we acknowledge that this fact could represent a potential confounder for our analyses. As in traditional meta-analyses, we have assessed the publication bias[64] and revealed that our results are robust as the clusters obtained from the primary outcomes presented FSN values within lower and upper boundaries. Importantly, this outcome proves a robust convergence of the foci emerged from our analyses even when the total amount of studies was small, particularly in the music production and imagery CBMA’s (details in Supplementary Table 6). With regard to future perspectives, we believe that the neuroscientific community would benefit from designing new ad-hoc studies to directly investigate the relationship between brain activity and different features of music processing such as music perception, production and imagination. Although our results provide a robust picture of the brain networks recruited by those different processes, carefully designed experiments which systematically modulate and compare stimuli and conditions would further refine the knowledge on this broad and complex topic.

Conclusions

In this comprehensive review, we have conducted a series of neuroimaging meta-analyses on decades of studies investigating the brain activity underlying music processing. We have shed new light on how the brain processes and integrates top-down and bottom-up mechanisms during complex multi-modal human activities such as music. The outcomes of our synthesis highlight three main concurrent music procedures, known as music perception (both top-down and bottom-up), music production (both top-down and bottom-up) and imagery (top-down). Our results show that the brain relies on different structures and mechanisms to process similar musical information. Indeed, music perception and music production depend on auditory cortices, sensorimotor cortices and cerebellum. Differently, music imagery shows a key recruitment of parietal regions. Taken together, our findings provide robust evidence that the brain requires different structures to process similar information which is made available either by the interaction with the environment (i.e., bottom-up) or by internally generated content (i.e., top-down).

Methods and materials

Search strategy, screening, and extraction

This comprehensive systematic review and meta-analysis followed procedures from the Cochrane Handbook for Systematic Reviews[65], and from the Center for Reviews and Dissemination (https://www.york.ac.uk/crd/). The review protocol was pre-registered in PROSPERO (CRD42019140177), and was carried out in accordance with the PRISMA statement[66]. A systematic literature search was conducted in PubMed, Scopus and PsycInfo for articles published up to February 26th, 2021. The search strategy for music perception, production and imagery was developed using keywords and MeSH terms (details in Supplementary Information p. 2). To ensure concurrence between the research question and data extracted, the following inclusion criteria was established: (1) healthy adults (18–80 age range); (2) studies investigating music listening, music playing, or music imagery, with no restrictions on music genre, parameter, or instrument; (3) using fMRI or PET; (4) whole-brain analyses; and (5) results reported in stereotactic coordinates either Talairach or Montreal Neurological Institute (MNI) three-dimensional-coordinate system. Studies were excluded using the following criteria: (1) review articles with no original experimental data, (2) populations with any form of disorder or disability; (3) paediatric population; (4) auditory stimuli not related to music; (5) functional connectivity analyses; (6) region-of-interest (ROI) analyses. Two reviewers (AP and VPN) independently screened by title and abstract and selected articles for full-text review and data extraction. Screening and data extraction were performed using the Covidence tool[67]. Any disagreements that arose between the reviewers were resolved through discussion. The following variables were extracted from each study: first author, year of publication, number of participants, age, sex, musical feature, year of education, year of musical training, age of onset of musical training, training hours per week, and MRI acquisition, processing, and analysis parameters. The main outcome to extract was coordinates resulting from functional brain activity related to music perception, production and/or imagination. If any of these points were not reported in the original article, authors were contacted to retrieve this information. Nine authors were contacted, with 3 positive answers.

Quality assessment of MRI studies

A set of guidelines used for the standardization of reporting MRI results was utilized to assess the quality of the included studies[23,24]. Such guidelines dictate a more consistent and coherent policy for the reporting of MRI methods to ensure that methods can be understood and replicated.

Activation likelihood estimation (ALE) and meta-analytic connectivity modelling (MACM)

All meta-analyses were performed using the activation likelihood estimation (ALE) method, implemented in GingerALE software v3.0.2[68], from the BrainMap[68] platform. Reported significant coordinates (foci) were extracted from each study. If necessary, coordinates were converted from Talairach coordinates to MNI space using the Lancaster transform (icbm2tal) incorporated in GingerALE[14,16]. The ALE method extracts the coordinates from the included studies and tests for anatomical consistency and concordance between the studies. The coordinates are weighted according to the size of the sample (number of subjects), and these weightings contribute to form estimates of anatomic likelihood estimation for each intracerebral voxel on a standardized map. This method treats anatomic foci not as single points, but as spatial probability distributions centred at the given coordinates. Thus, the algorithm tests the correlation between the spatial locations of the foci across MRI studies investigating the same construct and assesses them against a null-distribution of random spatial association between experiments[34]. Statistical significance of the ALE scores was determined by a permutation test using cluster-level inference at p < 0.05 (FWE), with a cluster-forming threshold set at p < 0.001. The primary outcome was brain functional activity related to music cognition with the aim of comprehensively examine the brain regions associated with music perception, music production, and music imagery, independently. Then, the resulting ALE maps from each group were tested for similarity (conjunction) and difference (subtraction) in a contrast analysis with the purpose of identifying common and distinct brain areas recruited while listening, playing and/or imagining music. The GingerALE software was used following procedures described in the manual[69]. In short, during the first step, the foci from the two groups of interest were merged into a single text file and an ALE meta-analysis was conducted on the new file. Then, three ALE maps are imported to the GingerALE algorithm which randomly divides the pooled foci (foci of two datasets to be compared) into two new datasets of the same size as the original dataset. Then, voxel-wise ALE scores are calculated by subtracting one ALE image from the other. This procedure is repeated 10,000 times to yield a higher number of permutations compared to that of original dataset[70]. Finally, the ‘true’ difference of ALE scores is established and compared to the differences under a null distribution. This yields a voxel-wise p value of the difference which was calculated using cluster-level inference at p < 0.05 (FWE), with a cluster-forming threshold set at p < 0.001. Meta-analytic connectivity modelling (MACM) was conducted to identify patterns of co-activation from regions-of-interest (ROI) resulting from the primary outcomes, aiming to functionally segregate each region’s putative contribution to behavioural domains and paradigm classes[71,72]. Such analyses were performed using Sleuth[73] to identify studies reporting activation within each music-related ROI boundary independently and included the experiment level search criteria of “context: normal mapping” and “activations: activation only”. Then, ALE meta-analyses were performed in GingerALE over all foci resulted after the search in Sleuth to identify regions of significant convergence. Statistical significance of the ALE scores was determined by a permutation test using cluster-level inference at p < 0.05 (FWE), with a cluster-forming threshold set at p < 0.001. Functional characterization of music-related clusters was based on the “Behavioral Domain” (BD) meta-data categories in the BrianMap database which include action, perception, emotion, cognition and interoception. All meta-analytic results (ALE maps) were visualized using Mango[21] on the MNI152 1 mm standard brain, and resulting coordinates were cross-referenced to the Harvard–Oxford Cortical and Subcortical Atlas and the Juelich Histological Atlas via NeuroVault[74] and FSLeyes[75], respectively.

Fail-Safe N analysis (FSN)

Coordinate-based meta-analyses such as ALE can be subject to different forms of publication bias which may impact results and invalidate findings (e.g., the “file drawer problem”). The Fail-Safe N analysis (FSN)[64] assesses the robustness of results against potential publication bias. This method refers to the amount of contra-evidence that can be added to a meta-analysis before the results change and can be obtained for each cluster that survives thresholding in an ALE meta-analysis. It is estimated that a 95% confidence interval for the number of studies that report no local maxima varies from 5 to 30 per 100 published studies of normal human brain mapping. Using the upper bound and the fact that the CBMA’s consist of 105 music perception studies, 19 music production studies, and 15 music imagery studies, an estimate for the number of unpublished experiments is 32, 6, and 5, respectively. Therefore, the minimum FSN was defined as 32 for music perception, 6 for music production, and 5 for music imagery. A higher FSN indicates more stable results and hence a higher robustness. Supplementary Information.
  70 in total

Review 1.  Surgical anatomy of the insula.

Authors:  M Guenot; J Isnard; M Sindou
Journal:  Adv Tech Stand Neurosurg       Date:  2004

Review 2.  Towards a neural basis of music-evoked emotions.

Authors:  Stefan Koelsch
Journal:  Trends Cogn Sci       Date:  2010-02-10       Impact factor: 20.229

3.  Neural mechanisms of rhythm perception: current findings and future perspectives.

Authors:  Jessica A Grahn
Journal:  Top Cogn Sci       Date:  2012-07-18

4.  Cytoarchitecture and cortical connections of the anterior cingulate and adjacent somatomotor fields in the rhesus monkey.

Authors:  R J Morecraft; K S Stilwell-Morecraft; P B Cipolloni; J Ge; D W McNeal; D N Pandya
Journal:  Brain Res Bull       Date:  2012-01-02       Impact factor: 4.077

5.  Interacting cortical and basal ganglia networks underlying finding and tapping to the musical beat.

Authors:  Shu-Jen Kung; Joyce L Chen; Robert J Zatorre; Virginia B Penhune
Journal:  J Cogn Neurosci       Date:  2012-11-19       Impact factor: 3.225

Review 6.  Evaluating the consistency and specificity of neuroimaging data using meta-analysis.

Authors:  Tor D Wager; Martin A Lindquist; Thomas E Nichols; Hedy Kober; Jared X Van Snellenberg
Journal:  Neuroimage       Date:  2008-11-20       Impact factor: 6.556

Review 7.  FSL.

Authors:  Mark Jenkinson; Christian F Beckmann; Timothy E J Behrens; Mark W Woolrich; Stephen M Smith
Journal:  Neuroimage       Date:  2011-09-16       Impact factor: 6.556

8.  Automated regional behavioral analysis for human brain images.

Authors:  Jack L Lancaster; Angela R Laird; Simon B Eickhoff; Michael J Martinez; P Mickle Fox; Peter T Fox
Journal:  Front Neuroinform       Date:  2012-08-28       Impact factor: 4.081

9.  Stimulus complexity and categorical effects in human auditory cortex: an activation likelihood estimation meta-analysis.

Authors:  Fabienne Samson; Thomas A Zeffiro; Alain Toussaint; Pascal Belin
Journal:  Front Psychol       Date:  2011-01-17

Review 10.  The angular gyrus: multiple functions and multiple subdivisions.

Authors:  Mohamed L Seghier
Journal:  Neuroscientist       Date:  2012-04-30       Impact factor: 7.519

View more
  4 in total

1.  The rediscovered motor-related area 55b emerges as a core hub of music perception.

Authors:  Netanell Avisdris; Ofir Shany; Tali Siman-Tov; Carlos R Gordon; Avigail Lerner; Omer Shuster; Roni Y Granot; Talma Hendler
Journal:  Commun Biol       Date:  2022-10-18

Review 2.  An ALE meta-analytic review of musical expertise.

Authors:  Antonio Criscuolo; Victor Pando-Naude; Leonardo Bonetti; Peter Vuust; Elvira Brattico
Journal:  Sci Rep       Date:  2022-07-11       Impact factor: 4.996

3.  The sweet spot between predictability and surprise: musical groove in brain, body, and social interactions.

Authors:  Jan Stupacher; Tomas Edward Matthews; Victor Pando-Naude; Olivia Foster Vander Elst; Peter Vuust
Journal:  Front Psychol       Date:  2022-08-09

Review 4.  Neural Correlates of Music Listening: Does the Music Matter?

Authors:  Mark Reybrouck; Peter Vuust; Elvira Brattico
Journal:  Brain Sci       Date:  2021-11-24
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.