Literature DB >> 31646154

Listening to real-world sounds: fMRI data for analyzing connectivity networks.

Po-Chih Kuo1, Yi-Li Tseng2, Karl Zilles3, Summit Suen1, Simon B Eickhoff4,5, Juin-Der Lee6, Philip E Cheng1, Michelle Liou1.   

Abstract

There is a growing interest in functional magnetic resonance imaging (fMRI) studies on connectivity networks in the brain when subjects are under exposure to natural sensory stimulation. Because of a complicated coupling between spontaneous and evoked brain activity under real-world stimulation, there is no critical mapping between the experimental inputs and corresponding brain responses. The dataset contains auditory fMRI scans and T1-weighted anatomical scans acquired under eyes-closed and eyes-open conditions. Within each scanning condition, the subject was presented 12 different sound clips, including human voices followed by animal vocalizations. The dataset is meant to be used to assess brain dynamics and connectivity networks under natural sound stimulation; it also allows for empirical investigation of changes in fMRI responses between eyes-closed and eyes-open conditions, between animal vocalizations and human voices, as well as between the 12 different sound clips during auditory stimulation. The dataset is a supplement to the research findings in the paper "Brain dynamics and connectivity networks under natural auditory stimulation" published in NeuroImage.
© 2019 The Authors.

Entities:  

Keywords:  Auditory; Connectivity networks; Real-world; fMRI

Year:  2019        PMID: 31646154      PMCID: PMC6804394          DOI: 10.1016/j.dib.2019.104411

Source DB:  PubMed          Journal:  Data Brief        ISSN: 2352-3409


Specifications Table The dataset can be used for assessing reproducible fMRI time courses under real-life situations. The dataset can be used for investigating connectivity networks under natural sound stimulation. The dataset can be used for decoding mental states during hearing animal vocalizations and human voices.

Data

The fMRI scans were acquired when subjects were under exposure to natural auditory stimulation. The same auditory stimulation was repeated twice under eyes-closed followed by eyes-open conditions. The released dataset contains the following items: (1) T1-weighted MPRAGE anatomical images with skull stripped, (2) 4D MRI images in analyzed format (i.e., fMRI data), (3) the onset time and duration of each sound clip during the experiment, and (4) sound stimuli used in the experiment.

Experimental design, materials, and methods

Participants and experimental procedure

Forty healthy adults (19 males, average age: 22.76 ± 3.25 years) were recruited for the experiment. The subjects were instructed to listen passively to the sound stimuli, which were presented using high-quality MR-compatible insert earphones under the eyes-closed condition (4 min duration) and then under the eyes-open condition (4 min duration). The duration of the entire experiment was 8 min, as shown in Fig. 1(A). Under the eye-open condition, the subjects were instructed to keep their eyes fixed on a central cross throughout the scanning condition. Each condition (eyes-closed or eyes-open) comprised 6 clips of human voices followed by 6 clips of animal vocalizations.
Fig. 1

Stimuli and experimental paradigm: (A) The 8-min task began with an eye-closed condition (4 min duration) followed by an eye-open condition (4 min duration). The sounds were played to the subjects under each condition comprised of human voices and animal vocalizations. (B) The sound clips comprised six types of human voices, including a baby prattling, a woman crying, a man guffawing, a woman laughing, a man sneezing, and a crowd chattering. This was followed by six types of animal vocalizations, including a bird chirping, a cow mooing, a dog barking, a sheep bleating, various farm sounds, and a rooster crowing. The duration of each sound and its order of appearance were randomly determined for each subject. (C) No significant differences were observed between human and animal sounds in terms of acoustic features; for example, harmonics-to-noise ratio (HNR), intensity, and pitch.

Stimuli and experimental paradigm: (A) The 8-min task began with an eye-closed condition (4 min duration) followed by an eye-open condition (4 min duration). The sounds were played to the subjects under each condition comprised of human voices and animal vocalizations. (B) The sound clips comprised six types of human voices, including a baby prattling, a woman crying, a man guffawing, a woman laughing, a man sneezing, and a crowd chattering. This was followed by six types of animal vocalizations, including a bird chirping, a cow mooing, a dog barking, a sheep bleating, various farm sounds, and a rooster crowing. The duration of each sound and its order of appearance were randomly determined for each subject. (C) No significant differences were observed between human and animal sounds in terms of acoustic features; for example, harmonics-to-noise ratio (HNR), intensity, and pitch.

Stimuli

As shown in Fig. 1(B), the sound clips comprised human voices (baby prattling, woman crying, man guffawing, woman laughing, man sneezing, and crowd chatter) and animal vocalizations (bird chirping, cow mooing, dog barking, sheep bleating, rooster crowing, and actual sounds from a farm). The sound clips were collected from a collaborative database (www.freesound.org) and the PacDV website (www.pacdv.com). Each type of sounds was normalized to a similar intensity (70.81 ± 2.591dB) and presented using the Psychophysics Toolbox (www.psychtoolbox.org). The length of each clip and its order of appearance in the experiment were randomly determined within the [16, 26]-secs interval for each subject. Human voices were always presented before the animal vocalizations, as shown in Fig. 1(A). The 4-min sound track was first presented with the eyes-closed, and then repeated once with the eyes-open.

Acoustic features of stimuli

Fig. 1(C) presents the three acoustic features (harmonics-to-noise ratio (HNR), intensity, and pitch), which were calculated from the sound stimuli using Praat software (www.praat.org). MIRtoolbox was also used to compute other features, including zero crossing rate, energy, average frequency of events, tempo, energy above cut-off frequency, energy below cut-off frequency, centroid, spread, skewness, kurtosis, flatness of spectral distribution, spectrum entropy, and silence ratio [1]. HNR was used to measure the ratio of the strength of periodic and noisy components. It was calculated as follows:where k is the strength of the strongest periodic component. Intensity was calculated using the mean amplitude of the signals, whereas the pitch was calculated using an autocorrelation method. The other features reflect the rhythm and timbre of the sounds. Energy was estimated from the root-mean-square of the amplitude. Tempo was calculated by detecting periodicities in the event detection curve. Spectral distribution was described using statistical measures, including centroid, spread, skewness, kurtosis, flatness, and entropy. The silence ratio was used to measure the percentage of frames with energy levels below the mean energy across all frames. Note that there was no significant difference between the human voices and animal vocalizations in terms of these 16 acoustic features (with a p-value ranging from 0.12 to 0.86).

Acquisition procedure

MR scans were acquired using a 3T MAGNETOM Skyra scanner (Siemens Healthcare, Erlangen, Germany) and a standard 20-channel head-neck coil. A total of 35 slices were acquired every 2 s using a single-shot gradient EPI, which provided whole-head coverage (TR = 2000 ms; TE = 30 ms; flip angle = 84°; 35 slices with slice thickness = 3.4 mm; FOV = 192 mm; voxel size = 3 × 3 × 3.74 mm). A further 192 slices of T1-weighted anatomical images were also collected for image co-registration (TR = 2530 ms; TE = 3.30 ms; flip angle = 7°; 192 slices; FOV = 256 mm; voxel size = 1 × 1 × 1 mm).

Data preprocessing

The data can be preprocessed using SPM12 (http://www.fil.ion.ucl.ac.uk/spm). Low-resolution EPI images from each subject can be realigned in six dimensions using rigid body transformation (translation did not exceed 2 mm and rotation corrections did not exceed 2°). The T1-weighted high-resolution image volume can be co-registered to the mean of the realigned EPI image and spatially normalized using a voxel size of 2 × 2 × 3 mm to the standard space, as defined by Montreal Neurological Institute (MNI) T1-weighted template. The signal-to-noise ratio of the functional images can be enhanced via spatial smoothing using a 4-mm full-width-at-half-maximum Gaussian filter. The trends of temporal blood oxygen level-dependent (BOLD) signals caused by drifts in the magnetic field can be corrected by removing the global mean of the signals [2], [3]. In intra-subject analysis, the fMRI time series can be divided into two replicates: the eyes-closed condition and eyes-open condition. For more details about the data and analysis methods, please see the original publication in Ref. [4].

Specifications Table

SubjectNeuroscience
Specific subject areaNeuroimaging
Type of dataImage
How data were acquiredData were acquired using a 3T MAGNETOM Skyra scanner (Siemens Healthcare, Erlangen, Germany) and a standard 20-channel head-neck coil.
Data formatRaw
Parameters for data collectionEPI images: TR = 2000 ms; TE = 30 ms; flip angle = 84°; 35 slices with slice thickness = 3.4 mm; FOV = 192 mm; voxel size = 3 × 3 × 3.74mm.T1 images: TR = 2530 ms; TE = 3.30 ms; flip angle = 7°; 192 slices; FOV = 256 mm; voxel size = 1 × 1 × 1mm.
Description of data collectionData were collected from 40 subjects. The subjects were instructed to listen passively to the sound stimuli under the eyes-closed condition and then under the eyes-open condition. Each condition comprised human voices followed by animal vocalizations.
Data source locationInstitution: Institute of Statistical Science, Academia SinicaCity/Town/Region: Taipei CityCountry: TaiwanLatitude and longitude (and GPS coordinates) for collected samples/data: 24°59′12.8″N 121°34′34.5″E
Data accessibilityRepository name: MendeleyData identification number: 10.17632/9x324z2x3bDirect URL to data: https://data.mendeley.com/datasets/9x324z2x3b
Related research articleAuthor's name: Po-Chih Kuo, Yi-Li Tseng, Karl Zilles, Summit Suen, Simon B. Eickhoff, Juin-Der Lee, Philip E. Cheng, Michelle LiouTitle: Brain Dynamics and Connectivity Networks Under Natural Auditory StimulationJournal: NeuroImageDOI: https://doi.org/10.1016/j.neuroimage.2019.116042
Value of the data

The dataset can be used for assessing reproducible fMRI time courses under real-life situations.

The dataset can be used for investigating connectivity networks under natural sound stimulation.

The dataset can be used for decoding mental states during hearing animal vocalizations and human voices.

  3 in total

1.  A general statistical analysis for fMRI data.

Authors:  K J Worsley; C H Liao; J Aston; V Petre; G H Duncan; F Morales; A C Evans
Journal:  Neuroimage       Date:  2002-01       Impact factor: 6.556

2.  The global signal and observed anticorrelated resting state brain networks.

Authors:  Michael D Fox; Dongyang Zhang; Abraham Z Snyder; Marcus E Raichle
Journal:  J Neurophysiol       Date:  2009-04-01       Impact factor: 2.714

3.  Brain dynamics and connectivity networks under natural auditory stimulation.

Authors:  Po-Chih Kuo; Yi-Li Tseng; Karl Zilles; Summit Suen; Simon B Eickhoff; Juin-Der Lee; Philip E Cheng; Michelle Liou
Journal:  Neuroimage       Date:  2019-07-22       Impact factor: 6.556

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.