Literature DB >> 29492004

Humans identify negative (but not positive) arousal in silver fox vocalizations: implications for the adaptive value of interspecific eavesdropping.

Piera Filippi1,2,3, Svetlana S Gogoleva4, Elena V Volodina5, Ilya A Volodin4,5, Bart de Boer1.   

Abstract

The ability to identify emotional arousal in heterospecific vocalizations may facilitate behaviors that increase survival opportunities. Crucially, this ability may orient inter-species interactions, particularly between humans and other species. Research shows that humans identify emotional arousal in vocalizations across multiple species, such as cats, dogs, and piglets. However, no previous study has addressed humans' ability to identify emotional arousal in silver foxes. Here, we adopted low- and high-arousal calls emitted by three strains of silver fox-Tame, Aggressive, and Unselected-in response to human approach. Tame and Aggressive foxes are genetically selected for friendly and attacking behaviors toward humans, respectively. Unselected foxes show aggressive and fearful behaviors toward humans. These three strains show similar levels of emotional arousal, but different levels of emotional valence in relation to humans. This emotional information is reflected in the acoustic features of the calls. Our data suggest that humans can identify high-arousal calls of Aggressive and Unselected foxes, but not of Tame foxes. Further analyses revealed that, although within each strain different acoustic parameters affect human accuracy in identifying high-arousal calls, spectral center of gravity, harmonic-to-noise ratio, and F0 best predict humans' ability to discriminate high-arousal calls across all strains. Furthermore, we identified in spectral center of gravity and F0 the best predictors for humans' absolute ratings of arousal in each call. Implications for research on the adaptive value of inter-specific eavesdropping are discussed.

Entities:  

Keywords:  eavesdropping; emotional arousal; emotional valence; inter-species communication; silver foxes; vocal communication

Year:  2017        PMID: 29492004      PMCID: PMC5804197          DOI: 10.1093/cz/zox035

Source DB:  PubMed          Journal:  Curr Zool        ISSN: 1674-5507            Impact factor:   2.624


Emotions are linked to internal brain and physiological states that may be caused by external stimuli (Mendl et al. 2010; Anderson and Adolphs 2014). At least two dimensions define emotional states: arousal, namely a state of the brain or the body reflecting responsiveness to sensory stimulation ranging from sleep (low arousal) to frenetic excitement (high arousal) (Russell 1980), and valence, the intrinsic attractiveness (positive valence) or averseness (negative valence) of an event (Russell 1980; Frijda 1987; Mendl et al. 2010). Some emotions can have a similar arousal level and differ in valence (e.g., happiness vs. anger) or vice versa, have similar valence and differ in arousal (e.g., annoyance vs. rage). Valence and arousal are described as “building blocks” of emotions (Anderson and Adolphs 2014). Changes in emotional states may be reflected in vocalizations and express the emotional state of the signaler (Darwin 1872; Gogoleva et al. 2010a, 2010c; Briefer 2012; Volodin et al. 2017). Indeed, a key aspect of the acoustic features of the voice is that they reflect changes in the configuration and action of muscles involved in vocal production (e.g., the diaphragm and laryngeal muscles). Therefore, the way air flows through the system and ultimately the quality of the sounds produced is critically affected (Davis et al. 1996). In an extensive review on emotional vocal communication in mammals, Briefer (2012) suggested that increases in frequency-related parameters of the voice (e.g., fundamental frequency, frequency range, and spectral shape) and in amplitude contour, increase in call rate and decrease in inter-call interval predict high arousal in a number of mammalian species. In addition, Morton (1977) observed that both mammals and birds use of harsh, low-frequency vocalizations in hostile agonistic contexts, and of more tone-like, high-frequency sounds in fearful or appeasing contexts. As to research on emotional valence expression, studies comparing acoustic features underlying vocalizations produced in positive and negative situations (controlling for arousal) are sparse since animals are more likely to call in negative contexts. By reviewing findings in a number of studies conducted in a number of mammalian species, Briefer (2012) suggested that the only acoustic parameter that consistently changes as a function of valence is duration. Specifically, animal vocalizations reflecting positive valence are shorter than vocalizations expressing negative valence. Critically, from the listeners’ side, correct identification of the emotional state of a signaler through accurate perception of acoustic modulation of the voice may drive adaptive survival (Nesse 1990; Anderson and Adolphs 2014) in the context of territory disputes, predators avoidance (Nesse 1990; Owings and Morton 1998; Desrochers et al. 2002; Cross and Rogers 2005; Kitchen et al. 2010), or social interactions (Gogoleva et al. 2010a, 2010c; Altenmüller et al. 2013; Bryant 2013). Furthermore, the ability to recognize emotional arousal in vocal expressions may be decisive for the survival of newborns, who require caregivers to perceive and react to their needs (Marmoset monkey Callithrix jacchus:Tchernichovski and Oller 2016; Zhang and Ghazanfar 2016; human Homo Sapiens: Fernald 1992). Survival may be facilitated by the ability to identify emotions not only in vocalizations emitted by conspecifics, but also by members of other species (Nesse 1990). This ability may provide information that is key to responding appropriately. Indeed, it has been shown that nonhuman animals’ “eavesdropping” on another species alarm calls increases opportunities for survival (Owings and Morton 1998; Kitchen et al. 2010; Fallow et al. 2011; de Boer et al. 2015; Magrath et al. 2015). Advantageous responses to inter-specific calls may occur as a result of acoustic similarity in the signals (Aubin 1991; De Kort and Carel ten 2001; Johnson et al. 2003). In other cases, listeners respond appropriately to calls that are acoustically different from their own (Templeton et al. 2005; Lea et al. 2008; Fallow and Magrath 2010), suggesting that responses are biologically rooted, or, in the case of species living in close territories, learned. For example, juvenile vervet monkeysCercopithecus aethiops pygerthyrus appropriate responses to playback of alarm calls given by superb starlings Spreo superbus vary depending on the rates of exposure to these alarm calls (Hauser 1988). Generally, the ability to respond appropriately to heterospecific calls, which may presuppose the ability to recognize their level of emotional arousal and valence (Mendl et al. 2010), is the result of a signaling system that affords inter-specific beneficial outcomes in dangerous contexts. In parallel to research on the acoustic correlates of emotional dimensions in vocal production, multiple studies examined perception of valence and arousal in mammals’ vocalizations. Research on arousal perception suggests that humans rate human, piglet, and dog vocalizations with higher fundamental frequency (F0) as expressing higher emotional arousal (Laukka et al. 2005; Faragó et al. 2014; Maruščáková et al. 2015). Moreover, McComb et al. (2009) suggest that humans perceive as more urgent and less pleasant cat purrs recorded while cats were actively seeking food than purrs recorded in non-solicitation contexts. The authors identified in voiced frequency peaks the acoustic predictors of humans’ accuracy in classifying cat vocalizations. Sauter et al. (2010) found that humans perceive human nonverbal vocalizations with higher F0 means, shorter duration, more amplitude onsets, lower minimum F0, and less F0 variation as expressing higher arousal. In a recent study, Filippi et al. (2016) suggest that humans are able to discriminate high versus low levels of arousal in negative-valenced vocalizations of terrestrial tetrapods spanning all classes of animals. In addition, they identified in F0 and spectral center of gravity the acoustic predictors of this ability, pointing to biologically rooted acoustic universals of arousal perception. Furthermore, studies on arousal perception across species suggests that shepherds’ high-pitched, quickly pulsating whistles have an activating effect on dogs (McConnell and Baylis, 2010) and that 2 species of deers Odocoileus hemionus and Odocoileus virginianus respond to infant distress vocalizations of human and nonhuman animals (infant marmots Marmota flaviventris, seals Neophoca cinerea, and Arctocephalus tropicalis, domestic cats Felis catus, bats Lasionycteris noctivagans, humans H.sapiens, and other ungulates: eland Taurotragus oryx, red deer Cervus elaphus, fallow deer Dama dama, sika deer Cers nippon, pronghorn Antilocapridae americana, and bighorn sheep Ovis canadensis) if the F0 falls within the deer’s frequency response range (Lingle and Riede 2014). Moreover, research on valence perception suggests that humans recognize the negative and positive contexts in which vocalizations of human infants H.sapiens, chimpanzees Pan troglodytes (Scheumann et al. 2014), domestic pigs Sus scrofa domesticus (Tallet et al. 2010; Maruščáková et al. 2015), dogs Canis familiaris (Pongrácz et al. 2006; Scheumann et al. 2014), and cats F.catus (Nicastro and Owren 2003) were recorded (but see Belin et al. 2008 for contrasting results on human perception of valence in cats F.catus—and monkeys Macaca mulatta vocalizations). Albuquerque et al. (2016) found that dogs can identify emotional valence in both conspecific and human vocalizations. Further studies addressing the acoustic predictors of valence perception suggest that humans’ rate human and dog vocalizations with shorter duration, and human vocalizations with lower SCOG as more positive (Faragó et al. 2014; but see Pongrácz et al. 2005). Maruščáková et al. (2015) found that humans rate domestic piglets’ vocalizations with increased F0 and duration as more negative. Notably, much research has examined humans’ perception of arousal or valence in vocalizations of multiple species, adopting a continuous rating scale, and their ability to infer whether the vocalizations were produced in a positive or negative context. However, humans’ ability to recognize different arousal levels in nonverbal vocalizations from the same species systematically varying in valence content remains largely unexplored. Furthermore, to the best of our knowledge, no previous work has investigated this issue focusing on the perception of heterospecific calls, linking the perception of voice modulation to the arousal state of the caller—as identified based on independent nonvocal indicators. Here, for the first time, we analyzed humans’ ability to recognize emotional arousal in silver fox vocalizations produced in a positive or negative context. Furthermore, we examined the acoustic predictors of both this ability and of the level of perceived arousal in silver fox calls, as reported using a rating scale spanning from 1 (very subdued) to 7 (very excited). To this aim, we adopted vocalizations produced by the following three strains of silver fox: Tame, Aggressive, and Unselected. These strains are the result of a genetic selection program in silver foxes. Belyaev (1979) hypothesized that selection of farm foxes for less fearful and less aggressive behavior would yield a strain of domesticated fox. To address this hypothesis, the authors started a program designed to recapitulate canine domestication in the silver fox at the Institute of Cytology and Genetics of the Russian Academy of Sciences, Novosibirsk, Russia (Trut 1999; Trut et al. 2004, 2009a). Thus, different behavioral phenotypes across these three strains of silver fox have been experimentally established by intense selective breeding (Kukekova et al. 2012). Specifically, Tame silver foxes have been experimentally selected in the course of more than 50 years of selection for positive behavior to people and show friendly response to humans, approaching any unfamiliar experimenter (Belyaev 1979; Trut 1999, 2001; Trut et al. 2009) and even kept as pets (Ratliff 2011). Aggressive foxes were selected for aggressive behaviour and can attack humans (Trut 1980, 2001; Kukekova et al. 2008a, 2008b). Unselected foxes were not deliberately selected for behavior and demonstrate aggressively fearful behavior to humans (Pedersen and Jeppesen 1990; Pedersen 1991, 1993, 1994; Trut 1999; Nimon and Broom 2001; Kukekova et al. 2008a, 2008b; Gogoleva et al. 2010c). In the presence of an unfamiliar human, the Unselected fox with its wild type attitudes toward people enlarges the animal–human distance and shows escape responses (Supplementary movies 1–3). Cross fostering, cross breeding, and embryo transplantation experiments have shown that behavioral differences between Tame and Aggressive foxes are genetically determined (Trut 1980, 2001; Kukekova et al. 2012). Critically, strict selection for tame behavior included a substantial decrease in the levels of adrenocorticotropic hormone (ACTH) and basal levels of corticosteroids in blood plasma of Tame foxes (Oskina 1996; Trut 1999; Trut et al. 2004, 2009; Oskina et al. 2008). Both basal and ‘post-stress’ (after 10-min restraint in human hands) levels of cortisol and ACTH in Aggressive foxes did not differ from those of Unselected foxes, whereas in Tame foxes they were much lower (Oskina et al. 2008). These findings suggest that, in contrast to tame silver foxes, Unselected and Aggressive foxes experience negative emotional arousal in response to human approach. For this reason, Belyaev's silver foxes (Belyaev 1979) provide a unique model for studying the human ability to distinguish between high- and low-arousal levels in heterospecific vocalizations of both negative and positive emotional valence. Converging empirical evidence on physiological and vocal correlates of arousal (Oskina 1996; Trut et al. 2009; Gogoleva 2010a, 2010c) indicate that, in response to humans approaching their cage, Tame foxes experience heightened emotional arousal with positive valence, while Aggressive and Unselected silver foxes experience high arousal with negative valence. Gogoleva et al. (2010a, 2010c) found that, across strains of silver fox, higher levels of emotional arousal due to human approach are reflected in an increased calling rate, and the proportion of time spent vocalizing. This applies to both high-arousal calls with positive valence—in the case of tame foxes, who experience comfort in relation to human physical approach—and high-arousal calls with negative valence, in the case of Aggressive and Unselected silver foxes, who experience discomfort in relation to human physical approach. Thus, vocal responses toward humans differ according to fox strain (Gogoleva et al. 2008, 2009, 2010b, 2010c, 2011), as a consequence of genetic differences between these 3 strains (Trut 1980, 2001; Kukekova et al. 2012). Call types that silver foxes in captivity produce toward humans, include whine, moo, cackle, growl, bark, pant, snort, and cough (Gogoleva et al. 2008, 2010a, 2010b, 2010c). Selection by behavior did not affect the vocal repertoire of silver fox; all strains (Tame, Aggressive, and Unselected) retain all call types toward conspecifics (Gogoleva et al. 2010b). However, toward humans, Tame foxes selectively produce cackles and pants but never cough or snort, while Aggressive and Unselected foxes selectively produce coughs and snorts but never cackles or pants (Gogoleva et al. 2008, 2009, 2010c, 2013). Importantly, of the total of 8 call types that silver foxes direct to humans, only the whine often occurs in all 3 strains (Gogoleva et al. 2008, 2010a, 2010c, 2013). In addition, Newton-Fisher et al. (1993) have reported on the whine call-type in wild red fox, suggesting that it is used in both agonistic and affiliative contexts. Therefore, whines constitute the most appropriate call type for investigating human perception of arousal across the 3 fox strains. The analysis of human perception of vocalizations of these 3 strains of silver foxes provides an ideal context for research on the adaptive value of emotion perception across species area for the following reasons: 1) calls emitted in comparable behavioral contexts across all 3 strains can be used; 2) our stimuli are all instances of one call type, the whine, which is produced in all 3 strains. This excludes variation in call types as a confound in the analyses; 3) both the arousal and valence states of silver foxes in response to human approach have been attested in terms of physiological measurements, namely hormonal responses, in previous studies (Oskina 1996; Trut 1999; Trut et al. 2004, 2009; Oskina et al. 2008). These 3 conditions enabled us to disentangle human sensitivity to high-arousal calls with negative valence from human sensitivity to high-arousal calls with positive valence. Although research shows that humans identify emotional arousal in vocalizations across multiple species, such as cats, dogs, and piglets, humans’ ability to identify emotional arousal in silver foxes has never been investigated. Here, we adopted low- and high-arousal calls emitted by 3 strains of silver fox—Tame, Aggressive, and Unselected—in response to human approach. Specifically, within this research framework, the present study aimed to address the following questions: 1) Are humans able to identify high-arousal calls in silver fox vocalizations? 2) If so, does this ability vary as a function of strain? 3) What are the acoustic features that predict both the ability to recognize higher levels of arousal in silver fox vocalizations and the level of perceived arousal in silver fox vocalizations? In line with previous research on arousal perception across animal species (Pongrácz et al. 2006; Lingle et al. 2012; Teichroeb et al. 2013; Lingle and Riede 2014; Scheumann et al. 2014; Filippi 2016; Filippi et al. 2016), we hypothesized that humans would be able to discriminate between low and high levels of arousal expressed in all 3 strains of silver fox using frequency-related parameters, which are identified as acoustic correlates of arousal in mammals and birds (Morton 1977; Briefer 2012). This investigation might provide key insights into the adaptive effects of the ability to identify different levels of arousal varying in valence in a nonhuman species.

Materials and Methods

Acoustic recordings and emotion classification

The stimuli adopted in our study are selected from acoustic recordings collected at the experimental farm of the Institute of Cytology and Genetics, Novosibirsk, Russia under a framework used in previous studies (Gogoleva et al. 2010a, 2010c). Three study groups included Tame (selected for tameness toward humans, 45–47 generations since the start of selection), Aggressive (selected for aggressiveness toward humans; 34–36 generations since the start of selection), and Unselected (unselected for any behavioral trait) adult female silver foxes. The foxes were kept and tested in individual outdoor cages (for keeping details see Gogoleva et al. 2010a, 2010c). Human-approach tests were made when foxes were in their home cages, out of breeding or pup-raising seasons. The same researcher (S. S. G.), unfamiliar to the foxes, performed all human-approach tests (1 per fox), while acoustic recordings were collected. Each test lasted 10 min and included 5 successive steps, each lasting 2 min. A test started at the moment of the researcher's approach to a focal fox cage at a distance of 50 cm. At Step 1, the researcher was motionless; at Step 2, the researcher performed smooth body and hand movements left to right, maintaining a distance of 50 cm; at Step 3, the researcher shortened the human–fox distance with 1-step forward, and performed body and hand movements forward and back, touching the cage door with her fingers; at Step 4, the researcher enhanced the human–fox distance with 1-step back; and repeated Step 2. Finally, at Step 5, the research was motionless as in Step 1. Thus, the human impact on an animal increased between the Steps 1 and 3, and decreased between Steps 3 and 5. The shifts in the levels of emotional arousal and valence were estimated by nonvocal indicators, that is, by increased degree of striving to approach the front door (i.e., by striving to contact with a human either friendly or aggressively, see Supplementary movies 1–3). Hence, the same recording procedure applied to different fox strains made it possible to obtain high- and low-arousal stimuli of negative valence for the Aggressive and Unselected foxes and high- and low-arousal stimuli of positive valence for the Tame foxes (Gogoleva et al. 2010a, 2010c). The unfamiliar human represented an external stimulus for the foxes. The level of arousal in the focal fox depended on the distance between the human and the focal fox. Specifically, the arousal level in the focal fox increased by decreasing the distance between the human experimenter and the focal fox. In silver foxes, the changes of emotional arousal states in response to human approach, have been established in previous studies (Pedersen and Jeppesen 1990; Pedersen 1993; Bakken 1998; Bakken et al. 1999; Trut 1999; Kukekova et al. 2008b). Critically, research show that, in response to humans, Unselected silver foxes show fearful behaviors, Aggressive silver foxes show aggressive behaviors, and Tame silver foxes show friendly behaviors (Trut 1999; Trut et al. 2009). The valence content of vocalizations produced by the focal foxes (negative for Aggressive and Unselected silver foxes, and positive for Tame silver foxes) was inferred on the basis of these studies. For audio recordings (distance between vocalizing fox and the microphone: 0.25–1 m), we used a Marantz PMD-222 (D&M Professional, Kanagawa, Japan) cassette recorder with an AKG-C1000S (AKG-Acoustics Gmbh, Vienna, Austria) cardioid electret condenser microphone, and Type II chrome audiocassettes EMTEC-CS II (EMTEC Consumer Media, Ludwigshafen, Germany). The system had a frequency response of 0.04–14 kHz at a tape speed of 4.75 cm/s. For the purposes of the present study, recordings were digitized (with each test step taken as a separate file) at a 22.05 kHz sampling rate with 16 bit precision and then high-pass filtered at 0.1 kHz with Avisoft-SASLab Pro (Avisoft Bioacoustics, Berlin, Germany, Specht 2002). SSG classified each call visually to 1 of 8 types (whine, moo, cackle, growl, bark, pant, snort, and cough) by spectrogram (Hamming window, FFT-length 1024 points, frame 50%, and overlap 87.5%), blindly to the fox strain and to the number of the test step (based on Gogoleva et al. 2008), Only whines were adopted as playback stimuli, as only the whine is shared by Tame, Aggressive, and Unselected foxes in the human-approach context (Gogoleva et al. 2008, 2010a, 2010c, 2013). For creating the playback stimuli, we selected 27 individual foxes (9 foxes per strain) From each individual, we selected 1 low-arousal whine (from Step 1) and 1 high-arousal whine (from Step 3), which provided 27 paired low-/high-arousal stimuli in total (Figure 1). Our choice was based on the quality of the recordings, background noise, and vocalizations of other animals. We equalized all experimental stimuli to the same root-mean square amplitude (70 dB). Fade in/out transitions of 5 ms were applied to all stimuli to remove any transients.
Figure 1

Spectrogram (below) and waveform (above) of: (A) low-arousal whine and (B) high-arousal whine of the same individual Aggressive silver fox.

Spectrogram (below) and waveform (above) of: (A) low-arousal whine and (B) high-arousal whine of the same individual Aggressive silver fox. Mean percentage of correct responses for stimuli belonging to Tame, Aggressive, and Unselected silver foxes, averaged across participants. Error bars represent 95% confidence intervals.

Acoustic analysis

To explore the effect of specific acoustic cues on humans’ perception of arousal in whine calls across the 3 strains of silver fox, for each call, we measured the following 5 parameters: duration, tonality (harmonics-to-noise-ratio: HNR), SCOG, dominant frequency (DF)—that is, the frequency with the highest amplitude in the spectrum, and mean F0. We based the choice of the parameters to include in our analysis on findings from previous studies: duration, HNR, and F0 are shown to be linked to the emotional state of the caller (Morton 1977; Taylor and Reby 2010; Briefer 2012; Zimmermann et al. 2013) and SCOG affects the perception of arousal in humans (Sauter et al. 2010; Faragó et al. 2014). We performed an automated acoustic analysis of the acoustic parameters in PRAAT (v. 5.2.26; Boersma 2002) for all parameters, except for duration, which was measured in Avisoft-SASlab Pro (Table 1). The duration was measured with the standard marker cursor in the main window of Avisoft. HNR was measured using the “To Harmonicity (cc)” command in PRAAT with standard settings. SCOG was measured using the “To spectrum” and “Get center of gravity” commands (Power = 2.0). DF was measured using the “To Ltas:” command (bandwidth = 10 Hz, with no interpolation). Finally, the analysis of F0 was restricted to harmonics (integer multiples of F0). F0 was measured with the “Get pitch” command (Pitch settings: View range = 30–4,000 Hz, analysis method = cross-correlation. Advanced pitch settings: very accurate: “yes”, voicing threshold = 0.3).
Table 1

Descriptive table of acoustic values of the low- and high-arousal calls of each strain. For each of the acoustic parameters, namely DF (dominant frequency), duration, HNR (harmonic-to-noise-ratio), SCOG (spectral center of gravity), and F0 (mean fundamental frequency), the following values are provided: minimum and maximum, mean, and standard deviation of the mean

StrainArousal levelAcoustic parameterNMinimumMaximumMeanSD
Aggressivehighdf (Hz)937820701024.667602.020
duration (s)90.7261.1840.9900.140
HNR (dB)93.41514.4519.5844.103
scog (Hz)9857.5222072.2061427.819467.095
F0 (Hz)9377.064858.885628.297198.311
lowdf (Hz)961820221130.000471.296
duration (s)90.6791.3581.0160.233
HNR (dB)97.21415.68411.8433.546
scog (Hz)9902.7811795.1371225.265311.676
F0 (Hz)9379.728851.781573.562171.136
Tamehighdf (Hz)94021206714327.866
duration (s)90.2621.3990.7000.370
HNR (dB)9−1.08716.07210.9505.529
scog (Hz)9554.5403074.6461274.658719.675
F0 (Hz)9401.6551259.259596.700271.998
lowdf (Hz)943821301226463.128
duration (s)90.2321.0740.5730.255
HNR (dB)94.17616.26810.7234.341
scog (Hz)9874.0892327.4721447.183434.674
F0 (Hz)9320.834842.181522.081157.799
Unselectedhighdf (Hz)93061698603.333460.821
duration (s)90.7491.8641.1020.342
HNR (dB)94.32519.05210.6824.745
scog (Hz)9470.3491931.269869.907556.211
F0 (Hz)9182.492817.729392.011182.205
lowdf (Hz)91981986591.333580.197
duration (s)90.4351.1730.8240.298
HNR (dB)96.24614.96910.0462.695
scog (Hz)9325.5901304.692607.799333.890
F0 (Hz)9213.290505.722347.699102.763
Descriptive table of acoustic values of the low- and high-arousal calls of each strain. For each of the acoustic parameters, namely DF (dominant frequency), duration, HNR (harmonic-to-noise-ratio), SCOG (spectral center of gravity), and F0 (mean fundamental frequency), the following values are provided: minimum and maximum, mean, and standard deviation of the mean In addition, the first author inspected the F0 contour of each stimulus visually. When the visible contour did not overlap with the first harmonic, the parameters “Pitch floor” and “Pitch Ceiling”, but sometimes also “Silence Threshold”, “Voicing Threshold”, and “Octave jump” within the “Advanced pitch settings” menu were adjusted until the values identified by the algorithm visually matched the frequency distance between harmonics seen in the PRAAT spectrogram view window. These settings’ adjustments were made for 23 (out of 54) stimuli.

Human participants

Twenty-seven participants (mean age = 26.26 years; SD = 5.65 years; 15 female), recruited at the Vrije Universiteit Brussel (Belgium) participated in this experiment in exchange for monetary compensation. The experimental design adopted for this study was approved by the university ethical review panel in accordance with the Helsinki Declaration. All participants gave written informed consent.

Experimental design

To avoid any bias in data collection and preliminary analyses, the experimenter (PF) was blind to the fox strain each caller belonged to. The experimental interface was created in PsychoPy (standalone version 1.81.oorc1; Peirce 2007). Participants were individually tested in a sound-attenuated room. The entire procedure was computerized. Stimuli were played binaurally over Shure SRH440 headphones. Participants were informed that the aim of the study was to understand whether humans are able to identify different levels of arousal expressed in animal vocalizations. Before the start of the experiment they were instructed to read an information sheet where the definition of arousal and the experimental procedure were explained. We provided the following definition of arousal: ‘Arousal is a state of the brain or the body reflecting responsiveness to sensory stimulation. Arousal level typically ranges from low (very subdued) to high (very excited). Examples of low-arousal states (e.g., of low responsiveness to sensory stimulation) are calmness or boredom. Examples of high-arousal states (e.g., of high responsiveness to sensory stimulation) are anger or excitement.” For familiarization with the experimental procedure, each participant completed 5 practice trials, each consisting of a pair of baby cries (www.freesound.org) varying in arousal level. During this practice phase, explicit instructions on the experimental procedure were displayed on the monitor. In the experimental phase, 27 pairs of calls were played in a randomized order across participants. Each trial in both phases was divided into 3 parts: Sound playback: One low and one high-arousal vocalization emitted by the same individual were played with an inter-stimulus interval of 1 s. Order within pairs was randomized across participants. One sound would play while the letter “A” appeared on screen. At the end of sound A playback, the letter “A” faded out and then the other sound played while the letter “B” appeared on screen. Relative rating of arousal: Participants were asked to indicate which vocalization expressed a higher level of arousal by clicking on the corresponding letter with the mouse. Given the short duration of our stimuli (see Table 1), to favor accurate assessment of sound features, participants could replay each sound ad libitum by pressing either letter (A or B) on the keyboard. No feedback was provided. Absolute rating of arousal: Participants were asked to rate the level of arousal expressed in each vocalization by using a Likert scale ranging from 1 = very subdued to 7 = very excited. Again, they could replay each vocalization separately by pressing the corresponding letter on the keyboard. No feedback was provided.

Statistical analysis

All statistical analyses were performed using R 3.1.2 (R Development Core Team 2013). A binomial test and a signal detection analysis were performed to assess participants’ accuracy in identifying high-arousal calls within each call pair across the 3 strains of silver fox. The dependent variable was the proportion of correct choices in participants’ responses (where chance = 0.50). We used a generalized linear mixed model (GLMM) to analyze humans’ overall accuracy in identifying the high-arousal vocalization across strains. GLMMs were used because they allow fixed and random factors to be defined. Data across all participants were modeled using a binomial distribution. Participant ID was entered as a random factor, fox strain was entered as a fixed factor, and correct/incorrect response was entered as the outcome variable. False discovery rate (FDR) adjustments were applied to conduct pairwise comparisons. FDR was controlled at α level 0.05 following the procedure proposed by Benjamini and Hochberg (1995): For m tests, rank the P-values in ascending order P(1) ≤ P(2) ≤ …/P(m), and denote by H(i) the null hypothesis corresponding to P(i); Let k be the largest i for which P(i) ≤ (α/m) * i and reject all null hypotheses H(1) … H (k). This means that, starting with the highest P-value each P is checked for this requirement; at the first P that meets the requirement its corresponding null hypothesis and all those having smaller P’s are rejected (Verhoeven et al. 2005). In addition, Cohen's (Cohen 1992) d effect sizes were calculated. To assess which acoustic parameters affect human ability to identify the vocalization expressing a higher level of arousal within each strain, we performed separate GLMMs for each strain. A separate GLMM was used to examine the acoustic parameters that predict participants’ correct identification of high-arousal calls within each pair of arousal calls. Here, participant ID was entered as a random factor, acoustic parameters (duration, DF, SCOG, HNR, and F0 ratios) were entered as fixed factors, and the correct or incorrect response was entered as the outcome variable. Finally, a multiple linear regression analysis was computed to detect the acoustic parameters that predict absolute ratings for arousal level in each call. In this analysis, we included acoustic parameters as fixed factors and mean ratings for perceived emotional arousal in each call as outcome variable. For all the analyses including acoustic parameters as fixed factors, we used a model selection procedure based on the Akaike’s information criterion adjusted for small sample size (AICc) to identify the model(s) with the highest power to explain variation in the dependent variable. The AICc was used to rank the GLMMs and to obtain model weights (model.sel function, MuMIn library). Selection of the model(s), that is, of the model(s) with the highest power to explain variation in the dependent variable, is based on lowest AICc. When the difference between the AICc values of 2 models (ΔAICc) is less than 2 units, both models are considered as good as the best model (Symonds and Moussalli 2011). Models with ΔAICc up to 6 have considerably less support by the data. Models with values greater than 10 are sufficiently poorer than the best AIC model as to be considered implausible (Anderson and Burnham 2002). Finally, we computed binary logistic regression models within the generalized linear model framework to assess any effect of order of sound playback within each call pair on participants response and of number of sound replays on participants’ accuracy within the relative rating task.

Results

The binomial test and signal detection analysis revealed that participants’ accuracy was above chance level for vocalizations of Aggressive silver foxes (proportion of correct responses: 0.59; P = 0.007; d’ = 0.455) and of Unselected silver foxes (proportion of correct responses: 0.65; P < 0.001; d’ = 0.771). Remarkably, we detected a significant effect in the opposite direction for stimuli of Tame silver foxes, where proportion of incorrect responses was higher than the expected 0.50 (proportion of incorrect responses: 0.60; P = 0.003; d’ = −0.507) (Figure 2). In line with this result, the analysis performed within the GLMM revealed a significant effect of strain (Wald χ22 = 31.681, P < 0.001). Pairwise comparisons performed using the FDR correction revealed a significant difference between the effect of Tame and Aggressive strains (Q < 0.001; d = −0.386) and between Tame and Unselected strains (Q < 0.001; d = −0.515).
Figure 2

Mean percentage of correct responses for stimuli belonging to Tame, Aggressive, and Unselected silver foxes, averaged across participants. Error bars represent 95% confidence intervals.

As shown in Table 2 (A), analyses performed within the GLMM revealed a significant effect of all the acoustic predictors included in our analysis on humans’ accuracy in identifying arousal level in whine calls across all strains. However, critically, the model selection procedure ranked the models where F0, HNR, or SCOG ratios were excluded from the analyses as the weakest models, not resulting in a ΔAICc < 2. A separate GLMM computed on each strain identified the following acoustic parameters as predictors of humans’ accuracy in identifying arousal level in whine calls: DF and F0 ratios for Aggressive foxes, SCOG, F0, and HNR ratios for Unselected foxes, and duration and HNR ratios for Tame foxes. Model selection procedure applied on each of these models was in line with these results (Table 2 (B)).
Table 2

Relative rating task: (A) Values of the GLMMs computed across and within silver fox strains. We assessed acoustic predictors of humans’ ability to identify vocalizations expressing higher levels of arousal across and within silver fox strains. Bold type indicate P ≤0.05; degrees of freedom = 1 for all fixed factors. (B) Outcome of model selection procedure based on AICc. Degrees of freedom = 6 for all models. Bold type indicates models with the strongest support based on log likelihood (logLik), akaike weights and the difference between the AICc values of two models (ΔAICc ≤2.0)

(A) Generalized linear mixed models(B) Model selection procedure
Across strains

Fixed effectχ2P(χ2)ModellogLikAICcΔAICcAkaike weight

duration4.2790.039exluding duration-422.123856.3620.0000.592
DF6.2580.012exluding DF-423.113858.3421.9790.220
F06.5700.010exluding F0-423.269858.6542.2910.188
HNR21.582<.001exluding HNR-430.775873.66617.3030.000
SCOG21.936<.001exluding SCOG-430.952874.02017.6580.000

Aggressive silver foxes

Fixed effectχ2P(χ2)ModellogLikAICcΔAICcAkaike weight

duration0.0110.916exluding duration-135.460283.2770.0000.438
HNR1.5410.214exluding HNR-136.225284.8071.5300.204
SCOG1.6190.203exluding SCOG-136.264284.8841.6070.196
DF2.7300.099exluding DF-136.820285.9952.7180.112
F04.3560.037exluding F0-137.633287.6214.3450.050

Unselected silver foxes

Fixed effectχ2P(χ2)ModellogLikAICcΔAICcAkaike weight

duration0.2780.598exluding duration-121.507255.3700.0000.551
DF0.8080.369exluding DF-121.772255.9010.5300.423
SCOG7.1260.008exluding SCOG-124.931262.2186.8470.018
F09.3810.002exluding F0-126.059264.4739.1030.006
HNR10.9420.001exluding HNR-126.839266.03410.6640.003

Tame silver foxes

Fixed effectχ2P(χ2)ModellogLikAICcΔAICcAkaike weight

SCOG0.0100.920exluding SCOG-116.492245.3400.0000.399
F00.0950.757exluding F0-116.535245.4250.0850.383
DF1.6910.193exluding DF-117.333247.0211.6810.172
Duration4.3360.037exluding duration-118.655249.6664.3250.046
HNR81.549<0.001exluding HNR-157.262326.87981.5390.000
Relative rating task: (A) Values of the GLMMs computed across and within silver fox strains. We assessed acoustic predictors of humans’ ability to identify vocalizations expressing higher levels of arousal across and within silver fox strains. Bold type indicate P ≤0.05; degrees of freedom = 1 for all fixed factors. (B) Outcome of model selection procedure based on AICc. Degrees of freedom = 6 for all models. Bold type indicates models with the strongest support based on log likelihood (logLik), akaike weights and the difference between the AICc values of two models (ΔAICc ≤2.0) Absolute rating task: (A) Values of the multiple linear regression model computed on human rating of arousal in silver fox calls. We assessed acoustic predictors of humans’ perceived arousal in silver fox calls, as reported using a rating scale spanning from 1 (very subdued) to 7 (very excited). Bold type indicate P ≤ 0.05; degrees of freedom = 1 for all fixed effects. (B) Outcome of model selection procedure based on AICc. Degrees of freedom = 6 for all models. Bold type indicates models with the strongest support based on loglikelihood (logLik), akaike weights, and the difference between the AICc values of two models (ΔAICc ≤2.0) The multiple linear regression analysis for the absolute ratings identified a significant effect of F0 and SCOG values of each call as the best predictors of the rated level of emotional arousal in each call—across strains (Table 3 (A)). The model selection procedure applied to this model was in line with these findings, and identified the models excluding DF as the weakest model, not resulting in a ΔAICc < 2. However, according to this procedure, the model excluding duration or HNR, which have ΔAICc < 6, as models that should not be discounted (Table 3 (B)). Results from a Shapiro Wilk test indicated that errors in the multiple linear regression analysis were normally distributed (W (54) = 0.981, P = 0.541). To assess whether our model met the assumption of multicollinearity, we obtained the variance of inflation factor (VIF) value for the best predictors, and found that it was not substantially greater than 1 for any of them (SCOG: 1.662; F0: 1.613; HNR: 1.044). In addition, tolerance values were not below 0.2 (SCOG: 0.602; F0: 0.620; HNR: 0.958). Thus, we can exclude that collinearity is a problem for this model (Bowerman and O'connell 1990). Finally, a Durbin–Watson test revealed that the residuals are not linearly auto-correlated (d = 2.301).
Table 3

Absolute rating task: (A) Values of the multiple linear regression model computed on human rating of arousal in silver fox calls. We assessed acoustic predictors of humans’ perceived arousal in silver fox calls, as reported using a rating scale spanning from 1 (very subdued) to 7 (very excited). Bold type indicate P ≤ 0.05; degrees of freedom = 1 for all fixed effects. (B) Outcome of model selection procedure based on AICc. Degrees of freedom = 6 for all models. Bold type indicates models with the strongest support based on loglikelihood (logLik), akaike weights, and the difference between the AICc values of two models (ΔAICc ≤2.0)

(A) Multiple linear regression model(B) Model selection procedure
Across strains

Fixed effectFP(F)ModellogLikAICcΔAICcAkaike weight

DF0.1500.451exluding DF-37.29788.3800.0000.558
duration0.6330.125exluding duration-38.31490.4162.0360.202
HNR0.7270.101exluding HNR-38.50690.8002.4190.166
F01.2150.035exluding F0-39.49192.7704.3890.062
SCOG2.0910.007exluding SCOG-41.16996.1257.7450.012
Finally, our analyses did not reveal any effect of order on correct response (z = 0.486, P = 0.627). No effect of number of sound replays on participants’ correct response was detected (z = 1.208, P = 0.227).

Discussion

We found that humans are able to identify high arousal in vocalizations (namely, whine calls) of Aggressive and Unselected silver foxes. Intriguingly, in Tame silver foxes human participants identified low-arousal calls as expressing high arousal. Thus, participants’ accuracy was significantly lower for whine calls produced by tame silver foxes, which have positive attitudes toward people, in comparison with Aggressive and Unselected foxes, which have negative attitude toward humans. In addition, we found that F0, HNR, and SCOG ratios predicted human accuracy in identifying high-arousal calls across all silver fox strains. Separate analyses revealed that different acoustic parameters affect human accuracy in identifying high-arousal calls within each strain. Specifically, our analyses suggest that DF and F0 ratios affect human accuracy in identifying high-arousal calls of Aggressive silver foxes. Furthermore, we found that SCOG, F0, and HNR ratios affect human accuracy in identifying high-arousal calls of Unselected silver foxes, and that duration and HNR ratios affect human accuracy in identifying high-arousal calls of Tame silver foxes. Finally, our analyses suggest that F0 and SCOG are reliable predictors of humans’ absolute ratings of arousal in our stimuli, although duration and HNR should not be discounted. Our findings are consistent with previous research showing that humans are able to perceive arousal in vocalizations of cats (Nicastro and Owren 2003; McComb et al. 2009), dogs (Pongrácz et al. 2005; Faragó et al. 2014; Albuquerque et al. 2016), and piglets (Tallet et al. 2010; Maruščáková et al. 2015). Our results confirm findings from Filippi et al. (2016), suggesting that humans are able to recognize arousal in vocalizations emitted by members of species varying in size, social structure, and ecology. However, we have extended this line of research by disentangling human sensitivity to high-arousal calls with negative valence from human sensitivity to high-arousal calls with positive valence in silver foxes. In line with these studies, our work confirms that acoustic parameters associated with pitch perception play a key role in affecting human participants’ ability for both identifying high arousal with negative valence in whine calls across the 3 strains of silver fox, and in assessing their absolute level of arousal on a rating scale. In addition, in line with research on vocal expression of arousal in mammals, which have identified in HNR an acoustic correlate of high arousal (Briefer 2012; but see Blumstein and Chi 2012), our analyses, which examines human perception of vocalizations with arousal content, suggest that HNR facilitates recognition of negative arousal in vocalizations of Unselected silver foxes and positive arousal in Tame silver foxes. Moreover, we found that HNR may have a role in affecting humans’ absolute rating of arousal in silver foxes. One interesting finding is that duration affects humans’ accuracy in identifying high-arousal calls with positive valence, that is, high-arousal calls emitted by Tame silver foxes. Previous research identified duration as a predictor of human perception of valence in animal calls, without controlling for arousal level (Faragó et al. 2014; Maruščáková et al. 2015). Our work complements this research, and particularly the study conducted by Faragó et al. (2014) on human perception of dog calls, in that we show that duration predicts humans’ accuracy in identifying high arousal in silver fox calls with positive valence. In addition, in line with findings reported in Faragó et al. (2014), we found that duration may be used for the absolute rating of arousal in calls across all strains. Crucially, further experimental investigations are needed to estimate the effect of perceived loudness of the calls on correct assessment of their relative level of arousal. It is possible that humans adopt the same kind of changes in perceived frequency-related parameters to assess arousal levels across silver foxes strains. However, recent research has provided evidence for the use of shared mechanisms in arousal perception across phylogenetically distant species (Belin et al. 2008; Altenmüller et al. 2013; Faragó et al. 2014; Filippi 2016; Filippi et al. 2016; Song et al. 2016). These findings suggest that mechanisms underlying perception and plausibly also expression of emotional arousal—which are related to stress induced higher effort in vocalization—may have emerged in the early stages of animal evolution as a result of selection pressures. One interesting implication of our finding is that high-arousal calls with positive valence may not be as salient as high-arousal calls with negative valence to the human ear. Further research may analyze the perceptual saliency of acoustic correlates of negative emotional arousal in vocalizations produced by individuals belonging either to the same or to different species. Humans’ ability to recognize arousal in calls with positive valence might have not been selected by evolution because they are not as crucial for survival as arousal calls produced in negative-valenced contexts. Indeed, while the latter are a direct response to different degrees of external threat or danger, arousal calls with positive valence are emitted in contexts that may not be directly linked to survival. Therefore, in contrast to the ability to identify arousal calls with positive valence, humans may have evolved the ability to recognize arousal calls with negative valence as an adaptive trait. Further work is required to establish whether humans are able to identify high-arousal silver foxes calls that were elicited by the approach of another animal species. This analysis would add further support to the investigation of acoustic variation as a correlate of emotional arousal rather than as a function of predator species. Notably, in the recordings adopted in our experiments, the emotional content of silver fox calls was inferred based on previous findings reporting on hormonal responses to human approach, and on observational indicators of the focal fox, namely on motor activity of the caller in response to the approaching behavior to the human experimenter. However, motor response in the 3 strains within the human-approach experimental setting described in the “Acoustic recordings and emotion classification” section was not statistically quantified. This is a crucial limitation of the present study. Future work should aim to classify arousal calls by combining quantitative assessment of observational correlates of arousal (e.g., motor response) with quantitative assessment of arousal in terms of physiological and/or neural responses of each specific caller during the production of calls. This is an important issue for future research, which should aim to integrate multiple types of data to quantify the exact degree of emotional arousal and valence. Specifically, further research is required to examine how behavioral observations of the contexts in which the vocalization is emitted can be mapped to data on brain activity (Belin et al. 2008; Panksepp 2011; Ocklenburg et al. 2013; Andics et al. 2014) and physiological states of the caller such as heart rate, adrenaline, or stress hormone levels at the time of vocal response to human approach (Paul et al. 2005; Briefer et al. 2015a, 2015b; Stocker et al. 2016). This quantitative assessment of arousal and valence of each caller, recorded during vocal production, may be adopted to fine-tune the examination of responses in listeners, assessing their neurological, physiological, and behavioral activity in response to each call. Crucially, further research in this direction might help to identify which emotion correlates in the focal Tame silver foxes led human participants to categorize low-arousal calls as expressing a high arousal. Generally, within this research framework, the analysis of acoustic correlates of emotional content in nonhuman animal vocalizations may be particularly valuable in assessing and improving animal welfare. In conclusion, our findings provided empirical evidence for humans’ sensitivity to arousal vocalizations across 3 strains of silver fox that have different genetically based predispositions toward approaching humans. We found that humans are able to identify high emotional arousal in whine calls with negative arousal—produced by Aggressive and Unselected silver foxes, but not in calls with positive valence, which were produced by Tame foxes. Our data did not identify in any of frequency-related parameters (F0, SCOG, DF) the best predictors of humans’ assessment of arousal content in whine calls of Tame silver foxes, suggesting that these types of acoustic parameters in the calls may be key to recognition of negative arousal in animal calls (Lingle et al. 2012; Teichroeb et al. 2013; Lingle and Riede 2014; Volodin et al. 2017). This work extends our understanding of vocal communication between species, providing key insights on the effect of acoustic correlates of emotional arousal and valence. Finally, this frame of investigation may enhance our understanding of the adaptive role of “eavesdropping” on heterospecific calls, providing key insights into the evolution of inter-species acoustic communication.

Supplementary material

Supplementary material can be found at https://academic.oup.com/cz. Click here for additional data file.
  47 in total

Review 1.  Genetics of behavior in the silver fox.

Authors:  Anna V Kukekova; Svetlana V Temnykh; Jennifer L Johnson; Lyudmila N Trut; Gregory M Acland
Journal:  Mamm Genome       Date:  2011-11-23       Impact factor: 2.957

Review 2.  An integrative and functional framework for the study of animal emotion and mood.

Authors:  Michael Mendl; Oliver H P Burman; Elizabeth S Paul
Journal:  Proc Biol Sci       Date:  2010-08-04       Impact factor: 5.349

Review 3.  Lateralisation of conspecific vocalisation in non-human vertebrates.

Authors:  Sebastian Ocklenburg; Felix Ströckens; Onur Güntürkün
Journal:  Laterality       Date:  2011-12-09

4.  Mobbing vocalizations as a coping response in the common marmoset.

Authors:  N Cross; L J Rogers
Journal:  Horm Behav       Date:  2005-08-15       Impact factor: 3.587

Review 5.  [An experiment on fox domestication and debatable issues of evolution of the dog].

Authors:  L N Trut; I Z Pliusnina; I N Os'kina
Journal:  Genetika       Date:  2004-06

6.  Dogs recognize dog and human emotions.

Authors:  Natalia Albuquerque; Kun Guo; Anna Wilkinson; Carine Savalli; Emma Otta; Daniel Mills
Journal:  Biol Lett       Date:  2016-01       Impact factor: 3.703

7.  Deer mothers are sensitive to infant distress vocalizations of diverse mammalian species.

Authors:  Susan Lingle; Tobias Riede
Journal:  Am Nat       Date:  2014-09-05       Impact factor: 3.926

Review 8.  A framework for studying emotions across species.

Authors:  David J Anderson; Ralph Adolphs
Journal:  Cell       Date:  2014-03-27       Impact factor: 41.582

Review 9.  Emotional and Interactional Prosody across Animal Communication Systems: A Comparative Approach to the Emergence of Language.

Authors:  Piera Filippi
Journal:  Front Psychol       Date:  2016-09-28

10.  Animal signals and emotion in music: coordinating affect across groups.

Authors:  Gregory A Bryant
Journal:  Front Psychol       Date:  2013-12-25
View more
  4 in total

1.  Positive and negative contexts predict duration of pig vocalisations.

Authors:  Mary Friel; Hansjoerg P Kunc; Kym Griffin; Lucy Asher; Lisa M Collins
Journal:  Sci Rep       Date:  2019-02-14       Impact factor: 4.379

2.  Exploring the cerebral substrate of voice perception in primate brains.

Authors:  Clémentine Bodin; Pascal Belin
Journal:  Philos Trans R Soc Lond B Biol Sci       Date:  2019-11-18       Impact factor: 6.237

3.  Polyphony of domestic dog whines and vocal cues to body size.

Authors:  Olga V Sibiryakova; Ilya A Volodin; Elena V Volodina
Journal:  Curr Zool       Date:  2020-08-13       Impact factor: 2.624

4.  Human listeners' perception of behavioural context and core affect dimensions in chimpanzee vocalizations.

Authors:  Roza G Kamiloğlu; Katie E Slocombe; Daniel B M Haun; Disa A Sauter
Journal:  Proc Biol Sci       Date:  2020-06-17       Impact factor: 5.349

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.