Literature DB >> 19136972

Neural correlates of categorical perception in learned vocal communication.

Jonathan F Prather1, Stephen Nowicki, Rindy C Anderson, Susan Peters, Richard Mooney.   

Abstract

The division of continuously variable acoustic signals into discrete perceptual categories is a fundamental feature of vocal communication, including human speech. Despite the importance of categorical perception to learned vocal communication, the neural correlates underlying this phenomenon await identification. We found that individual sensorimotor neurons in freely behaving swamp sparrows expressed categorical auditory responses to changes in note duration, a learned feature of their songs, and that the neural response boundary accurately predicted the categorical perceptual boundary measured in field studies of the same sparrow population. Furthermore, swamp sparrow populations that learned different song dialects showed different categorical perceptual boundaries that were consistent with the boundary being learned. Our results extend the analysis of the neural basis of perceptual categorization into the realm of vocal communication and advance the learned vocalizations of songbirds as a model for investigating how experience shapes categorical perception and the activity of categorically responsive neurons.

Entities:  

Mesh:

Year:  2009        PMID: 19136972      PMCID: PMC2822723          DOI: 10.1038/nn.2246

Source DB:  PubMed          Journal:  Nat Neurosci        ISSN: 1097-6256            Impact factor:   24.884


INTRODUCTION

One way perception enables adaptive behavior is by grouping variable stimuli into classes. An impressive example of this synthetic capacity of the nervous system occurs when continuous changes in a stimulus parameter are perceived as discrete perceptual categories1. Categorical perception plays a prominent role in processing acoustic signals important to social communication2, including vocalizations made by humans3,4, as well as by monkeys5 rodents6, birds7, and frogs8,9. In spoken English, for example, the distinction between the speech sounds that elicit the phonemes /ba/ and /pa/ stems from a difference in delay time between the initial sound made by opening the lips and the onset of vocal fold vibration (“voice onset time”). Voice onset time varies considerably from individual to individual, but this variation is perceived categorically3,4, with native English speakers recognizing a categorical boundary between /ba/ and /pa/ at ~ 12 ms3,4. Presumably, categorical processing of speech more generally facilitates comprehension by generating perceptual constancy in the face of individual variation in many dimensions including vocal pitch, timbre, and tempo. Despite the prominent role of categorical perception in the processing of speech and other natural signals, the neural mechanisms underlying this cognitive ability remain poorly studied. The fact that some animals have been shown to perceive human speech sounds categorically, even when they do not learn their own species-typical vocalizations6, suggests that the capacity for categorical perception reflects some fundamental and innate brain mechanism. At the same time, however, the underlying mechanism must also be sensitive to experience, because perceptual boundaries that define speech sounds vary across native speakers of different languages3,4, indicating that the boundary, like other features of speech, is learned. What has been lacking is a model system in which the neural underpinnings of categorical perception could be studied for a natural, learned communication system. Here we ask whether sensorimotor neurons known to be important to learned vocal communication in songbirds could also mediate categorical perception of learned vocalizations. Swamp sparrows (Melospiza georgiana) are especially appropriate for examining the neural events underlying categorical perception of learned vocalizations for several reasons. First, like all songbirds studied to date, swamp sparrows learn their song notes by imitation10, a feature of human speech that is otherwise rare among animals11. Second, behavioral experiments have shown that male swamp sparrows use categorical perception to distinguish fundamental acoustic elements in their species-typical vocal repertoire7. Swamp sparrow songs comprise repeated groups of 2 to 5 “notes” (Fig. 1A) composed of short pure-tonal frequency sweeps, with note categories differing primarily in duration, bandwidth and rate of change in frequency12. Previous work has shown that males use categorical perception to distinguish between note types that, like the phones produced in speech, are produced with considerable variation by different individuals but are grouped into natural categories. Specifically, note types in Categories I and VI share most spectral features but differ along a duration continuum7,12,13, and individual male swamp sparrows in a wild population in New York State perceive these differences categorically, with an estimated perceptual boundary at a duration of 13 ms (Figs. 1B-C)7.
FIGURE 1

Earlier behavioral tests with swamp sparrows reveal categorical perception of changes in individual song notes

(a) Male swamp sparrows sing a small repertoire of distinct song types composed of trilled syllables that comprise multiple notes (2-5 song types, each ~ 2 sec duration). (b, c, d) Categorical perception in swamp sparrows was demonstrated previously7 by (b) splicing out one note in the syllable (note C) and replacing it with a note with similar spectral characteristics but different duration. Replacement note durations differed by equal increments on a logarithmic scale. (c) Behavioral testing revealed that swamp sparrows perceive strong differences, evident as robust expression of aggressive displays, when note transitions spanned an estimated boundary of 13 ms (stimulus set 2) but perceive little or no difference when transitions did not cross that boundary (stimulus sets 1 and 3, figure adapted from Nelson and Marler7), indicating that swamp sparrows employ categorical perception in their song discrimination.

A third feature that advances the swamp sparrow as a model for investigating neural activity underlying categorical perception is that their brain, as in other songbirds, contains a discrete set of specialized sensorimotor structures devoted to learning, producing and perceiving song14-19. Previously, we showed that one of these structures, the nucleus HVC, contains a certain class of striatum-projecting neurons (HVCX cells; Figs. 2B-C, left) that respond to only one song type in the bird’s repertoire20,21. Three observations implicate song type-selective HVCX neurons in song perception. First, song perception is impaired by lesions to HVC15,16. Second, song perception is also impaired by lesions to the striatal portion of an anterior forebrain pathway into which HVCX cells project their axons22. Third, some HVCX cells display a sensorimotor correspondence reminiscent of mirror neurons in the monkey cortex that are hypothesized to play an important role in perception20,23.
FIGURE 2

Extracellular recordings reveal differences in the auditory response of identified single units in nucleus HVC

(a) A parasagittal schematic of the swamp sparrow brain at the level of HVC. Antidromic stimulation was used to identify HVC neurons19 as projecting to striatal areas implicated in song learning and perception (Area X), projecting to premotor structures (robust nucleus of the arcopallium, RA), of making local connections within HVC (HVC interneurons, omitted for clarity). Each cell type within HVC receives auditory input45 (dotted line; D = dorsal, R = rostral). (b, c) Auditory responses of HVCX neurons (N = 24 of 29 cells, 5 birds) were phasic and typically evoked by only one song type in the bird’s repertoire (“primary song type20”, N = 23 of 24 responsive cells, identity of primary song type varied across cells). Auditory responses of HVCINT cells (N = 18 of 18 cells, 5 birds) consisted of tonic increases in activity in response to many or all song types in the repertoire. In contrast, HVCRA neurons were unresponsive to any stimulus that was presented (top: raw data traces; middle: auditory response raster; bottom: auditory response PSTH, 10 ms binsize; N = 16 cells, 5 birds). (c) HVCX phasic responses occurred at a restricted phase in the syllable of the primary song type (top: auditory response raster; bottom: auditory response PSTH, 1 ms binsize), whereas HVCINT cells responded throughout the syllable duration.

Although HVCX cells appear to be well suited to a role in perception, the extent to which their auditory properties are correlated with the bird’s perception of categorical differences among song features is unknown. Therefore, we used a lightweight chronic recording device20,24 to assess whether HVCX cells in freely behaving adult male swamp sparrows respond categorically to changes in note duration, known from behavioral experiments to be a key feature that distinguishes between note type Categories I and VI7.

RESULTS

In initial experiments, we recorded from antidromically identified HVC neurons in freely behaving adult male swamp sparrows and presented the bird’s song types through a speaker located near its perch. These recordings confirmed that HVCX neurons responded to a single song type in the bird’s repertoire (the “primary song type”), and also revealed that robust auditory responses could be evoked in interneurons (HVCINT cells). In contrast, all of the cells we sampled of the other projection neuron type in HVC, which innervates the song premotor nucleus RA (HVCRA cells), were unresponsive to auditory stimuli. To test whether individual HVCX neurons or interneurons express categorical responses to changes in the duration of notes in the primary song type, we first identified the primary song type to which a cell responded and then identified a note in the repeated syllable comprising that song type that was in either a Category I (< 13 ms) or a Category VI (> 13 ms) note type as defined in previous studies7,12,13. We then replaced the Category I or Category VI note in the primary song type with a note that had similar spectral features but a different duration, following the same procedures used in prior behavioral studies (Fig. 3A) in which replacement notes were chosen that differed in their durations by equal increments on a logarithmic scale7 (see Methods for additional details). The resulting synthetic syllable was assembled into a trill to form a stimulus song, with all other notes, inter-note intervals and trill rate identical to the natural song. Repeating this procedure using different replacement notes to create each stimulus, we presented to each cell a set of song stimuli comprising 5 to 11 variants of the primary song type, each differing only in the duration of a single replacement note in each trilled syllable, with durations spanning a range from notes that would be classified as Category I to those that would be classified as Category VI.
FIGURE 3

Manipulation of song note duration reveals categorical auditory responses of HVCX neurons

(a) Individual notes in the primary song type syllable (note C in top spectrogram) were replaced by another swamp sparrow song note with similar spectral characteristics but different duration7. Auditory responses in this HVCX cell were strong among one group of replacement notes (4, 8, 16 ms) but weak among another group (27, 31 ms; third row: auditory response PSTH, 1 ms binsize; bottom row: stimulus syllables, gray boxes indicate replacement notes). (b) Categorical responses like those in (a) were evident in all HVCX cells in this bird that responded to the same song type (N = 4, left, open blue squares indicate the cell in panel (a)). In contrast, categorical activity was not evident in any of the HVCINT neurons that responded to the same song type in the same bird (N = 3, right). In behavioral testing, this same bird responded to differences between stimuli that crossed the putative neurophysiological boundary but not to differences between stimuli that did not cross the boundary (Supp. Fig. 4). (c) Among all cells tested in all birds (N = 22 of 24 auditory responsive HVCX cells, 5 birds and 18 of 18 responsive HVCINT cells, 5 birds), categorical activity was consistently evident for HVCX neurons (N = 19 of 22 cells, 5 birds), with stronger response strengths evoked by replacement notes in the same category as the note that was replaced. In contrast, categorical activity was very rarely evident in HVCINT neurons (N = 2 of 18 cells, 2 birds). (d) Categorical responses were observed in HVCX cells regardless of whether the duration of note that was replaced was naturally short or long, and the location of the categorical boundary (black triangle) was consistent across HVCX cells and birds (21 ± 4 ms, mean ± SD; 10 cells shown for clarity). A subset of 10 categorically responsive HVCX cells (2 birds) was tested further to probe the categorical boundary at a higher resolution (19, 22, 25 ms; e.g. thick blue line), revealing an estimated categorical boundary (20 ± 4 ms) very similar to that estimated using the full dataset (21 ± 4 ms; p = 0.47, unpaired t-test).

Categorical auditory responses are evoked by changes in duration of individual notes

Acoustic presentation of variants of the primary song type revealed that the auditory responses of HVCX neurons, but not interneurons, are highly sensitive to changes in note duration (Figs. 3A-D). Most HVCX neurons responded robustly to song stimuli containing replacement notes with durations within the range of the target note category but weakly or not at all to variants of the primary song type with replacement notes having durations outside that category (N = 19 of 22 HVCX cells, 5 birds). Robust responses were invariably evoked by stimuli containing replacement notes with durations that would fall unambiguously into the same category as the target note in the natural song (e.g., Fig. 3A, note durations 4, 8 ms; target note duration = 7 ms; replacement notes near the categorical boundary discussed below). Following the criteria established in previous studies of categorical perception7,25-27, these cells were deemed categorically responsive (see Methods for additional details). Categorical responsiveness was evident in HVCX cells both within and across birds (within bird: Fig. 3B, left, N = 4 of 4 cells, 1 bird, p < 0.001; across birds: Fig. 3C, left; p < 0.001, Mann-Whitney U test; N = 19 cells, 5 birds), regardless of whether the original target note was of the shorter Category I type or the longer Category VI type (Fig. 3D). Beyond the sheer abundance of categorically-responsive HVCX neurons, two observations indicate that the HVCX neuronal population encodes note duration in a categorical manner. First, the three HVCX cells that did not respond categorically also did not display bell-shaped tuning curves to intermediate note durations, as might be expected if note duration is represented continuously (Supp. Fig. 1). Second, our dataset differed significantly from a model in which note duration is encoded linearly by HVCX cells (Supp. Fig. 2). In contrast to HVCX neurons, HVCINT cells responded similarly regardless of whether the replacement note was of the same category or a different category as the target note (Figs. 3B-C, right; p = 0.24, Mann-Whitney U test; N = 16 of 18 cells, 5 birds). These results indicate that categorical responses to changes in note duration are displayed by only one subset of auditory responsive HVC neurons, namely those that project to a striatal pathway important in song perception15,22.

Estimating the neuronal response boundary

To probe the link between these categorical neuronal responses and song perception, we first estimated the categorical boundary evident in the auditory responses of HVCX neurons. Interpolation at the transition from strong to weak responses yielded an estimated categorical response boundary at 21 ± 4 ms (mean ± SD, N = 19 cells, 5 birds). We further probed the location of the categorical response boundary at higher resolution in a subset of HVCX cells using stimuli containing replacement notes with durations near the estimated boundary (e.g., thick blue line in Fig. 3D; 19, 22, 25 ms; N = 10 cells, 2 birds, Supp. Fig. 3). This higher resolution investigation of the categorical transition yielded an estimated boundary (20 ± 4 ms), very similar to that obtained using the full dataset (21 ± 4 ms, p = 0.47, unpaired t-test). Such a sharp response transition is a hallmark of categorical activity3,4,28 (see also Supplementary Methods), further establishing the auditory responses of HVCX neurons as categorical.

Categorical boundary in neural response predicts a novel perceptual boundary

Although the categorical responses of HVCX cells are consistent with their playing some role in the categorical perception of note type based on duration, we found it curious that the categorical boundary estimated from our neural data (~ 21 ms) was different from the perceptual boundary reported previously for swamp sparrows based on behavioral testing (~ 13 ms)7. This discrepancy suggests that the neural responses we measured are not linked to perception of note duration or, alternatively, that the perceptual boundary in the birds we studied differed from the previously reported values. All birds from which we collected neural data were obtained from a population in northwestern Pennsylvania, whereas the previous behavioral study examined a population from upstate New York, some 540 km distant7. As with most songbirds, geographically distinct populations of swamp sparrows learn different song dialects29, raising the possibility that the perceptual boundary separating Category I and Category VI notes may be influenced by learning and thus vary across populations. To test whether the discrepancy between our neural data and previous behavioral measures of a categorical boundary reflects a learned dialect difference, we first compared the acoustic characteristics of Category I and Category VI notes obtained from the Pennsylvania and New York populations. We found that the distributions of note durations differ between these two populations, as would be predicted by a dialect difference in this phonological feature (Fig. 4A). We next asked whether population-specific differences in note durations are paralleled by population differences in song perception. We tested the categorical perceptual boundary for note duration in the Pennsylvania population, replicating the behavioral experiment done by Nelson and Marler on the New York population7, in which habituation to repeated presentation of one song stimulus was followed by presentation of a test stimulus (i.e., stimulus transitions in Figs. 4B-C). This approach is a standard method for establishing categorical perception in human infants30 and non-human animals7,26.
FIGURE 4

The perceptual boundary in swamp sparrows’ categorization of note duration was predicted by the categorical boundary evident in auditory responses of HVCX neurons

(a) The distribution of Category I and Category VI11 note durations in the songs of swamp sparrows from Pennsylvania (PA, thick black line) differed from the distribution in the songs of a previously studied population of swamp sparrows from New York7 (NY, thin gray line; p < 0.001, Kolmogorov-Smirnov 2-sample test, N = 129 notes, 62 songs from 35 PA birds; N = 52 notes, 29 songs from 12 NY birds, 3 ms binsize), suggesting that the categorical boundary in each population (dotted lines) may be learned through population-specific auditory and vocal experience. (b) Stimuli in our behavioral tests contained note transitions that either: 1) crossed no putative categorical boundary (4 to 8 ms), or 2) crossed the NY perceptual boundary but not the PA neural boundary (8 to 16 ms), or 3) crossed the PA neural boundary but not the NY perceptual boundary (16 to 32 ms). (c) Behavioral testing revealed that PA swamp sparrows (top, filled bars) perceived strong differences when the transition in note duration spanned the PA neural boundary detected in HVCX neurons (stimulus set 3, p < 0.001, Kruskal-Wallis test; p < 0.05 for stimulus set 3 vs. 1 and 3 vs. 2, Tukey’s HSD) but perceived little or no difference when the transition spanned the perceptual boundary detected in NY birds (stimulus set 2; p > 0.05, stimulus set 1 vs. 2, Tukey’s HSD) or spanned no putative boundary (stimulus set 1). Comparison of the behavioral data obtained from PA birds (top, filled bars, mean ± SE) and from NY birds7 (bottom, open bars, mean ± SE) makes clear the population-specific differences in categorical perception of note duration (NY data adapted from Nelson and Marler7, asterisk indicates p < 0.05, Kruskal-Wallis ANOVA, responses of individual PA birds shown in Supp. Fig. 6).

Swamp sparrows are territorial, and the introduction of novel song stimuli on their territory evokes robust aggressive displays, such as wing waves7,31. We quantified this territorial behavior in a habituation/dishabituation paradigm to determine whether the perceptual boundary of Pennsylvania birds more closely approximated the perceptual boundary measured behaviorally for New York birds7 or the neural response boundary we measured for Pennsylvania birds. We found that Pennsylvania birds responded as though they did not detect changes in note duration that spanned the perceptual boundary reported for New York birds (Fig. 4B-C, p > 0.05, stimulus transition 2 vs. 1; Tukey’s HSD). However, Pennsylvania birds clearly detected changes in note duration that spanned the neural response boundary detected in HVCX neurons, indicating that the changes in duration also spanned a categorical perceptual boundary (p < 0.001, Kruskal-Wallis test; p < 0.05, stimulus transition 3 vs. 1; p < 0.05, stimulus transition 3 vs. 2; Tukey’s HSD). Thus, categorical perception of note duration in swamp sparrows as measured in the field by behavioral testing is accurately predicted by the auditory response properties of HVCX neurons measured in the laboratory. Although it is extremely challenging to conduct behavioral tests that adequately capture the responsiveness of a wild territorial male songbird in a laboratory setting while simultaneously collecting neurophysiological data, we were able to obtain both behavioral and neurophysiological data at different times from one bird (Supp. Fig. 4, see also Figs. 2B-C (left and middle columns), Fig. 3A-B and Supp. Fig. 5). Although these findings are limited to a single individual, they indicate a direct within-individual correspondence between neurophysiological data and behavioral data related to those we measured in a field setting.

Categorical auditory responses are not influenced by frequency modulation or bandwidth

Although prior behavioral work implicated note duration as the primary salient acoustic feature for categorical perception7, the stimuli used in that study and the initial set of stimuli used here varied note duration without controlling for changes in other covariant features of the manipulated note, namely the rate of frequency modulation (FM) or frequency bandwidth (BW). Thus it remains to be established whether note duration, rather than FM or bandwidth, is the salient feature underlying categorical perception of Category I and Category VI notes by swamp sparrows and the categorical responses of their HVC neurons. Although the larger stimulus set required to distinguish between these alternatives is impractical to employ in field studies of wild sparrows, such measurements could be made for the responses of single neurons. Therefore, we tested a subset of HVCX neurons (N = 10 cells, 2 birds; different subset than cells used in high-resolution testing of boundary location) with variants of the primary song type containing computer-generated replacement notes. The use of these synthetic notes allowed us to systematically change note duration while holding constant either the note’s FM or BW (Figs. 5A-C). Syllables containing synthetic notes that closely replicated the features of the natural note they replaced were just as efficacious as the natural syllable at eliciting responses from HVCX neurons (p = 0.47; Wilcoxon signed rank test; Fig. 5A). Furthermore, syllables containing synthetic notes with durations similar to those of the natural notes they replaced, but with different FM or bandwidths, also were highly effective at eliciting responses from HVCX neurons (FM: p = 0.13; BW: p = 0.43; Wilcoxon signed rank test; Figs. 5B-C). Thus, these data indicate that neither FM nor BW FM nor BW plays a primary role in influencing our results, and that note duration, the song feature previously implicated in affecting categorical perception of these note types7, is the salient song feature in driving categorical responses in HVCX neurons.
FIGURE 5

Categorical responses were evoked by changes in specific features of individual notes in the syllable

(a-c) A subset of 10 categorically responsive HVCX neurons (2 birds) was further tested to probe the acoustic stimulus features that directed categorical responsiveness. (a) In all cells, computer-generated duplicates of natural replacement notes were sufficient to evoke a categorical response (p = 0.47; Wilcoxon signed rank test), establishing the validity of further testing using synthetic stimuli. Synthetic stimuli in which note duration was varied while either (b) the rate of frequency modulation (FM; p = 0.13) or (c) the frequency bandwidth (BW; p = 0.43; Wilcoxon signed rank test) was held constant evoked responses like those evoked by natural replacement notes, implicating note duration as the salient song feature driving categorical responses of HVCX neurons.

Categorical responses are evident in neurons displaying a precise sensorimotor correspondence

In a prior study in which we recorded both auditory and singing-related activity in HVCX neurons of swamp sparrows, we established that individual cells display a precise sensorimotor correspondence20. Because neurons that display sensorimotor “mirroring” are hypothesized to mediate perception23, we investigated whether categorical responsiveness was evident in these same cells for which this sensorimotor correspondence had been established. Categorical auditory responsiveness was observed in all HVCX neurons for which we also recorded singing-related activity (N = 5 cells, 2 birds). Comparison of auditory and singing-related activity revealed that all of these neurons displayed a precise sensorimotor correspondence20 (e.g., Supp Fig. 5). These results suggest that neurons displaying a form of sensorimotor mirroring convey perceptual information about song.

DISCUSSION

This study provides the first evidence of neurons that encode perceptual information about a phonological feature of learned vocal behavior, specifically information about a categorical perceptual boundary. Furthermore, our neural results accurately predicted a previously unknown geographic dialectical difference in this perceptual boundary, one that we subsequently confirmed using behavioral tests with wild birds. Variation in this perceptual boundary across swamp sparrow populations strongly suggests that both categorical perception and categorical neural responses in sparrows are affected by experience. Finally, by identifying a cell type that expresses categorical activity, we provide insights into the neural circuitry for categorical perception of note duration. Categorical perception has been shown to play a prominent role in a wide variety of communication systems3-9, including human speech3,4. Although categorical processing of both human and non-human vocalizations has been demonstrated in a variety of animals3,4,6,8, neural correlates of this phenomenon await identification. Neurophysiological studies in other modalities, however, have identified that neurons can reliably encode information about perceptual categories. For example, neurons in the monkey inferotemporal cortex respond in a categorical manner to facial features32, an important component of social signaling in primates, and brain imaging studies in humans have detected regions that show categorical sensitivity to differences in facial expression33. These studies strengthen the idea that categorical perception is mediated by categorically responsive neurons, and that the activity of those cells could play a role in communication. The present results identify neurons that respond categorically to learned vocal signals, extending the analysis of the neural basis of perceptual categorization to another major facet of communication with special relevance to human speech. To our initial surprise, we found that the categorical response boundary for note duration measured from neurons differed markedly from the previously reported categorical perceptual boundary for this system using behavioral methods7. We found, however, that the response boundary we determined from physiological recordings of HVCX cells accurately predicted a previously unidentified population difference in this perceptual boundary, indicating that the perceptual boundary varies across geographically distinct swamp sparrow populations and consistent with the view that, as is the case for human speech3,4, these boundaries can be influenced by early experience. The fact that bird songs vary geographically as a consequence of learning is widely documented11,34, but our data are the first to suggest an effect of learning on the categorical perception of a natural signal other than speech. The observations that human infants across cultures display an initially similar ability to categorize speech sounds35 and that categorical boundaries of human speech can be detected by mammals that do not learn their vocalizations6 suggest that categorization of vocal signals may exploit innate neural mechanisms. However, the role of social experience in shaping human speech perception, including the perception of categorical boundaries, is also well documented3,4, indicating that the underlying neural mechanisms must also be pliable. Future studies in songbirds will be especially fruitful for addressing how acoustic and perhaps social experience shapes the development of categorical perception and categorically responsive neurons. The present study localized categorical responses to HVCX projection neurons, which innervate a striatal pathway that lesion studies implicate in song perception as well as song learning15,16,18,22 and may be regarded as analogous to mammalian corticostriatal neurons. Our finding that the auditory responses of these projection neurons are closely linked to song perception provides a physiological basis for understanding how lesions to this pathway could interfere with song recognition. Furthermore, the present findings show that categorical responses are exhibited by HVCX neurons that display precise sensorimotor mirroring20, providing a link between mirror neurons and perception. In the songbird, categorically responsive sensorimotor neurons could enable auditory perception to directly guide subsequent behavior, as evident in countersinging in response to hearing a neighbor’s song20. Intriguingly, neurons that represent learned visual categories have been detected in regions of monkey parietal cortex implicated in motor planning and decision-making36. The expression of categorical responses in sensorimotor neurons in both songbirds and primates could point to an efficient means through which the nervous system determines when variations in a sensory stimulus warrant similar or different behavioral responses. Establishing that categorical responses are expressed by striatal-projecting HVCX neurons, but not by interneurons, may also provide a clue as to how and where encoding of perceptual attributes occurs. Prior studies have shown that although HVCX neurons and HVC interneurons receive shared excitatory auditory input37,38, the auditory selectivity of interneurons more closely parallels the activity of auditory afferents to HVC37, and the highly selective auditory responses of HVCX neurons require inhibitory sculpting through interneurons39. The presence of categorical auditory responses in HVCX neurons and the absence of such responses in HVC interneurons suggest that categorical activity emerges through the local circuitry of HVC40 and the processes of synaptic integration in HVCX cells. Future studies using intracellular recordings could test this idea by determining whether HVCX neurons receive broadly responsive excitatory input from which categorical responses are sculpted by local inhibition. Precise mapping of the neuronal response boundary revealed that the action potential discharge of HVCX cells is exquisitely sensitive to changes in note duration on the millisecond timescale. Sensitivity to fine temporal features of song has been described for HVC neurons in other songbirds41, and intracellular and extracellular recordings from putative HVCX neurons reveal that their song-evoked auditory responses are strongly dependent on auditory context, such as sequences of notes or syllables42-44. This context sensitivity involves integration over hundreds of milliseconds42 and is thought to depend at least in part on circuit mechanisms local to HVC38,39. Although the present study cannot resolve whether categorical responsiveness to note duration arises in HVC, it does suggest that temporal aspects of song ranging from milliseconds to hundreds of milliseconds can be encoded by single neurons. The sensitivity of individual neurons to features over multiple timescales may constitute a strategy for optimizing encoding of complex stimuli that can vary in both local structure and global sequence. SUPPLEMENTARY FIGURE 1 A small minority of HVCX cells expressed responses to changes in note duration that were not categorical in nature. In 3 of 22 HVCX cells tested (2 birds), the auditory response did not fulfill the criteria to establish a response as categorical (see Methods). These data reveal no evidence of neurons selectively tuned to intermediate note durations, indicating that the population of HVCX neurons does not represent note duration continuously, further supporting our claim that HVCX neurons encode note duration in categorical manner. SUPPLEMENTARY FIGURE 2 A linear model of auditory responses of HVCX neurons to notes of different duration does not account for the observed dataset. We compared the relations between note duration and strength of response for HVCX cells in our dataset against the null hypothesis that those cells represented note duration in a linear manner (i.e., HVCX cells expressed a linear relation rather than the step-like transition characteristic of categorical activity). In this noisy linear model, the neural response is expected to be minimal at the shortest note duration we tested (normalized response = 0 at note duration 4 ms) and is expected to be maximal at the longest note duration we tested (normalized response = 1 at note duration 31 ms; stated in equation form: Y = mX + b, where m = 0.037 and b = -0.148). Random variance is included in the model to simulate local transitions with slopes much steeper than the mean value or possibly negative slopes (see Supplementary Methods). Because our dataset included both positive and negative slopes at the categorical boundary (e.g., Fig. 3D, Supp. Fig. 3), we used the absolute value of the slope in the tuning curve of each HVCX neuron to compare the steepest slope of each transition in our dataset against the results of the model (shown in panel a). Artificial data created using this model were compared against the dataset of 10 HVCX neurons for which we had a high-resolution estimate of the categorical boundary (Supp. Fig. 3). Three epochs of note duration were considered: 1) within the putative short-note category (X <= 14 ms), 2) within the putative long-note category (X >= 27 ms), and 3) values around the putative boundary (14 < X < 27 ms). (a) Within the putative boundary region, the slopes of the steepest transition for cells in our dataset (filled bar, mean ± SE) were significantly steeper than the slope of the model (open bar, p = 0.002, paired t-tests). (b) Within each category, the slopes of responses in our dataset (filled bar) were significantly different than the slope of the model (open bar, p = 0.02) and were indistinguishable from a slope of zero (p = 0.94, no absolute values used in within-category comparison of model and actual values). Together, these data indicate that a linear model cannot account for our observations and that, consistent with a categorical representation, the population activity of HVCX neurons in response to changes in note duration expresses a steep transition between two regions with slopes indistinguishable from zero. SUPPLEMENTARY FIGURE 3 The categorical boundary estimated from a subset of HVCX neurons tested using high-resolution stimuli corroborates the boundary estimated using the full dataset. A subset of 10 HVCX neurons (2 birds, all 10 cells shown here) was tested to probe the categorical boundary at a high resolution, revealing an estimated categorical boundary (filled triangle, 20 ± 4 ms) very similar to that estimated using the full dataset (21 ± 4 ms; p = 0.47, unpaired t-test). Categorical responses were observed in HVCX cells regardless of whether the duration of note that was replaced was naturally short (thick lines) or naturally long (thin lines). SUPPLEMENTARY FIGURE 4 Behavioral testing of one swamp sparrow for which neural data were also collected revealed a correspondence between the perceptual and neurophysiological boundaries. Song stimuli were presented as described in the Methods for experiments performed in the field, and the bird’s behavior was observed using a small camera placed inside the neurophysiological recording chamber. The bird was conditioned to song stimuli using the song of another swamp sparrow, and the number of aggressive displays (crest raises46 or wing waves7 quantified as described in the Methods) was strongly habituated at time 0. Changing the stimulus to the bird’s own song (12 minutes) evoked a dishabituation of response, indicating that the bird perceived the new song as different than the song to which it had been conditioned. In subsequent song changes in this habituation/dishabituation paradigm (see Methods), this bird responded as if he perceived differences between stimuli that crossed the neurophysiological boundary (i.e., changing note duration from 7 to 31 ms at 98 minutes and from 27 to 5 ms at 218 minutes) but not between stimuli that did not cross the boundary (i.e., changing from 31 to 27 ms at 192 minutes). The bird from which we obtained these behavioral results is the same bird from which we obtained the neural in Figs 2B-C (left and middle columns), Fig 3A-B and Supp. Fig. 5. SUPPLEMENTARY FIGURE 5 Categorical auditory responses were observed in HVCX cells that also expressed a precise sensorimotor correspondence20. (a) The auditory response of this HVCX neuron was selective for the primary song type (N = 5 cells, 2 birds; data from the same cell shown in all panels; top: auditory response PSTH, 10 ms binsize; bottom: stimulus syllable spectrogram; left: primary song type; center: reverse playback of the primary song type; right: another song type in the bird’s vocal repertoire). (b) Manipulation of the duration of note C in each syllable of the primary song type resulted in a categorical auditory response (data as in Fig. 3A). (c) This cell also expressed a precise correspondence in the timing of action potentials generated when the bird sang the primary song type (top) and during the auditory response (middle) to the primary song type (spectrogram, bottom). The presence of a neural correlate of the animal’s perception in auditory-vocal “mirror neurons20” strengthens the idea that cells expressing an auditory-vocal correspondence may facilitate perception of the signals used in vocal communication. SUPPLEMENTARY FIGURE 6 Responses of individual birds in behavioral testing of the perceptual boundary. Swamp sparrows in the Pennsylvania population perceived strong differences when the transition in note duration spanned the neurophysiological boundary detected in HVCX neurons of Pennsylvania birds (right, N = 8 birds, statistics as reported in legend of Fig. 4C) but perceived little or no difference when the transition spanned the perceptual boundary detected in New York birds (middle, N = 8 birds) or spanned no putative boundary (control, left, N = 8 birds). The y-axis indicates the difference between the number of territorial displays evoked in the first block of testing using the dishabituation test stimulus and the number of displays evoked in the final block of testing using the habituation stimulus (thus negative values are possible, see Methods for details). Small values are consistent with the bird perceiving little or no difference between the habituation and dishabituation stimuli, and large positive values are consistent with the bird perceiving the two stimuli as different. These data are summarized in Fig. 4C.
  40 in total

1.  Different subthreshold mechanisms underlie song selectivity in identified HVc neurons of the zebra finch.

Authors:  R Mooney
Journal:  J Neurosci       Date:  2000-07-15       Impact factor: 6.167

Review 2.  Speech perception.

Authors:  Randy L Diehl; Andrew J Lotto; Lori L Holt
Journal:  Annu Rev Psychol       Date:  2004       Impact factor: 24.137

3.  Neuronal correlates of face identification in the monkey anterior temporal cortical areas.

Authors:  Satoshi Eifuku; Wania C De Souza; Ryoi Tamura; Hisao Nishijo; Taketoshi Ono
Journal:  J Neurophysiol       Date:  2004-01       Impact factor: 2.714

4.  CULTURALLY TRANSMITTED PATTERNS OF VOCAL BEHAVIOR IN SPARROWS.

Authors:  P MARLER; M TAMURA
Journal:  Science       Date:  1964-12-11       Impact factor: 47.728

5.  The discrimination of speech sounds within and across phoneme boundaries.

Authors:  A M LIBERMAN; K S HARRIS; H S HOFFMAN; B C GRIFFITH
Journal:  J Exp Psychol       Date:  1957-11

6.  Precise auditory-vocal mirroring in neurons for learned vocal communication.

Authors:  J F Prather; S Peters; S Nowicki; R Mooney
Journal:  Nature       Date:  2008-01-17       Impact factor: 49.962

7.  Categorical perception of a natural, multivariate signal: mating call recognition in túngara frogs.

Authors:  A T Baugh; K L Akre; M J Ryan
Journal:  Proc Natl Acad Sci U S A       Date:  2008-06-24       Impact factor: 11.205

8.  The evolution of language.

Authors:  M A Nowak; D C Krakauer
Journal:  Proc Natl Acad Sci U S A       Date:  1999-07-06       Impact factor: 11.205

9.  Theoretical notes. Motor theory of speech perception: a reply to Lane's critical review.

Authors:  M Studdert-Kennedy; A M Liberman; K S Harris; F S Cooper
Journal:  Psychol Rev       Date:  1970-05       Impact factor: 8.934

10.  Acoustic parameters underlying the responses of song-specific neurons in the white-crowned sparrow.

Authors:  D Margoliash
Journal:  J Neurosci       Date:  1983-05       Impact factor: 6.167

View more
  62 in total

1.  Reciprocal inhibition of inhibition: a circuit motif for flexible categorization in stimulus selection.

Authors:  Shreesh P Mysore; Eric I Knudsen
Journal:  Neuron       Date:  2012-01-12       Impact factor: 17.173

2.  Adaptive switches in midbrain circuits.

Authors:  Tatyana O Sharpee
Journal:  Neuron       Date:  2012-01-12       Impact factor: 17.173

3.  Differential influence of frequency, timing, and intensity cues in a complex acoustic categorization task.

Authors:  Katherine I Nagel; Helen M McLendon; Allison J Doupe
Journal:  J Neurophysiol       Date:  2010-07-07       Impact factor: 2.714

4.  Temporal and rate code analysis of responses to low-frequency components in the bird's own song by song system neurons.

Authors:  Makoto Fukushima; Peter L Rauske; Daniel Margoliash
Journal:  J Comp Physiol A Neuroethol Sens Neural Behav Physiol       Date:  2015-08-30       Impact factor: 1.836

5.  Neural encoding and integration of learned probabilistic sequences in avian sensory-motor circuitry.

Authors:  Kristofer E Bouchard; Michael S Brainard
Journal:  J Neurosci       Date:  2013-11-06       Impact factor: 6.167

6.  Linear and nonlinear auditory response properties of interneurons in a high-order avian vocal motor nucleus during wakefulness.

Authors:  Jonathan N Raksin; Christopher M Glaze; Sarah Smith; Marc F Schmidt
Journal:  J Neurophysiol       Date:  2011-12-28       Impact factor: 2.714

7.  Neuron-specific cholinergic modulation of a forebrain song control nucleus.

Authors:  Stephen D Shea; Henner Koch; Daniel Baleckaitis; Jan-Marino Ramirez; Daniel Margoliash
Journal:  J Neurophysiol       Date:  2009-11-25       Impact factor: 2.714

8.  Imaging auditory representations of song and syllables in populations of sensorimotor neurons essential to vocal communication.

Authors:  Wendy Y X Peh; Todd F Roberts; Richard Mooney
Journal:  J Neurosci       Date:  2015-04-08       Impact factor: 6.167

Review 9.  Auditory-vocal mirroring in songbirds.

Authors:  Richard Mooney
Journal:  Philos Trans R Soc Lond B Biol Sci       Date:  2014-04-28       Impact factor: 6.237

10.  Perceptual and neuronal boundary learned from higher-order stimulus probabilities.

Authors:  Hania Köver; Kirt Gill; Yi-Ting L Tseng; Shaowen Bao
Journal:  J Neurosci       Date:  2013-02-20       Impact factor: 6.167

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.