Literature DB >> 35200259

Objective Detection of the Speech Frequency Following Response (sFFR): A Comparison of Two Methods.

Fan-Yin Cheng1, Spencer Smith1.   

Abstract

Speech frequency following responses (sFFRs) are increasingly used in translational auditory research. Statistically-based automated sFFR detection could aid response identification and provide a basis for stopping rules when recording responses in clinical and/or research applications. In this brief report, sFFRs were measured from 18 normal hearing adult listeners in quiet and speech-shaped noise. Two statistically-based automated response detection methods, the F-test and Hotelling's T2 (HT2) test, were compared based on detection accuracy and test time. Similar detection accuracy across statistical tests and conditions was observed, although the HT2 test time was less variable. These findings suggest that automated sFFR detection is robust for responses recorded in quiet and speech-shaped noise using either the F-test or HT2 test. Future studies evaluating test performance with different stimuli and maskers are warranted to determine if the interchangeability of test performance extends to these conditions.

Entities:  

Keywords:  F-test; Hotelling’s T2 test; auditory electrophysiology; frequency following response (FFR); objective detection

Year:  2022        PMID: 35200259      PMCID: PMC8869319          DOI: 10.3390/audiolres12010010

Source DB:  PubMed          Journal:  Audiol Res        ISSN: 2039-4330


1. Introduction

Neural encoding of speech features can be noninvasively assessed using the speech frequency-following response (sFFR). Over the past two decades, the sFFR has been used in a variety of research studies examining: effects of auditory expertise and deprivation on speech feature encoding [1,2,3,4,5,6,7]; neural representation of speech in adverse listening conditions [8,9,10]; objective hearing aid fitting [11]; bimodal hearing assessment [12,13,14]; and auditory development [15], to name a few examples. Despite the broad use of this technique, few studies, e.g., [11,16,17,18] have examined statistically-based automated response detection of sFFRs. This consideration is important in both clinical and research domains, as automated detection would aid in objective response identification and would provide a statistical basis for stopping rules when recording sFFRs. This brief report examined accuracy and test time required to objectively detect sFFRs in quiet and noise using F- and Hotelling’s T2 (HT2) tests in the frequency domain (see [19,20] for detailed reviews of statistical techniques for objective response detection). The F-test calculates an F ratio of signal power at a single frequency of interest (e.g., the fundamental frequency, F0) to the mean power of adjacent frequency bins in which no response components are expected. This approach is particularly well-suited for detecting auditory steady state responses in objective audiometry, e.g., [21,22]. The HT2 test is a multivariate analysis in which mean differences of measured and hypothesized values are tested for significance using multiple sFFR features (e.g., F0 and harmonics). The HT2 test is thus ideal for automated sFFR detection, as it analyzes information at multiple frequencies to determine if a response is present [6,23,24]. Further, this test may be more powerful in conditions where individual components are degraded (e.g., when the sFFR is measured in background noise). The purpose of this study was to objectively detect sFFR responses using both F- and HT2 tests and to compare detection accuracies and test times for both approaches. We hypothesized that the HT2 test would have better detection accuracy and shorter time-to-detect because it uses more information contained in the sFFR than the F-test.

2. Materials and Methods

2.1. Participants

The University of Texas at Austin approved the methods described in this study. Participants were 18 adults (10 females, age range 20–32). None reported a history of otopathology, neuropathology, or significant noise exposure. All passed otoscopy, tympanometry, and pure tone audiometry screening.

2.2. Stimulus and Recording Procedures

sFFRs were elicited diotically using a 170-ms/da/speech token (F0 = 100 Hz). Stimuli were presented in alternating polarity at 70 dB SPL through electromagnetically shielded ER-3C insert earphones at a rate of 4.3 Hz [13]. In the noise condition, continuous speech-shaped noise was presented diotically at 65 dB SPL. sFFRs were recorded with a Cz-C7 single channel montage for 2000 stimulus repetitions using a Neuroscan SynAmps2 system (Compumedics Neuroscan; Charlotte, NC, USA). Responses were artifact rejected at ±30 µV and filtered from 70–2000 Hz using Curry 8 software. Further analyses were conducted offline in MATLAB (The Mathworks, Natick, MA, USA).

2.3. Automatic Response Detection

F-statistic and HT2 analyses were performed on frequency-domain data extracted from the 60–180 ms epoch (i.e., steady-state vowel portion) of the “added” sFFR, which biases the responses to reflect speech envelope features (see [25]). The F-statistic approach was implemented on the spectrum of the cumulative average sFFR waveform each time a sweep was added to the average, as described by Picton and colleagues (2003) [20]: where F0 is the fundamental frequency (averaged across 99–101 Hz bins), j is the frequency of an adjacent noise bin, B is the number of noise bins, and x and y are amplitude and phase components of a specified frequency bin. In the present experiment, B = 18, and the upper and lower samples of noise began at F0 ± 5 Hz, respectively. The respective degrees of freedom were 6 and 36 for the numerator and denominator [19,20]. The HT2 test was also implemented as described by Picton and colleagues (2003) using the equation [20]: where is the number of sweeps, is a vector of measured dependent variable means with length , μ0 is a vector of hypothesized means with length , and S−1 is an inverse covariance matrix of . In the present experiment, both amplitude and phase components from F0, H2, H3, and H4 were used as dependent variables in the HT2 test; thus, . Because the data are circular, the vector of hypothesized means (i.e., the expected outcome if a response were not present) consists of zeros. T2 was converted to an F-statistic using the equation: where and The F-statistic produced by each analysis was used to find the statistical probability that an sFFR response was present as a function of sweep count; a schematized overview of this approach is provided in Figure 1. Detection rate was quantified as the number of sFFRs detected by each statistical test in each listening condition. These proportions were compared using a Chi-square test. We calculated a “time-to-detect” measure as the test time, in seconds, at which detection probability was ≥99% and remained above this threshold for at least the following 25 sweeps. A repeated-measures analysis of variance (RMANOVA) was used to assess the effects of two factors (quiet vs. noise; F-test vs. HT2 test) on time-to-detect measurements.
Figure 1

Automatic Response Detection Approach. (A) Average sFFR waveforms collected in quiet (left) and their corresponding spectra (right) are shown as a function of sweep count for a single participant (black-to-red color transition indicates an increase in cumulative sweeps contributing to the average from 200–1000). Frequency bins of interest for the F-statistic (F0 only) and HT2 test (F0-H4) are labeled on the final spectrum. (B) Amplitude and phase components are extracted from each frequency bin of interest and used as dependent variables for statistical tests. Note that each vector maintains the color-coding denoting sweep count from panel A. Note also that the F-test only uses magnitude and phase from the F0 vector, whereas HT2 uses magnitude and phase components from all vectors shown. (C) Cumulative probability functions are plotted as a function of sweep count for each test with 99% confidence denoted by the gray dotted line. In this single-subject example, F- and HT2 tests perform similarly, as an sFFR is detected with >99% confidence between 380–400 sweeps with each test.

3. Results

3.1. Detection Rate

Table 1 lists the percentage of participants from whom sFFR responses could be detected for each condition. Detection rate was similar across conditions [X(1) = 0.057, p = 0.811], demonstrating that comparable objective sFFR identification was achieved with either statistical approach in quiet or noise.
Table 1

Percentage of response detection across conditions.

QuietNoise
F-statistic94% (17/18)100% (18/18)
HT2100% (18/18)94% (17/18)

3.2. Time-to-Detect

Participants in which sFFR responses were not detected in any of the four conditions were removed from the time-to-detect analysis, leaving 16 participants. There was no main effect of statistical test (F-statistic vs. HT2 test; F(1,15) = 1.475, p = 0.243, 2 = 0.09) or noise condition (quiet vs. noise; F(1,15) = 0.172, p = 0.684, 2 = 0.011), nor was there an interaction between factors (F(1,15) = 3.893, p = 0.067, 2 = 0.206). As shown in Figure 2, there was a trend for shorter averaged time-to-detect estimate with the HT2 test than with F-statistic in quiet, and there is less variability between participants when estimated with HT2 than F-statistic in both quiet and noise conditions.
Figure 2

Time-to-detect by statistical test and condition. Each condition is coded with a different color for clarity. The middle line of the box represents the median, and the x represents the mean. The bottom line of the box represents the median of the 1st quartile, and the top line of the box represents the median of the 3rd quartile. The whiskers (vertical lines) extend to minimum and maximum values excluding outliers. Points that exceed 1.5 times of interquartile range are considered outliers. Outliers were not excluded from analyses because they represented the upper limit of test times for our sample.

4. Discussion

We predicted that the HT2 test, which utilizes more information contained in the sFFR spectrum than the F-test, would have superior detection performance than its counterpart for quiet and noise conditions. While there was a tendency for time-to-detect to be less variable when calculated from the HT2 test, we found no difference between statistical test performance in quiet or noise on our time-to-detect measure. These findings are broadly consistent with Easwar and colleagues (2020), who found that area under the curve metrics were similar between F- and HT2 tests. However, the same study indicated that the HT2 test had higher sensitivity than the F-test. Other reports have also shown time-to-detect advantages of the HT2 test for sFFRs measured in quiet, e.g., [6]. These discrepancies may be due to differences in sFFR stimuli between studies. For example, Vanheusden and colleagues (2019) used naturalistic monosyllabic words with a dynamic F0 and harmonics to evoke sFFRs, whereas the present study used a synthesized/da/stimulus with a stable F0 throughout the entire steady-state portion [6]. This likely resulted in robust F0 representation in the sFFR spectra; because this robust component was used in both F- and HT2 tests, performance of each test may have been more similar than predicted. We also expected that sFFR detection would be superior using the HT2 test in the noise condition. This prediction was based on the reasoning that noise is deleterious to sFFR spectral peaks, including the F0. Including more spectral peaks in the statistical test would therefore be expected to improve its ability to detect when a “true” sFFR is present. However, because we used speech-shaped noise as a masker, the sFFR components most impacted by the masker were the harmonics H2-H4 and not the F0. This is because the F0 component of the sFFR is dominated by neural phase-locking from higher frequency channels in which multiple harmonics overlap [18] and may not be appreciably affected by speech-shaped noise, which is low frequency dominant. This again would have made F- and HT2 test inputs more similar than expected and may have made their time-to-detect results more comparable in noise. To our knowledge, this is the first report to compare such sFFR detection methods in noise, and future studies with more realistic speech tokens and maskers are warranted. This brief report adds to the growing literature demonstrating that statistically-based methods for automatic sFFR detection are robust, at least in listeners with normal hearing. Future parametric work is needed to explore the performance of these tests in listeners with hearing loss e.g., [11] or other audiologic disorders as well as the aging e.g., [26]. While objective detection methods are commonly used in commercial devices in which auditory steady-state responses can be obtained (e.g., Interacoustics, Intelligent Hearing Systems, GSI-Audera, Bio-logic), fewer proprietarry options exist for sFFR measurement and analyis and many clinician-scientists conduct analyses outside of commercially-availible systems (e.g., in MATLAB). This brief report and similar articles provide an evidence base for objective measures that can be employed in commercially available software for sFFR detection and analysis.
  23 in total

1.  Multiple-ASSR thresholds in infants and young children with hearing loss.

Authors:  Anna Van Maanen; David R Stapells
Journal:  J Am Acad Audiol       Date:  2010-09       Impact factor: 1.664

2.  Aging degrades the neural encoding of simple and complex sounds in the human brainstem.

Authors:  Christopher G Clinard; Kelly L Tremblay
Journal:  J Am Acad Audiol       Date:  2013 Jul-Aug       Impact factor: 1.664

3.  A comparison of t test, F test, and coherence methods of detecting steady-state auditory-evoked potentials, distortion-product otoacoustic emissions, or other sinusoids.

Authors:  R A Dobie; M J Wilson
Journal:  J Acoust Soc Am       Date:  1996-10       Impact factor: 1.840

4.  Sensitivity of envelope following responses to vowel polarity.

Authors:  Vijayalakshmi Easwar; Laura Beamish; Steven Aiken; Jong Min Choi; Susan Scollie; David Purcell
Journal:  Hear Res       Date:  2014-12-09       Impact factor: 3.208

5.  Establishing auditory steady-state response thresholds to narrow band CE-chirps(®) in full-term neonates.

Authors:  Gabriela Ribeiro Ivo Rodrigues; Doris Ruthy Lewis
Journal:  Int J Pediatr Otorhinolaryngol       Date:  2013-11-22       Impact factor: 1.675

6.  Objective measures for detecting the auditory brainstem response: comparisons of specificity, sensitivity and detection time.

Authors:  M A Chesnaye; S L Bell; J M Harte; D M Simpson
Journal:  Int J Audiol       Date:  2018-03-14       Impact factor: 2.117

7.  Improved Detection of Vowel Envelope Frequency Following Responses Using Hotelling's T2 Analysis.

Authors:  Frederique J Vanheusden; Steven L Bell; Michael A Chesnaye; David M Simpson
Journal:  Ear Hear       Date:  2019 Jan/Feb       Impact factor: 3.570

Review 8.  Auditory brain stem response to complex sounds: a tutorial.

Authors:  Erika Skoe; Nina Kraus
Journal:  Ear Hear       Date:  2010-06       Impact factor: 3.570

Review 9.  Human auditory steady-state responses.

Authors:  Terence W Picton; M Sasha John; Andrew Dimitrijevic; David Purcell
Journal:  Int J Audiol       Date:  2003-06       Impact factor: 2.117

10.  Relationships between click auditory brainstem response and speech frequency following response with development in infants born preterm.

Authors:  Angela M Madrid; Kerry A Walker; Spencer B Smith; Linda J Hood; Beth A Prieve
Journal:  Hear Res       Date:  2021-05-23       Impact factor: 3.672

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.