Abstract
Some people using hearing aids have difficulty discriminating between sounds even though the sounds are audible. As such, cochlear implants may provide greater benefits for speech perception. One method to identify people with auditory discrimination deficits is to measure discrimination thresholds using spectral ripple noise (SRN). Previous studies have shown that behavioral discrimination of SRN was associated with speech perception, and behavioral discrimination was also related to cortical responses to acoustic change or ACCs. We hypothesized that cortical ACCs could be directly related to speech perception. In this study, we investigated the relationship between subjective speech perception and objective ACC responses measured using SRNs. We tested 13 normal-hearing and 10 hearing-impaired adults using hearing aids. Our results showed that behavioral SRN discrimination was correlated with speech perception in quiet and in noise. Furthermore, cortical ACC responses to phase changes in the SRN were significantly correlated with speech perception. Audibility was a major predictor of discrimination and speech perception, but direct measures of auditory discrimination could contribute information about a listener’s sensitivity to acoustic cues that underpin speech perception. The findings lend support for potential application of measuring ACC responses to SRNs for identifying people who may benefit from cochlear implants.
Similar content being viewed by others
Introduction
People with hearing loss experience difficulties in understanding speech, partly arising from reduced audibility and partly from their reduced ability to discriminate sounds that are audible. Hearing aids amplify sounds so that they become audible to the user but cannot enhance the ability of the user to analyze spectro-temporal variations in the speech wave to discriminate sounds. As hearing loss increases, the ability to perceive the spectral variations decreases1,2, thereby resulting in poor speech perception3. When hearing aids provide limited benefits to users, they can be considered candidates for cochlear implantation.
Current evaluations of the benefit of hearing aids typically relies on using speech perception tests to measure the accuracy with which words or sentences can be repeated. However, performance in language-based hearing tests relies not only on extraction of acoustic cues of speech, but also semantic (meaning), syntactic (grammar) and other circumstantial (speaker identity, subject matter, listening environment, etc.) cues. When the goal is to assess the listener’s ability to discriminate sounds, it has been proposed that non-linguistic stimuli that are modulated in the frequency domain to mimic the spectral variations found in speech are valuable for this purpose4,5,6.
The non-linguistic stimuli, commonly referred to as spectral ripple noise (SRN), can be varied in terms of number of modulations per octave and depth of modulation7. Davies-Venn et al. demonstrated that the ability to perceive SRN modulation depth in noise (Spectral Modulation Detection: SMD) was significantly correlated with speech perception in quiet and in noise4. In a similar vein, Litvak et al. found a correlation between SMD and consonant and vowel perception8. These studies that showed a significant correlation between behavioral discrimination measured using SRN and speech perception assessed using language-based material lend support to the use of SRN to evaluate auditory spectral processing abilities.
However, measurement of behavioral discrimination using SRN cannot be applied to people who are not able to provide behavioral responses reliably, including but not limited to those who may have a language barrier or cognitive limitations or use non-verbal communication modes or too young. Therefore, it is important to examine whether electrophysiological methods for measuring auditory discrimination using the SRN is viable and valid. Won et al. reported a significant correlation between behavioral detection of a phase inversion in an SRN stimulus and electrophysiological measure of cortical auditory evoked potential (CAEP) to sound changes or the acoustic change complex (ACC) to the SRN stimulus in three normal-hearing adults9. Higher amplitude of the ACC was significantly correlated with better behavioral discrimination ability. The ACC can be reliably recorded with good test–retest reliability not only from listeners with normal hearing but also from those with hearing loss, using hearing aids or cochlear implants (see review by Kim et al., 2015)10. As psychoacoustic measures of discrimination were related to speech perception8,11,12, and psychoacoustic measures of discrimination had been related to electrophysiological measures of discrimination13, it may be surmised that objective ACC measures induced by spectral variation in SRNs are related to subjective speech perception. There is some evidence that ACCs elicited using synthesized vowel tokens were related to behavioural detection in normal-hearing adults14; and that the latency of ACCs elicited using changing pure tones were correlated with speech perception scores in adults using cochlear implants15. The current study extends previous work by investigating the relationship between ACCs and behavioral discrimination elicited using SRN and speech perception.
It is widely known that CAEPs are influenced by characteristics of the stimulus, including the level of presentation. Cheek et al. showed that synthesized vowel tokens presented at 70 dBA elicited higher ACC amplitude than tokens presented at 40 dBA in normal-hearing listeners14. However, this effect may be different in people aided using hearing aids as audibility of the stimulus is determined not only by the signal level relative to the hearing threshold (as in normal-hearers) but also to the hearing aid gain and the internal noise level with amplification16. Because hearing sensitivity loss is generally more severe in the high than in the lower frequencies, and spectral processing abilities reduce with increase in hearing loss, it is important to increase knowledge about the effects of high- or low-pass filtering and presentation level of the SRN on the amplitude of ACCs in people using amplification, so that ACC measurements can be used to optimize hearing aids or to assess whether cochlear implants need to be considered to meet individual listening needs.
In this study, we aimed to investigate the relationship between subjective speech perception and objective ACC responses to sound changes measured using SRN. The findings will have significant implications on whether ACC responses to SRN can be used to guide intervention for hearing loss including potential decisions for cochlear implantation. Firstly, we evaluated the relationship between behavioral measures of phase inversion discrimination (or PID) and speech perception in quiet and in noise by adults with normal hearing (NH) and those with hearing loss (HL). In line with current knowledge, we hypothesized that better PID is related to better speech perception in quiet and in noise. Secondly, we examined the relationship between behavioral PID and cortical ACC responses to SRN. We hypothesized that better behavioral discrimination (lower detection thresholds for phase inversion) is related to stronger cortical ACC responses (higher cortical response amplitude). We also hypothesized that ACC response amplitude increases with higher presentation level. Thirdly, we examined the relationship between cortical ACC responses to SRN and speech perception in quiet and in noise. We hypothesized that people who had better speech perception performance also had better ACC performance and better psychoacoustic PID in response to SRN. Performance in all three measures will be affected by the participants’ hearing threshold level.
Results
Speech perception
For speech perception in quiet at 50% correct, the mean speech reception threshold (SRT) for normal hearing (NH) listeners was 18.7 dB SPL (SD = 5.7). The mean SRT for hearing impaired (HI) listeners was 56.9 dB SPL (SD = 12.9). For sentence perception in babble noise at 50% correct, the mean SRT in terms of signal-to-noise ratio (SNR) for the NH group was − 0.1 dB (SD = 1.1). The mean SNR for HI listeners was 3.9 dB (SD = 2.3 dB). A 2 × 2 repeated measures ANOVA with SRT (in Quiet and in Noise) as dependent variables and hearing status (NH and HI) as categorical variable showed significant interactions. Whereas NH listeners had better speech perception than HI listeners in quiet, there was no significant difference in performance in noise (p < 0.0001).
Phase inversion discrimination threshold (PID)
Table 1 summarizes the PID thresholds for NH and HI listeners for a low-pass- (LP) and a high-pass-filtered (HP) stimulus. A 2 × 2 repeated measures ANOVA with PID thresholds (HP and LP) as dependent variables and hearing status (NH and HI) as categorical variable showed that on average, NH listeners had better (lower) discrimination thresholds than HI listeners (p < 0.05). The main effect of filter condition was also significant – the discrimination thresholds for a low-pass filtered SRN were significantly lower (better) than those obtained for a high-pass filtered SRN (p < 0.001). The interaction effect was not significant (p > 0.05).
Electrophysiological measure of onset detection and ACC discrimination
The cortical P1-N1-P2 complex is a cortical auditory evoked potential elicited by the onset or offset of an acoustic stimulus or changes within an acoustic stimulus17. The cortical response to the onset of a stimulus is referred to here as “onset” denoting detection of a sound at the cortical level. The cortical response to a phase change within the stimulus is referred to as the ACC. Table 2 shows the z-scores for onset detection and ACC discrimination, separately for NH and HI participants. A 2 × 2 × 2 repeated measures ANOVA was performed on the z-scores (two response types: onset and ACC; two frequencies: high-pass and low-pass filtered; and two presentation levels: normal and low) with hearing status (NH vs HI) as a categorical factor. The main effects of response type, frequency, and level were significant (p < 0.0001, p < 0.0001, and p < 0.05 respectively). There were significant interactions between response type and hearing status (p < 0.01). Post-hoc analyses using Bonferroni comparisons revealed that NH subjects had significantly better ACC scores (lower z-score) than HI subjects, but there was no significant difference in onset scores. There were significant interactions between response type and frequency (p < 0.05). On average, ACC responses at high frequencies were poorer (lower in amplitude) than those at lower frequencies. Also, there were significant interactions between response type and level (p < 0.05). On average, onset detection was better at normal levels than at lower levels, but discrimination ACC was not significantly different between levels (p > 0.05).
Correlation analyses
Results of correlations among speech perception, phase inversion discrimination and cortical discrimination and detection are summarized in Table 3. On average, significant correlations were found between speech perception and phase inversion discrimination and cortical ACCs for the 23 participants. There were also significant correlations between speech perception and averaged hearing level (Quiet: r = 0.91, p < 0.001; Noise: r = 0.85, p < 0.001).
Phase inversion discrimination (PID) thresholds and speech perception scores were correlated. For high-pass filtered stimuli, correlations with speech perception in quiet and noise were significant (Quiet: r = 0.62, p < 0.01; Noise: r = 0.72, p < 0.001). Similarly, PID for low-pass filtered stimuli were significantly correlated with speech perception (Quiet: r = 0.46, p < 0.05; Noise: r = 0.59, p < 0.01). As shown in Fig. 1, better behavioral discrimination (lower PID expressed as lower modulation depth in dB) measured using high-pass and low-pass filtered SRNs were associated with better speech perception in quiet and in noise (lower speech reception threshold), for adults with normal hearing and those with hearing impairment. Linear regression showed that speech perception in noise SRT was significantly associated with phase inversion discrimination, after controlling for better-ear hearing level (High-pass: beta = 0.11, p < 0.01; Low-pass: beta = 0.10, p < 0.01).
Phase inversion discrimination was significantly correlated with cortical ACCs (p < 0.01) but not cortical onset responses (p > 0.05). As shown in Fig. 2, better behavioral discrimination (lower PID) was associated with stronger cortical ACC responses (lower z-scores depicting higher cortical response amplitude).
Speech perception was significantly correlated with cortical ACC (p < 0.05) but not onset responses (p > 0.05). As shown in Figs. 3 and 4, subjects with cortical ACCs that were higher in amplitude were better at perceiving speech in quiet and in noise. The strongest correlation (r = 0.70, p < 0.0001) was found between speech perception in noise and ACCs measured with high-pass filtered SRNs presented at normal conversational level (HP20NL). Linear regression analysis revealed that this association between speech perception and ACC z-score was no longer significant (p > 0.05) after controlling for the effect of audibility. It is possible that there are associations, but the current sample size was not large enough to detect them.
Discussion
The current study investigated the relationship between subjective speech perception and objective measures of auditory cortical responses induced by phase changes in SRN. A significant relationship would suggest a potential applicability of using objective measures of ACC to SRN as a predictor of hearing aid benefit for speech perception. This method may be used to identify patients who may benefit from cochlear implantation, especially those who cannot perform reliably in behavioral speech perception testing.
This study found a significant correlation between behavioral SRN modulation depth threshold and sentence perception in quiet and in noise in NH and HI listeners (Fig. 1). Better phase inversion discrimination was associated with better speech perception. This association was significant after controlling for the effect of hearing level (p < 0.01). Our results obtained using a phase inversion paradigm are consistent with previous research that used a SMD task to find modulation depth thresholds. Davies et al. and Won et al. demonstrated that the subjective ability to perceive SRN modulation depth and discrimination were significantly correlated with word recognition in quiet and in noise4,9. This lends support to applying SRN stimuli in electrophysiological measurements of spectral discrimination ability.
Secondly, this study found a significant relationship between behavioral discrimination and cortical ACC responses elicited by SRN. Better behavioral discrimination measured using high-pass and low-pass filtered stimuli was significantly associated with higher ACC response amplitude, but not with onset response. Consistent with previous studies, our results confirmed that ACC can be elicited by a change of spectral characteristics alone in non-linguistic stimuli presented at normal conversational level of around 65 dB SPL and at a low input level of 55 dB SPL4,9,13,15. On average, onset response amplitude was higher at higher presentation level, but the amplitude of ACC response was not significantly different between levels. This may be partly related to the extent to which cortical auditory neurons are sensitive to the direction of change of acoustic parameters18,19. On average, there was no significant difference in onset detection between NH and HI groups implying that the stimuli were audible and detected by both groups efficiently. However, the cortical ACC responses induced by SRN phase inversion was significantly stronger for NH than for HI people suggesting that HI participants had difficulty discriminating phase changes within the acoustic stimuli. As hearing aids increase audibility but cannot improve discrimination, this electrophysiological test using filtered SRN can be used to identify patients using hearing aids who can hear sounds but cannot discriminate between sounds especially in the high frequencies for cochlear implant candidacy evaluation. Our findings complement those reported by Choi et al. (2016) on using SRNs with spectro-temporal modulations for behavioral testing and extends the application of SRNs for electrophysiological testing12.
Thirdly, we found that ACC z-scores were significantly correlated with PID thresholds for high-pass and low-pass filtered SRN in NH and HI adults. These findings complement those reported by Won et al. (2011) that showed a significant relationship between ACC and spectral ripple discrimination9. Higher amplitude in ACC (lower z-score) was associated with better thresholds in phase discrimination (lower modulation depth). These results show that the ACC response to phase changes in the SRN could be used as an objective index of the listener’s capacity for extracting acoustic cues that are important in differentiating speech sounds13,20.
This study shows that cortical ACC induced by SRN phase inversion correlated significantly with speech perception in quiet and in noise (p < 0.001). Higher ACC amplitude was associated with better speech perception ability. This finding is consistent with the report of He et al. (2014) that showed correlations between ACCs evoked by changes in the stimulating electrode in cochlear implants of children with auditory neuropathy and their word scores21. We found that speech perception in noise was significantly correlated with ACC z-score for high-pass filtered SRN with a 20 dB modulation depth presented at 65 dB SPL in the free field (r = 0.7, p < 0.001).
However, when we partialled out the effect of audibility from the speech perception scores and looked at the relationship between speech perception and ACC z-score, there was no longer a significant relationship between speech perception in noise and ACCs for the SRN (p > 0.05). This is unlike the finding of a reduced but significant relationship between behavioral phase discrimination and speech perception in noise after controlling for the effect of audibility. Taken together, the findings suggest that audibility is a strong predictor of behavioral and electrophysiological measures of phase discrimination as well as speech perception. This lends support to the current adoption of audiometric criteria for referrals for cochlear implant candidacy in many countries22. Nevertheless, previous studies have failed to show a correlation between pre-operative hearing threshold levels and post-operative speech perception outcomes23,24,25,26. Considerable variability in post-operative outcomes exists. Some evidence suggests that variability may be explained partly by the auditory sensitivity of a user of cochlear implants to acoustic cues underlying speech perception and their attention to those cues27. Measurement of a listener’s auditory ability for sound discrimination pre-operatively may contribute prognostic information on likely outcomes. This may be especially useful in cases where audiometric thresholds do not meet candidacy criteria according to local regulatory authorities. Importantly, cortical measurements would assist with identifying cochlear implant candidates in patients diagnosed with auditory neuropathy spectrum disorder 28. The use of SRNs for assessing auditory discrimination has the potential for adoption across regions of diverse languages. The viability and validity of using electrophysiological measures of ACC to assess auditory discrimination ability using the SRNs suggest that objective methods can be applied to patients who cannot provide reliable behavioral responses.
Future studies will extend the current work to a larger sample to adequately control for the effect of audibility in electrophysiological measurement of ACC using SRN. It would be useful to perform further research examining in detail how to optimize SRNs for electrophysiological measurements for clinical applications. Further work will also extend ACC assessments to infants and young children for determining detection and discrimination of sounds with hearing aids. In infants with congenital hearing loss, providing early implantation is crucial to maximizing outcomes29.
Conclusion
The current study showed a significant correlation between speech perception and spectral discrimination abilities measured using psychoacoustic methods and electrophysiological methods, with SRN as stimuli in adults with normal hearing or with hearing loss using hearing aids. Cortical ACC induced by a phase inversion in an SRN was strongly correlated with speech perception in quiet and in noise. Cortical ACC was also significantly correlated with phase inversion discrimination measured using psychoacoustic methods. The ACC responses elicited using non-linguistic stimuli have the potential to be used as a clinical tool to objectively predict speech abilities for people who cannot provide reliable behavioral responses in speech perception testing. This objective marker can become a useful criterion for optimizing hearing aids and for identifying candidates for cochlear implantation.
Methods
Subjects
Twenty-three adult listeners who were native speakers of English participated in this study, comprising 13 with normal hearing (NH) and 10 with hearing impairment (HI). The NH group comprised 10 females and 3 males aged between 19 and 55 years (mean = 27, SD = 9.6) who had hearing thresholds within the normal range in both ears (averaged hearing thresholds at 0.5, 1, 2 and 4 kHz or 4FA < 20 dB HL). The HI group comprised 3 females and 10 males aged between 36 and 81 years (mean = 68.1, SD = 15.0). Their hearing loss ranged from moderate to profound degrees. All HI participants had worn hearing aids for at least 6 months. The study protocol was approved by the Hearing Australia Human Research Ethics committee. All participants provided written informed consent to participate in this study. This study was carried out in accordance with approved guidelines.
Equipment for behavioral testing
All the behavioral tests were conducted in free field in an acoustically treated booth. The stimuli were presented from a loudspeaker (B&W 600 Series) placed 180 cm from the subject, via a sound card (Fireface 800) controlled by a MATLAB (The Mathworks, Natick) graphical user interface running on a computer. All stimuli for the behavioral discrimination test were equalized in level in terms of loudness using a loudness model 30 implemented in Matlab (The Mathworks, Natick). A half-inch Brüel & Kjaer microphone was used for calibration. Stimuli were presented at an overall level of 20 sones, measured at the subject position with the subject absent. Speech perception tests were calibrated using a Matlab user interface and a Brüel & Kjaer Hand-Held Analyzer type 2250. The overall presentation level of speech stimuli was adjusted to be the same as an equivalent continuous noise (Leq) at 75 dB SPL.
Audiological assessment
Otoscopy was performed before pure tone audiometry was performed using insert earphones ER3 and an Interacoustic AD 28 Diagnostic Audiometer31. Hearing thresholds were measured at octave frequencies between 0.25 and 8.0 kHz in each ear using standard procedures32. The four-frequency average (4FA) hearing loss was calculated by computing the mean of the thresholds obtained at 0.5, 1.0, 2.0, and 4.0 kHz in each ear. The hearing aids used by HI participants at their personal settings were measured using an HA2-2cc coupler and a broad-band speech-weighted stimulus using the Aurical Hearing Instrument Test (HIT) box and the OTOsuite software (Natus Medical Inc., Taastrup, Denmark). Average real-ear-to-coupler differences were used to estimate levels of an amplified signal in the real-ear33. The real-ear levels were used in calculations of audibility.
Speech perception
Two different types of speech stimuli were used. Consonant–vowel-consonant (CVC) words were presented in a carrier sentence in quiet, and the Bamford-Kowal-Bench (BKB)-like sentence test material34 were presented in babble noise. Training was provided before testing.
Speech perception in quiet
Recorded stimuli were presented at 65 dB SPL. Each carrier sentence contained two meaningless words; e.g. I saw the [CVC] in the [CVC]; spoken by a female native speaker of Australian English. Subjects were required to repeat the words they heard. The input level was adjusted in 5 dB steps for the first two reversals, after which the step size was reduced to 2 db. The speech reception threshold or SRT at which 50% of the words were correctly repeated was calculated, based on the average of the last 10 reversals. The test stopped when at least 24 stimuli were presented, and when 10 reversals were obtained with the 2-dB step size with a variance of less than 1 dB. The SRT, expressed in dB SPL, was calculated as the mean of the last 10 reversals.
Speech perception in noise
Recorded BKB-like sentences spoken by a male native speaker of Australian English were presented in a babble noise generated using two collocated talkers produced by native Australian English speakers. Subjects were sked to repeat the sentences. The SRT for 50% of the sentences correctly repeated was measured. The target speech was fixed at 65 dB SPL, and the babble noise was adjusted adaptively. The test started at 0 dB SNR, and adjustments were in 5 dB steps for the first two reversals, then in 2-dB steps. The test stopped when the variance was smaller than 1 dB. The SRT, expressed in dB SNR, was calculated as the mean of the last 10 reversals.
Behavioral discrimination test
Phase inversion discrimination (PID) thresholds were measured using spectrally modulated non-linguistic stimuli (SRN). The modulation depth thresholds for which a 180° phase inversion was perceived 50% of the time was measured.
All the stimuli were created by superposition of 4000 pure tone sounds equally distributed on a logarithmic scale. Two bandwidth conditions were included: a low-pass (LP) filtered SRN with a frequency bandwidth ranging from 250 to 1,500 Hz, and a high-pass (HP) filtered SRN with a frequency bandwidth ranging from 2,000 to 11,200 Hz. The spectral modulation starting phase for the ripple was randomly selected from a uniform distribution (0 to π rad). The ripple density was one ripple per octave. The stimuli had a total duration of 500 ms with a 10 ms up and down ramp. The inter-stimulus-interval (ISI) was 500 ms. To measure the PID threshold, a 3-Alternative-Forced-Choice (3AFC) paradigm, with a two-down one-up procedure was used. During the test, one interval had a 180° phase shift (i.e. deviant signal) compared to two other stimuli (i.e. standard signal). The participant had to select the one that was different from the other two. The position of the deviant stimulus in each triplet was randomized across trials. The modulation depth was initially set to 20 dB for the NH group and 25 dB for the HI group. The modulation depth varied automatically in steps of 2 dB for the first two reversals and 0.5 dB for the last ten. The threshold was defined as the average modulation depth of the last 10 reversals. The overall presentation level was 20 sones. All testing was preceded by training.
Electrophysiological tests
The electrophysiological test consisted of the recording of the electroencephalogram (EEG) in response to SRN stimuli with similar characteristics to the one used for behavioral testing. Electrophysiological assessment consisted of recording the CAEPs elicited by the onset of the SRN and the acoustic change complex (ACC) elicited by a transition between two SRN with phase inversion from 0° to 180° (see Fig. 5). The HEARLab system (Frye Electronics, Tigard, OR, USA) was used to record the raw EEG data.
Stimuli
All the stimuli were created by concatenating two SRNs that have a 180° phase inversion, at one ripple per octave. The total duration of the stimulus was 3 s with an SRN transition at 2 s after onset. To avoid spectral splatter at the transition, each stimulus was windowed with a 40 ms rise-fall ramp and the stimuli were concatenated with a 20 ms overlap. Two different modulation depths were investigated: 20 and 50 dB for the high-pass filtered condition. A pilot comparison of responses to high-pass filtered stimuli that have either 20 dB modulation depth or 50 dB modulation depth revealed no significant difference in onset or ACC responses. Therefore, 20 dB modulation depth was used for subsequent testing. The HP and LP stimuli used for PID threshold assessments were used. These were presented at normal conversational level (NL) of 20 sones (or around 65 dB SPL) and at a lower input level (LL) of 55 dB SPL. Four stimuli were used: HP20NL (high-pass filtered, 20 dB modulation depth, presented at normal level); HP20LL (high-pass filtered, 20 dB modulation depth, presented at low level); LP20NL (low-pass filtered, 20 dB modulation depth, normal level); and LP20LL (low-pass filtered, 20 dB modulation depth, low level).
Electrophysiological procedure
Testing was conducted in the free field using a single loudspeaker, positioned 100 cm from the subject, at 0° azimuth. All recordings were obtained in a sound-treated booth while the subject sat in a comfortable chair attending to a muted, close-captioned movie of their choice. EEG recording was conducted using four electrodes: one reference electrode placed at FCz and two active electrodes (“M1” on left mastoid and “M2” on right mastoid). The ground electrode was placed on the forehead35. For each stimulus, 120 epochs were recorded. The presentation order of stimuli was randomized across participants using a Latin square design36. Each condition lasted 13 min for a total of an hour and fifteen minutes approximately.
The EEG data were post-processed using a custom-designed Matlab script. Before analog-to-digital conversion, the signal was high-pass filtered at 0.33 Hz by an analog first-order filter. The signal was down-sampled to 1 kHz and low-pass filtered at 30 Hz with a 128-order zero-time delay filter. The raw EEG of each epoch was cut into two segments corresponding to the onset (CAEP induced during the transition from silence to sound stimulation) and the ACC (CAEP induced by the phase inversion of the SRN). The recording window for each response consisted of a 200 ms pre-stimulus interval and 700 ms post-stimulus interval. An artifact rejection algorithm was then applied to exclude any epoch that had a maximum absolute value threshold greater than 160 μV or a mean epoch value greater than 60 μV. A weighted average of the two channels was calculated for each epoch. Then, each recorded epoch was reduced to voltage levels in nine bins, covering the range from 51 to 348 ms, with each bin being 33 ms wide37. The Hotelling’s T2 statistics was used to obtain an objective measure of response detection—a statistically significant p-value means that an electrical activity corresponding to a CAEP was detected to be significantly different from zero or background noise. The z-score was computed from the p-value of the Hotelling’s T2 for both the onset and the ACC responses for each stimulus condition.
Calculation of audibility of stimuli
Audibility was the difference between the signal level and the maximum of hearing threshold and noise level. In HI listeners, real-ear measurements were used to calculate the level of the amplified signal at the ear-drum in the better ear. Audibility was estimated as the maximum value across different bands of one-third octave spectrum on the basis that audibility will most strongly be determined by the frequency region in which it is greatest.
Data analysis
To determine the influence of hearing status (NH or HI) on performance in speech perception, PID, and cortical onset and ACC responses, analysis of variance (ANOVA) with repeated measures was used. To determine the influence of presentation level (normal and low), and frequency (HP or LP) on cortical onset and ACC responses, ANOVA with repeated measures was used. To examine the relationship between PID and cortical responses, and between PID and speech perception, and between cortical responses and speech perception, Pearson’s product moment correlation analyses were carried out. Linear regression analyses using the SRT in noise as the dependent variable, and better-ear audibility together with either behavioral discrimination or ACC-z scores as predictor variables were conducted to estimate the effect of audibility on the relationship between discrimination and speech perception.
References
Glasberg, B. R. & Moore, B. C. J. Auditory filter shapes in subjects with unilateral and bilateral cochlear impairments. J. Acoust. Soc. Am. 79, 1020–1033 (1986).
Ching, T., Dillon, H., Katsch, R. & Byrne, D. Maximizing effective audibility in hearing aid fitting. Ear Hear. 22(3), 212–224 (2001).
Boothroyd, A., Mulhearn, B., Gong, J. & Ostroff, J. Effects of spectral smearing on phoneme and word recognition. J. Acoust. Soc. Am. 100, 1807–1818 (1996).
Davies-Venn, E., Nelson, P. & Souza, P. Comparing auditory filter bandwidths, spectral ripple modulation detection, spectral ripple discrimination, and speech recognition: Normal and impaired hearing. J. Acoust. Soc. Am. 138, 492–503 (2015).
Lopez Valdes, A. et al. Objective assessment of spectral ripple discrimination in cochlear implant listeners using cortical evoked responses to an oddball paradigm. PLoS ONE 9, e90044 (2014).
Won, J. H., Drennan, W. R., Nie, K., Jameyson, E. M. & Rubinstein, J. T. Acoustic temporal modulation detection and speech perception in cochlear implant listeners. J. Acoust. Soc. Am. 130, 376–388 (2011).
Supin, A. Y., Popov, V. V., Milekhina, O. N. & Tarakanov, M. B. Frequency resolving power measured by rippled noise. Hear. Res. 78, 31–40 (1994).
Litvak, L. M., Spahr, A. J., Saoji, A. A. & Fridman, G. Y. Relationship between perception of spectral ripple and speech recognition in cochlear implant and vocoder listeners. J. Acoust. Soc. Am. 122, 982–991 (2007).
Won, J. H. et al. Relationship between behavioral and physiological spectral-ripple discrimination. JARO - J. Assoc. Res. Otolaryngol. 12, 375–393 (2011).
Kim, J. R. Acoustic change complex: Clinical implications. Korean J. Audiol. 19, 120–124 (2015).
Henry, B. A., Turner, C. W. & Behrens, A. Spectral peak resolution and speech recognition in quiet: Normal hearing, hearing impaired, and cochlear implant listeners. J. Acoust. Soc. Am. 118, 1111–1121 (2005).
Choi, J. E. et al. Evaluation of cochlear implant candidates using a non-linguistic spectrotemporal modulation detection test. Sci. Rep. 6, 1–10 (2016).
He, S., Grose, J. H. & Buchman, C. A. Auditory discrimination: The relationship between psychophysical and electrophysiological measures. Int. J. Audiol. 51, 771–782 (2012).
Cheek, D. & Cone, B. Evidence of vowel discrimination provided by the acoustic change complex. Ear Hear. https://doi.org/10.1097/AUD.0000000000000809 (2020).
Liang, C., Houston, L. M., Samy, R. N., Abedelrehim, L. M. I. & Zhang, F. Cortical processing of frequency changes reflected by the acoustic change complex in adult cochlear implant users. Audiol. Neurotol. 23, 152–164 (2018).
Van Dun, B., Kania, A. & Dillon, H. Cortical auditory evoked potentials in (un)aided normal-hearing and hearing-impaired adults. Semin. Hear. 37, 9–24 (2016).
Martin, B. A., Tremblay, K. L. & Korczak, P. Speech evoked potentials: From the laboratory to the clinic. Ear Hear. 29, 285–313 (2008).
Eggermont, J. J. Between sound and perception: Reviewing the search for a neural code. Hear. Res. 157, 1–42 (2001).
Zhang, L. I., Tan, A. Y. Y., Schreiner, C. E. & Merzenich, M. M. Topography and synaptic shaping of direction selectivity in primary auditory cortex. Nature 424, 201–205 (2003).
Martin, B. A. & Boothroyd, A. Cortical, auditory, event-related potentials in response to periodic and aperiodic stimuli with the same spectral envelope. Ear Hear. 20, 33–44 (1999).
He, S., Grose, J. H., Teagle, H. F. B. & Buchman, C. A. Objective measures of electrode discrimination with electrically evoked auditory change complex and speech-perception abilities in children with auditory neuropathy spectrum disorder. Ear Hear. 35, e63 (2014).
Vickers, D., De Raeve, L. & Graham, J. International survey of cochlear implant candidacy. Cochlear Implants Int. 17, 36–41 (2016).
Ching, T. Y. C., Incerti, P. & Hill, M. Binaural benefits for adults who use hearing aids and cochlear implants in opposite ears. Ear Hear. 25, 9–21 (2004).
Blamey, P. J. et al. Factors predicting postoperative sentence scores in postlinguistically deaf adult cochlear implant patients. Ann. Otol. Rhinol. Laryngol. 101, 342–348 (1992).
Waltzman, S. B., Niparko, J. K., Fisher, S. G. & Cohen, N. L. Predictors of postoperative performance with cochlear implants. Ann. Otol. Rhinol. Laryngol. 104, 15–18 (1995).
Gantz, B. J., Abbas, P. J., Tyler, R. S., Woodworth, G. G. & Knutson, J. F. Multivariate predictors of audiological success with multichannel cochlear implants. Ann. Otol. Rhinol. Laryngol. 102, 909–916 (1993).
Moberly, A. C., Lowenstein, J. H. & Nittrouer, S. Word recognition variability with cochlear implants: ‘Perceptual attention’ versus ‘auditory sensitivity’. Ear Hear. 37, 14 (2016).
He, S. et al. Acoustically evoked auditory change complex in children with auditory neuropathy spectrum disorder: A potential objective tool for identifying cochlear implant candidates. Ear Hear. 36, 289 (2015).
Ching, T. Y. C. et al. Age at intervention for permanent hearing loss and 5-year language outcomes. Pediatrics 140(3), 1–11 (2017).
Moore, B. C. J. & Glasberg, B. R. A revision of Zwicker’s loudness model. Acta Acust. United Acust. 2, 335–345 (1994).
Clemis, J. D., Ballad, W. J. & Killion, M. C. Clinical use of an insert earphone. Ann. Otol. Rhinol. Laryngol. 95, 520–524 (1986).
Carhart, R. & Jerger, J. F. Preferred method for clinical determination of pure-tone thresholds. J. Speech Hear. Disord. 24, 330 (1959).
Dillon, H. Hearing Aids. in Hearing Aids (ed. Thieme) 81–127 (Boomerang Press, 2012).
Bench, J. & Bamford, J. Speech-Hearing Tests and the Spoken Language of Hearing-Impaired Children (Academic Press, 1979).
Homan, R. W., Herman, J. & Purdy, P. Cerebral location of international 10–20 system electrode placement. Electroencephalogr. Clin. Neurophysiol. 66, 376–382 (1987).
Bradley, J. V. Complete counterbalancing of immediate sequential effects in a Latin square design. J. Am. Stat. Assoc. 53, 525–528 (1958).
Golding, M., Dillon, H., Seymour, J. & Carter, L. The detection of adult cortical auditory evoked potentials (CAEPs) using an automated statistic and visual detection. Int. J. Audiol. 48, 833–842 (2009).
Acknowledgements
Research reported in this study was partly supported by the National Health and Medical Research Council Grant APP1088931 awarded to TYCC. The research was also partly supported by the National Institute on Deafness and Other Communication Disorders of the National Institutes of Health under Award Number R01DC008080 awarded to TYCC. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. We acknowledge the contributions of Bram Van Dun, James Galloway, and Harinath Garudadri to the design of the research.We acknowledge Mark Seeto for his assistance with statistical analysis, Vicky Zhang for her assistance with hearing aid measurements, and Carmen Kung for her comments on an earlier version of the manuscript. We thank all the participants for contributing to this study.
Author information
Authors and Affiliations
Contributions
The authors confirm contribution to the paper as follows: study conception and design: H.S., F.B. and T.Y.C.C.; data collection: H.S.; analysis and interpretation of results: H.S. and F.B.; draft manuscript preparation: T.Y.C.C. and H.S. All authors reviewed the results and approved the final version of the manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Sohier, H., Bardy, F. & Ching, T.Y.C. Relationship between objective measures of hearing discrimination elicited by non-linguistic stimuli and speech perception in adults. Sci Rep 11, 19554 (2021). https://doi.org/10.1038/s41598-021-98950-5
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-021-98950-5
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.