Listening in a noisy environment is challenging for individuals with normal hearing and can be a significant burden for those with hearing impairment. The extent to which this burden is alleviated by a hearing device is a major, unresolved issue for rehabilitation. Here, we found adult users of cochlear implants (CIs) self-reported listening effort during a speech-in-noise task that was positively related to alpha oscillatory activity in the left inferior frontal cortex, canonical Broca’s area, and inversely related to speech envelope coherence in the 2–5 Hz range originating in the superior-temporal plane encompassing auditory cortex. Left frontal cortex coherence in the 2–5 Hz range also predicted speech-in-noise identification. These data demonstrate that neural oscillations predict both speech perception ability in noise and listening effort.
We live in complex noisy environments. Typical human interactions occur in places where we need to focus our attention on one talker while ignoring others. Anecdotally, most of us have experienced some degree of listening effort following a conversation in a noisy pub and have noticed that we need to expend a lot of “cognitive effort”1,2 to successfully follow the conversation. The neural mechanisms of this “listening effort” are not well understood. To date, the study of listening effort has mostly focussed on indirect measures of brain activity (e.g., arousal mediated pupil diameter). In this study we describe a direct, brain-based biomarker of listening effort that is mediated through left frontal language areas of the brain and sensory regions of the auditory cortex. We chose to study people who are hearing impaired and known to have greater degrees of listening effort compared to normal hearing listeners3,4. Increased effort has been related to increased frustration, fatigue, and decreased concentration in this population, impacting work or school performance, resulting in overall poor quality of life, social withdrawal and depression5,6. The current philosophy in hearing loss management is to restore to sound frequency thresholds to normal ranges with no regard to manage how hard it is to listen.
Research on the effects of listening effort on hearing is still in its infancy. One challenge is to quantify listening effort. The current battery of objective tests of listening effort includes pupil diameter (pupillometry), dual-task measures, where listening effort is inferred from reductions in secondary task performance while the primary task is a listening task such as speech in noise perception; see review in Gagné7, or reaction times while listening in a challenging environment (see review in McGarrigle1). Although there is a rich history of using these methods to quantify listening effort, they are indirect measures of brain activity. Having a brain measure of listening effort in hearing impaired populations has far reaching implications. If the site(s) of dysfunction leading to increased listening effort can be inferred from brain physiology, clinicians could use this to guide optimal rehabilitation, for instance focussing on cognitive intervention in conjunction with hearing technology. Cognitive factors including executive function, working memory, attention and motivation are thought to be strongly engaged during effortful listening2. A recent meta-analysis of brain imaging studies (fMRI, functional magnetic resonance imaging and PET, positron emission tomography) suggested that effortful listening involves the recruitment of brain regions beyond primary auditory cortex8. Speech-in-noise typically recruited left inferior frontal gyrus (IFG), left inferior parietal lobule, and right insula. Effort resulting from spectral degradation typically recruited insula bilaterally and the left superior temporal gyrus (STG). Effort involving linguistic complexity typically activated the left IFG, right middle frontal gyrus, left middle temporal gyrus and bilateral STG. The common feature of these studies is that effortful listening was associated with recruitment of primary auditory cortex, the canonical language network (left perisylvian regions, including Broca’s and Wernicke’s areas), and extracanonical regions involved in language processes9.
A limited number of electroencephalogram (EEG) studies have addressed neurophysiological correlates of listening effort. Wisniewski and colleagues10 employed a speech in noise measure (English Sentence Matrix test) while participants reported subjective ratings of effort. Effort was correlated with an increase in frontal theta (4–7 Hz) activity. Wöstmann and colleagues11 observed that alpha power (7–13 Hz) varied proportionately with self-reported listening effort (not related to the stimuli that were used in the study) in a listening task that altered speech spectral detail. We12 and others13 have shown that listening in challenging environments (e.g., speech in noise or vocoded speech) is associated with more alpha modulation, especially in parietal cortex, relative to less challenging tasks. The mechanism underlying an increase in alpha power is incompletely understood and an increase in parietal alpha may be related to inhibition of competing visual inputs14 or inhibition of dorso-ventral attention pathways enhancing speech object processing12. Decreases in temporal-lobe alpha, related to auditory cortex activation, have been observed during speech perception in noise and were a better predictor of speech identification than parietally-generated alpha12.
Measures of neural entrainment to speech and other auditory signals may also provide insight to the mechanisms of listening effort. Neural entrainment, also referred to as ‘neural tracking’ or ‘cerebro-acoustic coherence’, is a measure that relates acoustic feature fluctuations to brain activity fluctuations. The simplest rhythmic feature is sinusoidal amplitude modulation giving rise to auditory steady-state responses15, defined as a peak in the EEG amplitude spectrum at the amplitude modulation rate. More recently, research focus has shifted towards natural speech, where the amplitude envelope16,17,18,19 or phonetic information20 is related to brain fluctuations. Neural entrainment to speech envelopes has been associated with successful selective attention in dichotic listening tasks21,22,23 but has never been used to quantify the degree of listening effort. Here, we also demonstrate that the degree of neural entrainment to speech envelopes is related to speech perception in noise and to listening effort arising from different areas of the brain.
We examined the relationship between alpha oscillations, neural entrainment and listening effort in cochlear implant (CI) users. A CI is a prosthetic device that delivers auditory stimulation via electrodes surgically implanted in the cochlea24. The incoming sound is received via a microphone behind the ear and processed in a series of filter banks. The filtered sound envelope is used to modulate an electrical pulse train that stimulates the auditory nerve. The representation of the sound is degraded in both spectral and temporal domains. The spectral degradation is partly due to the reduced number of channels or stimulating electrodes and to the large current spread to other electrodes resulting in broader psychoacoustic filters compared to normal acoustic hearing25. This degradation results in reduced speech in noise perception26 and reduced perceptual segregation between target and masker stimuli27. Temporal processing degradation in CI users has often been quantified using amplitude modulation detection thresholds related to speech perception ability28,29. Reduced spectro-temporal encoding ability in CI users may lead to recruitment of extra cognitive resources to understand speech in noise. CI users typically require more effort while listening to speech in noise compared to normal hearing counterparts3,4, even when the intelligibility of the speech is adjusted for equal performance4.
In this study we used a speech-in-noise procedure (the ‘digits-in-noise’; DIN) that is becoming widely used30 and is closely related to pure tone thresholds in people with hearing impairment31. Previously, we found that EEG alpha oscillations in normal hearing listeners are modulated by attention and accuracy in performing the DIN12. Here we demonstrate that listening effort during DIN testing in CI users is related to: (1) increased alpha oscillations in left frontal regions, similar to fMRI activation in normal hearing8,32 and, (2) decreased neural entrainment to the speech envelope of the DIN stimulus. The general pattern of oscillatory activity is the same in both normal hearing listeners12 and CI users in this study, therefore we believe that these findings are not limited to just CI users but rather generalizable to the normal hearing population.
Attention modulates brain oscillations in CI users
All the DIN stimuli for the electrophysiological recordings were presented at a digit signal-to-noise ratio (SNR) corresponding to the individual’s speech reception threshold (SRT). The SRT is the digit signal-to-noise ratio yielding 50% correct performance12. The SRT was determined in each participant in a prior behavioural session, before the EEG DIN data were collected. CI users had behavioural SRTs ranging from −6 to −13 dB (mean −7.7 dB; SD 2.7). This is significantly (t(21) = −10.8; p < 1e-6) greater than the SRTs (mean = −16.4 dB, SD = 0.8) observed in our previous DIN study in normal hearing listeners12 confirming that CI users have lower performance perceiving speech in noise.
CI users listened to DIN in both passive and attentive listening conditions, where attentive listening required verbal identification of the presented digits. During passive listening, CI users were instructed to ignore the digits while they watched a closed-captioned movie of their choice. Figure 1 shows the grand mean time-frequency representations across all CI users during passive and attentive listening conditions. The most prominent differences in response to the two listening conditions was the increase (or event-related synchronization; ERS33) in alpha power (8–12 Hz) during digit presentation (Fig. 1a), bursts of event-related desynchronization (ERD) for gamma (35–40 Hz), synchronized with each of the presented digits, and a beta (20–28 Hz) ERD at final digit offset. The scalp topography indicated a centroid over parietal for the alpha ERS and frontal and central peaks for the beta and gamma. Beamformer34 source analysis was performed for each participant based on the maximum ERD/ERS time-frequency windows in the grand mean data. The dominant generators were the right parietal cortex for alpha, left IFG for beta, and bilateral anterior temporal lobes for gamma (Fig. 1b).
Left frontal alpha power predicts listening effort
Listening effort was assessed using the NASA Task Load Index35 after each recording block. This self-reported rating was then used to perform the correlations with the electrophysiological data offline. Figure 2 shows that alpha power during the DIN task is significantly correlated with self-reported effort for the digit identification task. Correlations close to r = +1 were observed bilaterally in frontal regions encompassing left inferior frontal gyrus (IFG) and insula (Fig. 2a; peak at Talairach −39, 11, 10 coordinates, IFG according to http://sprout022.sprout.yale.edu/mni2tal/mni2tal.html). Significant correlation clusters were observed in these regions after multiple comparisons corrections36 (Fig. 2b; criterion: p = 0.048) CI users with higher alpha ERS showed greater listening effort whereas those with greater alpha desynchronization (ERD) showed less listening effort (Fig. 2c). No significant correlations were observed between alpha power source and DIN identification performance. Note that although the dominant alpha ERS source was found to be in parietal regions (Fig. 1), the peak correlations between alpha power and listening effort were observed in left frontal regions. This apparent brain source discrepancy is a result of using different computational approaches for the two measures. Parietal alpha ERS was computed from the peak activity for the beamformer source localization across subjects whereas the left frontal alpha listening effort was the result of computing correlations between individual subject’s alpha source and listening effort ratings, yielding a Pearson correlation value for each brain voxel. The left frontal regions showed the most consistent alpha relationship with listening effort even though the source was dominated by parietal regions.
No significant correlations between beta or gamma sources and listening effort were observed. However, interestingly, a negative correlation (r = −0.93) between gamma power and digit identification was observed in the occipital cortex (data not shown).
Brain coherence with the speech envelope is related to listening effort
Neural entrainment to the digit speech envelope was assessed using Dynamic Imaging of Coherent Sources (DICS34) as implemented in BESA37. DICS calculated the coherence between the envelope of the DIN stimuli (low pass filtered at 10 Hz) and neural sources in the 2–5 Hz range over a time range that encompassed the digit onset and offset (1–5 seconds). The envelope of typical human syllables also occurs over the 2–5 Hz range38. Figure 3 shows that lower speech-brain coherence was associated with higher listening effort. Correlations between the voxel-wise DIN envelope coherence and listening effort were computed across subjects. Negative correlations approaching r = −1 were observed bilaterally in the temporal lobe encompassing the auditory cortex (Fig. 3a). Only left hemisphere clusters survived corrections for multiple correlation comparisons (Fig. 3b). As with the alpha source described above, correlations between DIN identification and coherence were not significant.
Brain coherence in left frontal regions predicts DIN performance
Speech-brain coherence was compared between trials with correct and incorrect digit identification (Fig. 4). The CI user had to identify all three digits correctly for a trial to be classified as “correct”. Although high coherence in left and right auditory cortex regions on both correct and incorrect trials was observed, a paired-test yielded differences between correct and incorrect identification only in the left prefrontal cortex (criterion: p = 0.005, corrected for multiple comparisons; Talairach coordinates −32, 40, 16; BA10) where higher speech-brain coherence was observed on correct trials. Additionally, Supplemental Fig. S1 demonstrates significant coherence with the speech envelope in left auditory temporal regions.
The purpose of this study was to discover functional and structural neural substrates of self-reported measures of listening effort. Three novel findings were observed: (1) left frontal IFG alpha power was positively associated with listening effort; (2) speech envelope auditory cortex coherence in the 2–5 Hz range was negatively associated with listening effort; (3) accuracy of speech identification was positively associated with envelope speech-brain coherence in the left frontal cortex.
The term ‘listening effort’ was coined by translational auditory scientists to describe phenomena of ‘tiredness’ and the ‘compensation’ required to listen when hearing impaired or using hearing prosthetics. This study identifies neural correlates associated with these phenomena. More recently, focus has shifted to a recognizably cognitive approach. The observed correlation between left IFG alpha activity and self-report listening effort is consistent with a recent meta-analysis report summarizing the brain regions associated with effortful listening8. When listening in challenging environments, the classical auditory regions of the superior temporal gyrus (STG) cannot effectively process the speech sound, although they are activated39 and cortical regions supporting executive function, attention, memory and sensorimotor processing are recruited32. With effortful listening, the most commonly observed activated regions of the brain include the frontoparietal network including left IFG and left inferior parietal lobule. The left IFG includes canonical language production regions such as Broca’s area, known to be activated during covert or internalized speech production40 and working memory involving speech41. Increased alpha power in a cortical region likely reflects decreased neural activation42 by suppressing non-crucial or potentially distracting information allowing for more efficient neural processing. Increases in alpha have, for example, been reported for ipsilateral spatial attention to auditory43,44,45,46,47,48,49, visual50 and somatosensory51 stimuli and for modality switching of attention52. In this study we observed a continuum of alpha event-related activation, with more ERD for lower effort and more ERS for higher effort ratings. The interpretation of the direction of this alpha-listening effort (ERD, lower effort; ERS higher effort) effect is difficult. Alpha ERD in the left IFG would suggest neural activation with less listening effort. This is opposite to what we would expect given that this increases in this region are observed with greater listening effort8. Opposite alpha effects were observed when speech stimuli were used as a distractor stimuli in an auditory working memory study, leading the authors to suggest that the listener’s goal, rather than acoustic degradation, drives increases in alpha53. In the present study, the increased alpha may be related to attentional modulation of the distracting noise. Another interpretation is that increased listening effort is manifested as suppression of language production, given that the task involved a verbal response. This may be related to the observation that poor speech SNRs recruit speech motor systems as a compensatory mechanism for speech in noise perception54. Nonetheless, the involvement of canonical language areas appears to be involved in the current listening effort paradigm.
A limited number of previous studies have examined neural correlates of listening effort. Our data complement and extend the suggestion by Wöstmann and colleagues11 that alpha oscillations are related to listening effort. In their study, normal hearing listeners, young and elderly, had to make predictive number sequence judgments to filtered speech digits. Listening effort in their study was quantified in a general post experiment questionnaire unrelated to the experimental task in time or content. General self-reported listening effort was found to be significantly related to a derived measure of alpha that was modelled on how alpha changes with different acoustic manipulations and stimulus predictability. In the current study a direct relationship between alpha oscillations and listening effort was observed and localized with brain source analysis. We related alpha power during the time participants were listening to the digits in noise to effort ratings of the digits after each block, thereby demonstrating a more direct relationship. The other most relevant study is that from Wisniewski and colleagues10 who found that frontal theta was correlated to sentences presented in noise. This effect was attributed to the presumed role that theta plays in working memory. Although there was alpha ERS during sentence presentation in the data of Wisniewski and colleagues10 (see Fig. 2 in their study), the relationship with listening effort was not reported. One possible explanation for the differences in theta activity between this study and that of Wisniewski and colleagues is that listening to sentences involves greater demands on working memory than does listening to three digits, therefore yielding more robust theta oscillations. In a series of combined EEG and pupillometry studies, McMahon and colleagues13,55 failed to find relationships between alpha oscillations and listening effort. One possible explanation for the discrepancy between the current study and McMahon’s is related to how the alpha oscillations were quantified. Here, correlations across voxels in brain space were measured, whereas in the McMahon studies, observations were limited to parietal sensors. Although we and others have shown that alpha has a dominant generator in the parietal cortex, the relationship with listening effort appears to be left frontally-generated.
In contrast to the alpha oscillations, beta and gamma oscillations showed no relationships with listening effort. The observed beta ERD in left IFG is consistent previous left IFG beta ERD in expressive language paradigms56,57 suggesting that at the offset of the digits, speech production is initiated. Although distinct bursts of gamma ERD occurred with digit presentation and localized to anterior temporal lobes, the functional significance of these are difficult to interpret.
In this study we observed two novel speech-brain coherence effects: the strength of speech-brain coherence in auditory cortex was inversely related to listening effort, and speech-brain coherence in left frontal regions was positively related to correct digit identification. Speech-brain coherence is presumably reduced by the addition of energetic noise masking, which poses a particular challenge for CI users58. Previous work suggests that how well the auditory neural system entrains to speech is a good predictor of speech perception in noise in normal hearing59 and in hearing impaired listeners using a hearing aid22. We hypothesized that auditory cortex coherence to the speech envelope would be inversely related to number of correctly identified digits. No such relationship was observed. Rather, a significant inverse relationship with listening effort was found, suggesting that degree of sensory disruption in the auditory cortex from the noise masker is associated with a proportional increase in listening effort. This observation is consistent with previous work showing that reducing the SNR of speech is associated with increased listening effort, as measured with pupillometry60. Both frontal IFG and auditory cortex were associated with listening effort and therefore would suggest a functional relationship exists between auditory cortex and left IFG. Although it may be tempting to infer that reduced auditory cortex sensory encoding is driving the left IFG, connectivity measures between alpha rhythms and speech-brain coherence are difficult given that they quantify different types of brain activity. However, recent work examining connectivity measures between left frontal regions and left auditory cortex in the delta/theta (1–7 Hz) range, typical of speech acoustics, have indicated that left frontal regions are a top-down driver for the auditory cortex during natural speech61. Functional connectivity from left IFG to auditory cortex has been shown to be stronger in listeners with higher QuickSIN performance62. Additionally, frontal-auditory connectivity, seen in both structural and electrophysiological measures, appears to relate to the ability to learn new words, which may be especially important in adults learning to hear again and/or adapt to a cochlear implant63. These findings therefore suggest that correct digit identification is associated with left frontal speech-brain coherence driving auditory cortex. Left frontal speech-brain coherence is also consistent with previous fMRI/ECoG (electrocorticography) noise vocoding and sine-wave speech learning studies demonstrating increased left IFG activity after successful speech vocoder training64,65,66,67.
Although we have described some novel neural metrics for listening effort in cochlear implant users, some study caveats exist. Given that our experimental task involved a verbal feedback of the digits, language production areas of the brain will necessarily be recruited. A non-verbal task would help clarify the role of IFG in listening effort. Additionally, the alpha and coherence correlations with listening effort across subjects were based on a relatively small sample size. Studies with larger samples sizes and with different types of hearing loss are warranted.
Having direct neural correlates of listening effort in CI users has important clinical implications. Pupillometry, dual-task, and reaction time, used in the majority of studies6, are indirect measures associated with neural processing. One advantage of auditory electrophysiology, a direct measure of neural activity, is that the location of a disorder can be approximated. For example, with auditory neuropathy, absent auditory brainstem responses but robust cortical responses suggest abnormal subcortical encoding. With deaf users of CIs, a cortical measure indicating possible sites of dysfunction (e.g. primary auditory cortex or frontal cortex) could provide clinicians with optimal rehabilitation strategies (e.g. CI programming or cognitive training, respectively). These results may also provide a reference framework for a cortically driven CI to reduce listening effort, similar to the growing field of cognitively controlled hearing aids68.
The current study follows a previous study12 that examined digits in noise speech perception in normal hearing adults. The stimuli and recordings were identical and described only briefly below.
Ten adult cochlear implant participants (6 females; mean age: 49.5 years, range: 23–74 years) were recruited through Cincinnati Children’s Hospital Medical Center, according to an Institutional Review Board. All experimental protocols used in this study were approved by the Cincinnati Children’s Hospital Medical Center Institutional Review Board (Study number 2013-0105). All the methods used in this study were performed in accordance to the guidelines and regulations outlined in the Cincinnati Children’s Hospital Medical Center Institutional Review Board (Study number 2013-0105).
Participants had no clinically significant neurological or mental health issues. Participants received a monetary incentive and provided informed consent. All CI users had at least one-year CI experience prior to testing, a summary of the patient demographics is shown in Table 1.
Stimuli and procedure
Procedures and rationale for recording, equalizing, and homogenizing the speech and noise stimuli have been presented in detail previously12,31,69. Briefly, all speech stimuli were recorded from a female talker of standard American English. The speech stimuli consisted of monosyllabic digits 0 to 9 (excluding the disyllabic 7), where the “0” was pronounced “Oh” (/oʊ/). Measured digit durations varied from 434–672 ms (SD 57 ms). The process created 27 unique digit files (9 digits, for each of the three digit positions). The long-term average speech spectrum for all 9 digits was mixed to create spectrally matched noise maskers.
A final step was a stimulus homogenization procedure31,69 that equated the audibility of the digits in noise for each digit and for each position. These steps were necessary because the speech reception threshold (SRT) performance measure of the DIN is based on equal audibility of all digits12.
Digits in Noise Test (DIN)
A customized Matlab program was designed to present the triplet digits in noise in successive trials, enabling the estimation of SRT, defined as the SNR yielding 50% correct identification for each set of three digits69 A graphical user interface (GUI) resembling a telephone touch key pad (i.e., 3 rows and 3 columns for digits 1 to 9 and bottom middle for the 0 digit; see Fig. 1a in Dimitrijevic and colleagues12) was incorporated for user response after stimulus presentation. The user initiated the beginning of the test and heard the noise masker, then the carrier phrase “The numbers”. The first digit occurred 1.5 seconds later, followed by the second and third digits (ISI = 1195 ms; onset to onset). The noise masker was turned off 1.5 seconds after the offset of the third digit. The entire stimulus (noise and masker) lasted 6 seconds. The GUI then allowed the participant to indicate which digits were heard. Signal to noise ratio (SNR) of successive triplets was varied adaptively from an initial level of +2 dB. Trials following a correct response (all three digits) had a reduced SNR by 2 dB (noise constant, digit amplitudes reduced). Incorrect responses were followed by an increased SNR, also by 2 dB. Twenty-five trials were presented and the average SNR over the last 11 trials was the SRT. All sounds were presented through a single loudspeaker at 0° azimuth 1.5 meters in front of the subject.
Electrophysiological recordings were performed after the behavioral SRT determination. EEG recordings used the same stimuli as behavioral testing, except that no introductory phrase was used and all trials had the same SNR. Two listening tasks were used; ‘passive’ listening, where the participants were instructed to ignore any sounds while they watched a closed caption and silent movie of their choice, and ‘attentive’ listening, where participants fixated a white cross on an otherwise blank computer screen. In the attentive listening task, rather than responding with the GUI, the participant verbally reported heard digits. The experimenter noted the perceived digits and then initiated the next trial. Signal SNR was initially set to the previously measured SRT (performance close to 50% for all three digits, see below). However, pilot behavioral data showed that participants perform better (lower SRTs) with repeated testing, as previously reported for the DIN, but only between the first and second blocks of trials30. Because we aimed to have roughly an equal number of correct and incorrectly identified trials (100 each) during EEG testing, we adopted an adaptive threshold approach, as above, using 2 blocks of 25 trials (50 trials), but with a starting SNR 2 dB below the previous behavioral SRT. Recordings were performed in 8 blocks (4 active and 4 passive) of 25 trials yielding 200 total trials. Participants took short breaks after each block. The attentive task always occurred first, followed by the passive listening task. The stimuli presented in the passive blocks were identical to that of the active blocks.
The electrophysiological data were collected using a 64-channel actiCHamp Brain Products recording system (Brain Products GmbH, Inc., Munich, Germany). An electrode cap was placed on the scalp with electrodes placed at equidistant locations70. The infracerebral cap used covers a larger area than is typical in a 10–20 system. The reference channel was located at vertex (Cz) while the ground electrode was located on the midline 50% of the distance to nasion. Continuous data were digitized at 1000 Hz and stored for offline analysis.
Self-report measures of listening effort
After each recording block CI users were asked to rate listening effort on a scale of 1–10 using the NASA Task Load Index35. Specifically, the participants indicated their effort level on a 10 point scale to the question: “How hard did you have to listen to accomplish your level of performance in that block?”, the same approach has been used in an earlier EEG study of listening effort10. Listening effort was not measured in the first two participants, leaving a sample size of 8.
The electrophysiological data were first processed using Brain Vision Analyzer ver. 2.0 (Brain Products GmbH, Inc., Munich, Germany). Data were high-pass filtered (0.1 Hz) to remove baseline drifts and down sampled to 250 Hz. Visual inspection and manual sorting of the data included removal of extreme stereotypical artifacts related to subject movement (exceeding 500 mV). Independent component analysis (ICA), as implemented in Brain Vision Analyzer (identical algorithm to EEGLAB71), was applied to reduce ocular and cardiac artifacts. Cochlear implant artifacts were identified by observing components with a centroid of activation on the CI side and the time course of ICA activity as we have described previously72.
Time-frequency analysis: Data were average referenced and segmented into epochs −1500 to 7000 ms relative to speech masker onset. All time-frequency analyses were performed in BESA 6.0 (Brain Electrical Source Analysis, GmbH, Germany) using 2 Hz frequency resolution across the epoch. BESA uses a two-step complex demodulation for time-frequency analysis described in detail in Hoechstetter and colleagues37. Changes in spectral power were quantified as a percent change from baseline (post-stimulus – pre-stimulus/pre-stimulus) x 100.
Brain source analysis: After the time-frequency analysis, a DICS beamformer34, as implemented in BESA was applied to a time-frequency region of interest. The choice of the time-frequency region of interest was based on the condition specific grand mean time-frequency analysis. The BESA DICS beamformer was implemented for the speech-brain coherence where phase coherence between a reference signal (low-pass filtered triple digit speech signal) and brain source were examined across trials in the 2–5 Hz range. A similar approach has been previously described in participants listening to vocoded speech73.
Statistical analysis: All statistical analyses were performed in BESA Statistics 2.0 in a similar manner to that previously described70. Differences between conditions were assessed by performing a paired t-test in source space and then corrected for multiple comparisons using Monte-Carlo resampling techniques36. Clusters of voxels with p-values of less than 0.05 were considered significant. Correlations between brain activity and measures of listening effort were implemented in the “Correlation” option in BESA Statistics 2.0. This procedure provides a correlational measure of brain activity associated with behavioral measure (e.g., listening effort). The source activity (e.g., alpha ERS/ERD over a particular time window) in each participant was performed. A single behavioral measure was then used to correlate brain activity in each voxel across all participants. This process yields a correlation value for each voxel. Corrections for multiple comparisons using Monte-Carlo resampling techniques36.
McGarrigle, R. et al. Listening effort and fatigue: what exactly are we measuring? A British Society of Audiology Cognition in Hearing Special Interest Group ‘white paper’. Int. J. Audiol. 53, 433–40 (2014).
Pichora-Fuller, M. K. et al. Hearing Impairment and Cognitive Energy: The Framework for Understanding Effortful Listening (FUEL). Ear Hear. 37, 5S–27S (2016).
Winn, M. Rapid Release From Listening Effort Resulting From Semantic Context, and Effects of Spectral Degradation and Cochlear Implants. Trends Hear. 20, 1–17 (2016).
Hughes, K. C. & Galvin, K. L. Measuring listening effort expended by adolescents and young adults with unilateral or bilateral cochlear implants or normal hearing. Cochlear Implants Int. 14, 121–129 (2013).
Alhanbali, S., Dawes, P., Lloyd, S. & Munro, K. J. Self-Reported Listening-Related Effort and Fatigue in Hearing-Impaired Adults. Ear Hear. 38, e39–e48 (2016).
Hétu, R., Riverin, L., Lalande, N., Getty, L. & St-Cyr, C. Qualitative analysis of the handicap associated with occupational hearing loss. Br. J. Audiol. 22, 251–64 (1988).
Gagné, J. P., Besser, J. & Lemke, U. Behavioral assessment of listening effort using a dual-task paradigm: A review. Trends Hear. 21, 1–25 (2017).
Alain, C., Du, Y., Bernstein, L. J., Barten, T. & Banai, K. Listening under difficult conditions: An activation likelihood estimation meta-analysis. Hum. Brain Mapp. 2695–2709, https://doi.org/10.1002/hbm.24031 (2018).
Price, C. J. A review and synthesis of the first 20years of PET and fMRI studies of heard speech, spoken language and reading. Neuroimage 62, 816–847 (2012).
Wisniewski, M. G. et al. Frontal midline θ power as an index of listening effort. Neuroreport 26, 94–99 (2015).
Wöstmann, M., Herrmann, B., Wilsch, A. & Obleser, X. J. Neural Alpha Dynamics in Younger and Older Listeners. J. Neurosci. 35, 1458–1467 (2015).
Dimitrijevic, A., Smith, M. L., Kadis, D. S. & Moore, D. R. Cortical Alpha Oscillations Predict Speech Intelligibility. Front. Hum. Neurosci. 11, 88 (2017).
McMahon, C. M. et al. Monitoring Alpha Oscillations and Pupil Dilation across a Performance-Intensity Function. Front. Psychol. 7, 745 (2016).
Strauß, A., Wöstmann, M. & Obleser, J. Cortical alpha oscillations as a tool for auditory selective inhibition. Front. Hum. Neurosci. 8, 350 (2014).
Picton, T. W., John, M. S., Dimitrijevic, A. & Purcell, D. Human auditory steady-state responses. Int. J. Audiol. 42, 177–219 (2003).
Bleichner, M. G., Mirkovic, B. & Debener, S. Identifying auditory attention with ear-EEG: cEEGrid versus high-density cap-EEG comparison. J. Neural Eng. 13, 066004 (2016).
Ding, N. & Simon, J. Z. Cortical entrainment to continuous speech: functional roles and interpretations. Front. Hum. Neurosci. 8, 311 (2014).
Zion Golumbic, E. M. et al. Mechanisms underlying selective neuronal tracking of attended speech at a ‘cocktail party’. Neuron 77, 980–991 (2013).
Aiken, S. J. & Picton, T. W. Human cortical responses to the speech envelope. Ear Hear. 29, 139–157 (2008).
Di Liberto, G. M., O’Sullivan, J. A. & Lalor, E. C. Low-frequency cortical entrainment to speech reflects phoneme-level processing. Curr. Biol. 25, 2457–2465 (2015).
Ding, N. & Simon, J. Z. Neural coding of continuous speech in auditory cortex during monaural and dichotic listening. J. Neurophysiol. 107, 78–89 (2012).
Petersen, E. B., Wöstmann, M., Obleser, J. & Lunner, T. Neural tracking of attended versus ignored speech is differentially affected by hearing loss. J. Neurophysiol. 117, 18–27 (2017).
Horton, C., D’Zmura, M. & Srinivasan, R. Suppression of competing speech through entrainment of cortical oscillations. J. Neurophysiol. 109, 3082–3093 (2013).
Zeng, F. Trends in Cochlear Implants. Trends Amplif. 8, 1–34 (2004).
Nelson, D. A., Donaldson, G. S. & Kreft, H. Forward-masked spatial tuning curves in cochlear implant users. J. Acoust. Soc. Am. 123, 1522–43 (2008).
Won, J. H., Drennan, W. R. & Rubinstein, J. T. Spectral-ripple resolution correlates with speech reception in noise in cochlear implant users. J. Assoc. Res. Otolaryngol. 8, 384–92 (2007).
Gaudrain, E., Grimault, N., Healy, E. W. & Béra, J. C. Effect of spectral smearing on the perceptual segregation of vowel sequences. Hear. Res. 231, 32–41 (2007).
Fu, Q.-J. Temporal processing and speech recognition in cochlear implant users. Neuroreport 13, 1635–9 (2002).
Park, M.-H., Won, J. H., Horn, D. L. & Rubinstein, J. T. Acoustic temporal modulation detection in normal-hearing and cochlear implanted listeners: effects of hearing mechanism and development. J. Assoc. Res. Otolaryngol. 16, 389–99 (2015).
Smits, C., Goverts, T. & Festen, J. M. The digits-in-noise test: assessing auditory speech recognition abilities in noise. J. Acoust. Soc. Am. 133, 1693–706 (2013).
Vlaming, M. S. M. G., MacKinnon, R. C., Jansen, M. & Moore, D. R. Automated screening for high-frequency hearing loss. Ear Hear. 35, 667–79 (2014).
Mattys, S. L., Davis, M. H., Bradlow, A. R. & Scott, S. K. Speech recognition in adverse conditions: A review. Lang. Cogn. Process. 27, 953–978 (2012).
Pfurtscheller, G. & Lopes da Silva, F. H. Event-related EEG/MEG synchronization and desynchronization: basic principles. Clin Neurophysiol 110, 1842–1857 (1999).
Gross, J. et al. Dynamic imaging of coherent sources: Studying neural interactions in the human brain. Proc. Natl. Acad. Sci. USA 98, 694–9 (2001).
Hart, S. & Staveland, L. Development of NASA-TLX (Task Load Index). Adv. Psychol. 52, 139–183 (1988).
Maris, E. & Oostenveld, R. Nonparametric statistical testing of EEG- and MEG-data. J. Neurosci. Methods 164, 177–90 (2007).
Hoechstetter, K. et al. BESA source coherence: a new method to study cortical oscillatory coupling. Brain Topogr. 16, 233–8 (2004).
Rosen, S. Temporal information in speech: acoustic, auditory and linguistic aspects. Philos. Trans. R. Soc. Lond. B. Biol. Sci. 336, 367–73 (1992).
Wild, C. J. et al. Effortful Listening: The Processing of Degraded Speech Depends Critically on Attention. J. Neurosci. 32, 14010–14021 (2012).
Shergill, S. S. et al. Modulation of activity in temporal cortex during generation of inner speech. Hum. Brain Mapp. 16, 219–227 (2002).
Hsieh, L., Gandour, J., Wong, D. & Hutchins, G. D. Functional heterogeneity of inferior frontal gyrus is shaped by linguistic experience. Brain Lang. 76, 227–252 (2001).
Jensen, O. & Mazaheri, A. Shaping functional architecture by oscillatory alpha activity: gating by inhibition. Front. Hum. Neurosci. 4, 186 (2010).
Kerlin, J. R., Shahin, A. J. & Miller, L. M. Attentional gain control of ongoing cortical speech representations in a ‘cocktail party’. J. Neurosci. 30, 620–8 (2010).
Thorpe, S., D’Zmura, M. & Srinivasan, R. Lateralization of frequency-specific networks for covert spatial attention to auditory stimuli. Brain Topogr. 25, 39–54 (2012).
Banerjee, S., Snyder, A. C., Molholm, S. & Foxe, J. J. Oscillatory alpha-band mechanisms and the deployment of spatial attention to anticipated auditory and visual target locations: supramodal or sensory-specific control mechanisms? J. Neurosci. 31, 9923–32 (2011).
Weisz, N., Hartmann, T., Müller, N., Lorenz, I. & Obleser, J. Alpha rhythms in audition: Cognitive and clinical perspectives. Front. Psychol. 2, 1–15 (2011).
Huang, S., Chang, W. T., Belliveau, J. W., Hämäläinen, M. & Ahveninen, J. Lateralized parietotemporal oscillatory phase synchronization during auditory selective attention. Neuroimage 86, 461–469 (2014).
Frey, J. N. et al. Selective modulation of auditory cortical alpha activity in an audiovisual spatial attention task. J. Neurosci. 34, 6634–9 (2014).
Tune, S., Wöstmann, M. & Obleser, J. Probing the limits of alpha power lateralizationas a neural marker of selective attention in middle-aged and older listeners. Eur. J. Neurosci. 0–1, https://doi.org/10.1111/ejn.13862 (2018).
Foxe, J. J. & Snyder, A. C. The Role of Alpha-Band Brain Oscillations as a Sensory Suppression Mechanism during Selective Attention. Front. Psychol. 2, 154 (2011).
Haegens, S., Händel, B. F. & Jensen, O. Top-Down Controlled Alpha Band Activity in Somatosensory Areas Determines Behavioral Performance in a Discrimination Task. J. Neurosci. 31, 5197–5204 (2011).
Mazaheri, A. et al. Region-specific modulations in oscillatory alpha activity serve to facilitate processing in the visual and auditory modalities. Neuroimage 87, 356–62 (2014).
Wostmann, M., Lim, S.-J. & Obleser, J. The Human Neural Alpha Response to Speech is a Proxy of Attentional Control. Cereb. Cortex 27, 3307–3317 (2017).
Du, Y., Buchsbaum, B. R., Grady, C. L. & Alain, C. Noise differentially impacts phoneme representations in the auditory and speech motor systems. Proc. Natl. Acad. Sci. 111, 7126–7131 (2014).
Miles, K. et al. Objective Assessment of Listening Effort: Coregistration of Pupillometry and EEG. Trends Hear. 21, 1–13 (2017).
Kadis, D. S. et al. Characterizing the normal developmental trajectory of expressive language lateralization using magnetoencephalography. J. Int. Neuropsychol. Soc. 17, 896–904 (2011).
Youssofzadeh, V., Williamson, B. J. & Kadis, D. S. Mapping Critical Language Sites in Children Performing Verb Generation: Whole-Brain Connectivity and Graph Theoretical Analysis in MEG. Front. Hum. Neurosci. 11, 173 (2017).
Fu, Q.-J. & Nogaki, G. Noise susceptibility of cochlear implant users: the role of spectral resolution and smearing. J. Assoc. Res. Otolaryngol. 6, 19–27 (2005).
Vanthornhout, J., Decruy, L., Wouters, J., Simon, J. Z. & Francart, T. Speech Intelligibility Predicted from Neural Entrainment of the Speech Envelope. JARO - J. Assoc. Res. Otolaryngol. 19, 181–191 (2018).
Krueger, M. et al. Relation Between Listening Effort and Speech Intelligibility in Noise. Am. J. Audiol. 26, 378–393 (2017).
Park, H., Ince, R. A. A., Schyns, P. G., Thut, G. & Gross, J. Frontal Top-Down Signals Increase Coupling of Auditory Low-Frequency Oscillations to Continuous Speech in Human Listeners. Curr. Biol. 25, 1649–1653 (2015).
Bidelman, G. M., Katherine, M. & Pridgen, M. H. Brainstem-cortical functional connectivity for speech is differentially challenged by noise and reverberation. Hear. Res. 367, 149–160 (2018).
Assaneo, M. F. et al. Spontaneous synchronization to speech reveals neural mechanisms facilitating language learning. Nat. Neurosci., https://doi.org/10.1038/s41593-019-0353-z (2019).
Hervais-Adelman, A. G., Carlyon, R. P., Johnsrude, I. S. & Davis, M. H. Brain regions recruited for the effortful comprehension of noise-vocoded words. Lang. Cogn. Process. 27, 1145–1166 (2012).
Eisner, F., McGettigan, C., Faulkner, A., Rosen, S. & Scott, S. K. Inferior frontal gyrus activation predicts individual differences in perceptual learning of cochlear-implant simulations. J. Neurosci. 30, 7179–86 (2010).
Giraud, A. L. Contributions of Sensory Input, Auditory Search and Verbal Comprehension to Cortical Activity during Speech Processing. Cereb. Cortex 14, 247–255 (2004).
Khoshkhoo, S., Leonard, M. K., Mesgarani, N. & Chang, E. F. Neural correlates of sine-wave speech intelligibility in human frontal and temporal cortex. Brain Lang. 0–1, https://doi.org/10.1016/j.bandl.2018.01.007 (2018).
Fiedler, L., Obleser, J., Lunner, T. & Graversen, C. Ear-EEG allows extraction of neural responses in challenging listening scenarios - A future technology for hearing aids? Proc. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. EMBS 2016-Octob, 5697–5700 (2016).
Smits, C., Kapteyn, T. S. & Houtgast, T. Development and validation of an automatic speech-in-noise screening test by telephone. Int. J. Audiol. 43, 15–28 (2004).
Han, J. H. & Dimitrijevic, A. Acoustic change responses to amplitude modulation: a method to quantify cortical temporal processing and hemispheric asymmetry. Front. Neurosci. 9, 1–15 (2015).
Delorme, A. & Makeig, S. EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J Neurosci Methods 134, 9–21 (2004).
Dimitrijevic, A., Pratt, H. & Starr, A. Auditory cortical activity in normal hearing subjects to consonant vowels presented in quiet and in noise. Clin. Neurophysiol. 124, 1204–15 (2013).
Peelle, J. E., Gross, J. & Davis, M. H. Phase-locked responses to speech in human auditory cortex are enhanced during comprehension. Cereb. Cortex 23, 1378–87 (2013).
The authors would like to thank the anonymous reviewers for their very helpful and insightful comments. David R. Moore is supported by the NIHR Manchester Biomedical Research Centre. D.M. and A.D. were also partially supported by an NIH grant R01DC014078.
The authors declare no competing interests.
Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Dimitrijevic, A., Smith, M.L., Kadis, D.S. et al. Neural indices of listening effort in noisy environments. Sci Rep 9, 11278 (2019) doi:10.1038/s41598-019-47643-1