Introduction

Language is abstract and arbitrary, while even single words differ in their emotional quality. Event-related potential (ERP) studies show that our brains differentiate between negative and neutral words, even when emotion is not directly relevant to the experiment1,2,3,4,5,6. However, it has been recently shown that emotional information processing crucially depends on the focus of attention, enabling a dissociation of ERP unfolding across the time course of stimulus processing.

Different components of the ERP are associated with different stages of information processing. They can broadly be distinguished in rather early components (such as the P1 and N1), mid-latency (such as the P2, Early Posterior Negativity, EPN), and late components (such as the Late Positive Potential, LPP). Among early ERPs, the P1 (~ 80 to 100 ms post-stimulus) is an occipitally scored positivity, followed by the N1 (~ 120 to 170 ms) as occipito-temporal negativity. Both components reflect the early stages of stimulus processing7,8. They are related to early stimulus gain processes9,10, and are strongly influenced by low-level visual information11,12,13. Concerning mid-latency ERPs, the P2 component is a positive polarization and peaks at around 200 ms, with a more variable distribution, sometimes distinguishing an anterior and posterior P2 component14. The anterior P2 has been more closely related to exogenous attention14,15. Word repetition has been shown to increase the P216. The early posterior negativity (EPN) is observed as differential occipital-temporal negativity when contrasting emotional and neutral stimuli and is typically observed between 200 and 300 ms. The EPN is related to early emotional tagging and attention processes toward relevant information2,17,18. Among late ERPs, the late positive potential (LPP) is part of the family of late positivities, emerging from approximately 400 ms onwards up to seconds after stimulus appearance. It is identified by contrasting emotional and neutral stimuli and indicates stimulus evaluation and controlled attention processes, with a centro-parietal topographic distribution, but this varies and seems to depend on stimulus types, and tasks19,20.

Concerning early emotion effects and their modulation by attention asks, effects are mixed for the P1. Some studies found larger amplitudes for negative compared to neutral words21, with some only in males22 samples or the left hemisphere23, while other studies reported decreased P1 amplitudes for negative compared to neutral words, depending on target relevance or word frequency3,24, and other studies did not find differences4,12,25,26,27,28,29,30,31. No differential effects concerning negative valence were observed between an emotional vs. color judgment task4, when attending to the lexical or emotional information26, when performing a lexical decision task rather than reading29, or when instructed to inhibit a word32. For the N1, several studies reported larger N1 amplitudes for negative than neutral words33,34, sometimes restricted in the left hemisphere17,23. Nevertheless, other studies found emotion effects depending on word frequency3 or target status24. Some studies reported effects restricted to the right hemisphere, or only in response to positive words26, or the absence of effects12,25,30,35. The N1 to negative as compared to neutral words was not modulated by emotional vs. color judgment tasks4,36,37, or when attending to lexical vs. emotional information26,29. For the following P2, increased P2 amplitudes for negative words have been observed31,37,38, sometimes being left- or right-lateralized15,36, or descriptively larger for concrete negative words39. Other studies reported effects only for positive (concrete) words or no effects for negative words1,40. Thus, early emotion effects differ considerably between studies and may reflect differences in the used languages, specific stimulus sets, and attention tasks. Systematic studies with large sample sizes, well-controlled stimuli, and variation of task conditions are strongly needed.

More reliably, effects for emotional words are found during mid-latency and late processing stages. The EPN arises at about 200 ms and is related to early lexical access17, perceptual tagging2, and attention processes41. The LPP occurs from about 400 ms after presenting a word and reflects later stages of attention, stimulus evaluation, and episodic memory encoding25,33,42. Nevertheless, several studies showed for negative words either no EPN43,44,45 or no LPP effects26,27,35. Effects seem to depend on attentional conditions. For example, EPN effects were present for tasks requiring emotional judgments but not when attending to the color of stimuli4. Hinojosa et al.44 showed a similar EPN pattern, although not significant for negative words. Following these studies, attention to the semantic meaning seems necessary to elicit mid-latency emotion effects in words. Hinojosa and colleagues44 also showed LPP effects for negative words when participants had to identify a word among pseudowords (i.e., had to attend to the meaning) but not when words had to be identified among non-recognizable stimuli. In this regard, late emotion effects for negative words were absent in several studies during structural (font consistency)46, color37, lexical3, or semantic26 tasks. Further, for the LPP, a study reported increasing effects when negative words were target-relevant as compared to neutral words47, but see48. This pattern of findings would be in line with a recent meta-analysis across different visual stimuli (with smaller effect sizes for word stimuli) that reported no reliable late amplitude effects during non-emotional tasks (e.g., watching, reading, or classification according to non-emotional attributes), but reliable effects during explicit emotion decision tasks49. However, as pointed out above, there are several conflicting findings, mostly due to studies that report late emotion effects during color, lexical, or semantic tasks1,4,40,46,50. This might be explained by attentional spillover to task-irrelevant word features.

To reduce the variability in experimental conditions and to better differentiate processing stages during emotion processing, we recently developed a design that systematically varied feature-based attention to emotional visual stimuli51. Here, participants pay attention to a stimulus-unrelated feature (e.g., overlaid thin lines), to the stimulus (e.g., specific emotion irrelevant stimulus features), or to the emotional meaning (e.g., negative or neutral content). Studies using faces or complex scenes showed dissociable modulations of the EPN and LPP across attention tasks51,52,53. Emotional EPN effects were absent in the perceptual but present in the other tasks, while LPP differences were only present when attention was directed to the emotional information51,52,53. For pictures and faces, increased N1/N170 responses were found regardless of task51,52,53. Thus, this kind of task allows the separation of more automatic (early) processing stages from subsequent mid-latency and late stages, which require sufficient attentional resources or task relevance during the processing of emotional stimuli. Importantly, the dissociation between ERP components requires brief presentation times to avoid attentional spillover to task-irrelevant stimulus features54. It remains an open question whether a similar dissociation of processing stages can also be revealed while processing negative vs. neutral words.

Different stages of emotion processing are dissociated by tasks that systematically increase attention to emotionally relevant stimulus features. To test how modulations depend on the given attended feature in word stimuli, we used three different tasks during which negative and neutral words (adjectives and nouns) were presented and examined differential responses across the whole processing stream (P1, N1, P2, EPN, and LPP). Participants had to decide (1) if the overlaid line orientation was either horizontal or vertical, (2) if the word was a noun or an adjective, or (3) if the word valence was negative or neutral. We expected that the later the component of the ERP, the stronger the increasing attention to emotionally relevant information would increase emotion effects (for the detailed registration, see https://osf.io/nrmsb). We explored emotion effects and interactions with the task for earlier ERPs (P1, N1, and P2). Concerning registered effects, we expected the EPN to increase amplitudes for negative words in the grammatical (adjective/noun) and emotion decision tasks compared to the perceptual task. Finally, LPP emotion effects should be increased in the emotion decision task compared to grammatical and perceptual decision tasks.

Results

Behavior

Regarding hit rate, the number of correct choices was not affected by emotion (F(1,39) = 1.37, p = 0.250, ηP2 = 0.034), but by task (F(2,78) = 55.55, p < 0.001, ηP2 = 0.588), with higher accuracy for the perceptual compared to both the grammatical (p = 0.048) and the emotion task (p < 0.001). The grammatical task also elicited higher accuracy than the emotion task (p < 0.001). Emotion and task did not interact (F(1.30,50.56) = 2.49, p = 0.113, ηP2 = 0.060; see Table 1). Regarding reaction times, main effects of emotion (F(1,39) = 15.83, p < 0.001, ηP2 = 0.289) and of task were identified (F(2,78) = 91.91, p < 0.001, ηP2 = 0.702), but no significant emotion by task interaction (F(1.32,51.40) = 2.43, p = 0.117, partial η2 = 0.059). Reaction times were significantly shorter for negative than neutral words (p < 0.001) and shorter for the perceptual compared to both the grammatical and compared to the emotion task (ps < 0.001). The grammatical task also elicited shorter reaction times than the emotion task (p < 0.001).

Table 1 Behavioral results across the three attention tasks.

Event-related potentials

For mean amplitudes of all examined ERPs, see Table 2 below. For hemisphere effects for the P1, N1, and EPN, see respective control analyses below in "Analyses of hemispheric differences in early emotion effects" section.

Table 2 Mean amplitudes for all ERPs across the three attention tasks.

P1 (90–110 ms)

For the P1, there were no main effects of emotion (F(1,39) = 0.16, p = 0.696, ηP2 = 0.004, see Fig. 1), and of task (F(2,78) = 1.20, p = 0.306, ηP2 = 0.030), and no interaction of emotion and task (F(2,78) = 0.74, p = 0.483, ηP2 = 0.019).

Figure 1
figure 1

Occipital cluster showing P1 effects. (a) Scalp topographies depict the differences between negative and neutral words. (b) ERP waveforms show the time course over highlighted sensors. For bar plots, error bars show 95% confidence intervals for amplitudes averaged across selected sensors. Lines connect individual data points. (c) Respective difference plots displayed below contain 95% bootstrap confidence intervals of intra-individual differences.

N1 (140–180 ms)

For the N1, there was no main effect of emotion (F(1,39) = 1.09, p = 0.303, ηP2 = 0.027). There was a main effect of task (F(2,78) = 12.63, p < 0.001, ηP2 = 0.245), with larger N1 amplitudes for both the emotion and grammatical task compared to the perceptual task (ps < 0.001), the former two conditions not differing from each other (p = 0.425). There was no interaction of emotion and task (F(2,78) = 1.96, p = 0.148, ηP2 = 0.048).

Early posterior negativity (200–350 ms)

For the EPN, there was no main effect of emotion (F(1,39) = 3.24, p = 0.080, ηP2 = 0.077), and no main effect of task (F(2,78) = 1.39, p = 0.256, ηP2 = 0.034). There was an interaction of emotion and task (F(2,78) = 5.07, p = 0.008, ηP2 = 0.115, see Fig. 2). Importantly, emotion differences were larger for the emotion compared to the perceptual task (Mdifference =  − 0.36, SD = 0.90, t(39) =  − 2.54, p = 0.015), and for the grammatical compared to the perceptual task (Mdifference =  − 0.44, SD = 0.94, t(39) =  − 2.96, p = 0.005). Negative-neutral differences did not differ between the emotion and grammatical tasks (Mdifference = 0.08, SD = 0.96, t(39) = 0.54, p = 0.595).

Figure 2
figure 2

Occipito-temporal cluster showing N1 and EPN effects. (a) Scalp topographies depict the differences between negative and neutral words. (b) ERP waveforms show the time course over highlighted sensors. For bar plots, error bars show 95% confidence intervals for amplitudes averaged across selected sensors. Lines connect individual data points. (c) Respective difference plots displayed below contain 95% bootstrap confidence intervals of intra-individual differences.

P2 (150–200 ms)

Similar to N1 effects, the central P2 was not affected by emotion (F(1,39) = 0.06, p = 0.812, ηP2 = 0.001), but by task (F(2,78) = 7.53, p < 0.001, ηP2 = 0.162). There were larger P2 amplitudes for both the emotion and grammatical task compared to the perceptual task (ps < 0.01), the former two conditions not differing from each other (p = 0.463). There was no interaction of emotion and task (F(2,78) = 2.43, p = 0.095, ηP2 = 0.059).

Late positive potential (380–800 ms)

For the LPP, there was a main effect of emotion (F(1,39) = 11.38, p = 0.002, ηP2 = 0.226), with a larger positivity for negative than neutral words. There was no effect of task (F(2,78) = 2.09, p = 0.131, ηP2 = 0.051). There was an interaction of emotion and task (F(2,78) = 6.55, p = 0.002, ηP2 = 0.144, see Fig. 3). Importantly, emotion differences were larger for the emotion compared to the perceptual task (Mdifference = 0.68, SD = 1.27, t(39) = 3.38, p = 0.002), and for the emotion compared to the grammatical task (Mdifference = 0.45, SD = 1.04, t(39) = 2.74, p = 0.009). Negative-neutral differences did not differ between the grammatical and perceptual tasks (Mdifference = 0.23, SD = 1.30, t(39) = 1.12, p = 0.270).

Figure 3
figure 3

Central cluster showing P2 and LPP effects. (a) Scalp topographies depict the differences between negative and neutral words. (b) ERP waveforms show the time course over highlighted sensors. For bar plots, error bars show 95% confidence intervals for amplitudes averaged across selected sensors. Lines connect individual data points. (c) Respective difference plots displayed below contain 95% bootstrap confidence intervals of intra-individual differences.

Analyses of hemispheric differences in early emotion effects

Hemispheric differences for P1 effects

For the P1, a main effect hemisphere was found with larger P1 amplitudes over right compared to left sensors (F(1,39) = 7.42, p = 0.010, ηP2 = 0.160). Hemisphere interacted with emotion effects (F(1,39) = 6.85, p = 0.013, ηP2 = 0.149, see Fig. 4). Post-hoc tests showed no significant differences over left sensors (MDifference =  − 0.167, SEM = 0.107, p = 0.127) but significant differences over right sensors (MDifference = 0.223, SEM = 0.098, p = 0.029), with a larger P1 amplitude for negative words. There was no three-way interaction between hemisphere, emotion, and attention task (F(2,78) = 1.71, p = 0.186, ηP2 = 0.042).

Figure 4
figure 4

Occipito-temporal clusters, showing P1, N1, and EPN emotion effects across tasks for the left and right hemispheres. (a) and (d) Scalp topographies depict the differences between negative and neutral words. (b) and (e) ERP waveforms show the time course over highlighted sensors. For bar plots, error bars show 95% confidence intervals for amplitudes averaged across selected sensors. Lines connect individual data points. (c) and (f) Respective difference plots displayed below contain 95% bootstrap confidence intervals of intra-individual differences.

Hemispheric differences for N1 effects

For the N1, a main effect hemisphere was found with larger N1 amplitudes over left compared to right sensors (F(1,39) = 13.39, p < 0.001, ηP2 = 0.256). Hemisphere did not interact with emotion effects (F(1,39) = 0.39, p = 0.536, ηP2 = 0.010). There was no three-way interaction between hemisphere, emotion, and attention task (F(2,78) = 1.15, p = 0.323, ηP2 = 0.029).

Hemispheric differences for EPN effects

For the EPN, a main effect hemisphere was found with larger EPN amplitudes over left compared to right sensors (F(1,39) = 33.11, p < 0.001, ηP2 = 0.459). Hemisphere interacted with emotion effects (F(1,39) = 9.31, p = 0.004, ηP2 = 0.193, see Fig. 4). Post-hoc tests showed significant differences over left sensors (MDifference =  − 0.279, SEM = 0.090, p = 0.004), with larger EPN amplitude for negative words, but no significant differences over right sensors (MDifference = 0.065, SEM = 0.073, p = 0.379). There was no three-way interaction between hemisphere, emotion, and attention task (F(2,78) = 2.04, p = 0.137, ηP2 = 0.050).

Discussion

This study investigated how different tasks affect different ERP components to negative vs. neutral words to test whether the dissociation of early, mid-latency, and late emotional differentiation depends on the attended feature of the stimulus. The tasks varied the attentional focus on word-irrelevant features, emotion-irrelevant aspects, or emotional meaning of words. Our findings reveal a systematic pattern of emotional sensitivity varying with the temporal hierarchy of different ERPs, comparable to those observed for other visual, emotional stimuli. We observed interaction effects between emotion and task for the EPN, and LPP, showing an increase of EPN emotion differences during the grammatical and emotion task, while LPP emotion differences were restricted to the emotion task. For the P1, N1, and P2 we observed no main effects of emotion and no interactions between emotion and task, while unregistered control analyses showed reversed P1 emotion effects over right versus left sensors.

These right-lateralized P1 effects should be interpreted cautiously since other studies observed the opposite pattern23, and most studies do not observe P1 increases for negative words12,25,27,28,30,31, or effects of attention tasks4,26,29. Taking these considerations into account, we did not find reliable emotion effects on early ERPs. Besides the P1, this concerns the N1. For the N1, effects in previous studies are mixed, with emotion effects in some studies3,24,33, but not in others12,25,30,35. The conflicting emotion effects across studies may be due to a combination of specific stimuli, task parameters, variable effects in smaller samples, or individual differences, such as differences in morphosyntactic processing34. Our study focused on the effects of negative vs. neutral stimuli. Thus, it remains an open question whether there might be findings for positive stimuli, also applying to the findings concerning the P2 component1,15,29,55. While we did not observe early emotion effects with a typical word set used in the field, we do not rule out that specific word by emotion stimulus conditions (e.g., word frequency, word length, concreteness, stimulus presentation duration) might exist, which should be addressed in future high powered studies12. Furthermore, early effects might be evident in other analytical EEG/ERP data approaches.

We also found no effects for the P2. In the literature, both increased P2 amplitudes for negative words have been observed31,37,38, as well as no differences between negative and neutral words1,40. Emotion effects are often shown to be larger in concrete words39 and have been reported selectively for positive words in some studies1. Thus, similar to the P1 and N1 components, we might have missed specific valence effects in this study.

Concerning the generally increased N1 and P2 amplitudes in the grammatical and emotion task, this could reflect more elaborated processing of the word meaning through top-down instructions56. This would align with the observed larger reaction times in the latter two tasks. Please note that we did not register to interpret these general amplitude changes.

In contrast to the earlier components of the ERP, we formulated preregistered hypotheses regarding the EPN and and the LPP. Findings supported the outlined expectations. Our study found no EPN effects when participants attended to the lines. This might be surprising, given that EPN effects are frequently reported across various tasks2,46,48. We reason that the absence of emotional differentiation is due to combining a brief presentation duration with the attention directed to the line orientation. Longer durations might also enable participants to decode the emotional information. Brief stimulus durations are necessary to ensure the relevant task focus and avoid additional cognitive processes51,52,54. In contrast to the perceptual task, when participants attended to the word meaning or were explicitly asked to evaluate the emotional meaning of the words, larger EPN amplitudes were observed for negative compared to neutral stimuli. Our findings suggest that such differential processing is attenuated or abolished when participants attended to perceptual information, in line with some previous findings4,44, but see45. Additional analyses with the hemisphere as a factor showed a left-lateralization of EPN emotion modulations in line with the literature2,17. The EPN has been suggested to signal early attentional selection41, which typically is increased by emotionally (arousing) stimuli57,58 but also increased for other salient stimuli59. While the EPN was originally thought to be generated in the primary and secondary visual cortex, picture-wise correlation approaches showed stronger correlations with subcortical structures, including the amygdala, anterior cingulate cortex, or the striatum60,61. However, studies using separate or combined EEG/fMRI recordings that enable the localization of EPN generators in word stimuli are missing.

In contrast to earlier ERPs, a differential LPP effect for negative words was only seen during the emotion task. In line with this finding, several studies report late effects during tasks that require the processing of the emotion37,44,46,47,48. Further, late emotion effects were larger when participants attended to the emotion but compared to perceptual37, or semantic (touchable/not touchable) features45. While late effects are generally smaller in verbal stimuli, these are reliable during explicit emotion decision tasks but during non-emotional tasks, such as watching, reading, or classification according to non-emotional attributes49. Thus, these findings support the idea that during the LPP stage, evaluative and controlled attention processes occur20,62. The LPP likely results from multiple and distributed sources63. It has been suggested that these include visual cortices, temporal cortices, the amygdala, the insula, and the orbitofrontal cortex63. Stimulus-specific effects can be expected, as for emotional words, fMRI studies show that different frontal regions (inferior frontal gyrus, dorsomedial prefrontal, and cingulate cortex) are involved39,64. Studies with suited designs are needed to disentangle the differential involvement of brain areas in the generation of the EPN and the LPP.

Our findings further support the general notion that at least the dynamics and the functional significance of EPN and LPP effects are highly similar across different visual stimulus categories, scenes, faces, or words. For reviews, see6,42,65,66,67. In most studies, EPN and LPP effects are highly correlated60, while our attention task manipulation enables a clear dissociation. The findings are remarkably similar to our recent studies with other stimulus categories51,52,53. For example, we found that faces and scenes elicited increased N1/N170 modulations for negative stimuli were task-independent, while an EPN effect was not observed during the perceptual task but found in similar amplitude for the other two tasks51,52. LPP differences were only present when attention was directed to the emotional expression of the face51,52. Thus, a similar task dependency can be shown across different categories of emotional stimuli, at least for the EPN and the LPP.

Limitations and future directions

Concerning our study's findings, some constraints have to be mentioned. We only focused on comparing ERPs to negative and neutral words. Future studies might investigate whether findings depend on valence and/or arousal. While positive words would be interesting to examine, several reasons led to the inclusion of only negative and neutral words. First, we used a design comparable to other recent studies focusing on negative versus neutral stimuli51,52. Secondly, we aimed to have a similar two-forced choice task in all three tasks. An additional differentiation of positive words would likely increase the difficulty of the emotion task. However, future studies might also use other word stimulus sets or systematically vary stimulus features to better understand possible early effects and include positive words. Furthermore, we would like to note that the dissociation between ERP components requires brief presentation times to avoid attentional spillover to task-irrelevant stimulus features. Here, the brief presentation durations ensured that the participants' attention was only directed at a specific task, showing successful modulations of emotional effects across tasks, similar to previous studies. Nevertheless, future studies might use tasks with varying presentation durations to test whether effects differ with longer presentation times54. Finally, we used the collapsed conditions or collapsed differences to identify ERP components of interest. We decided to predefine our ERPs of interest (see methods section) based on studies with similar tasks but different visual stimuli51,52. However, other methods can also be used to identify ERPs without biases and may result in different time windows or sensors.

Conclusion

We found no evidence of early (P1, N1, and P2) ERP differences between negative vs. neutral words across three different attention tasks. However, we observed task-sensitive mid-latency (EPN) and late (LPP) differential processing of emotional words. EPN effects required attention to the word's meaning, while the LPP effect was only seen during the emotional task. These findings reveal a systematic pattern of emotional sensitivity varying with the temporal hierarchy of different components of the ERP, showing the graded increase in processing steps depending on the participants` task set.

Methods

Participants

In total, forty-eight participants enrolled in this study and provided complete data. Eight participants were excluded due to EEG rejection criteria, defined as more than 10 interpolated electrodes or more than 40% of usable trials rejected. All participants provided written informed consent to participate in the study. Participants received 10 Euros per hour for participation. The final sample of forty participants (30 female), exhibited a mean age of 23.58 years (SD = 3.95, Median 23, Min = 18; Max = 35), all had normal or corrected-to-normal vision, were right-handed, native German speakers, with no reported history of psychiatric disorders. We followed the updated data-sampling plan and collected 40 usable datasets (see https://osf.io/nrmsb), based on power calculations from a recent previous study using the same attention tasks for faces51. Our sample size exhibited a power of > 99% to detect the previously observed large effect sizes (ηp2 = 0.149 and 0.155) for the EPN and LPP interactions. Participants performed an unrelated auditory attention task first, and participation in this task was optional. Data were uploaded to the attached OSF project (https://osf.io/eyndu/). The University of Münster medical ethics committee has approved the study protocol. All experiments were performed in accordance with relevant guidelines and regulations of the University of Münster.

Stimuli

The words were taken from previously collected rating datasets, rated by students regarding valence and arousal, and matched for linguistic variables. For the experiment, 60 negative and 60 neutral words (30 nouns, 30 adjectives) were used, differing in valence and arousal (see Table 3). Word length and word frequency strongly affect word processing; shorter and more frequent words are processed more quickly68,69,70,71. Secondly, many (high-frequent) orthographic neighbors have been argued to elicit lateral inhibitory mechanisms at a lexical level72. Lines were overlaid to the words using presentation (www.neurobehavioralsystems.org), showing three horizontal or vertical lines (horizontal lines 2 lengths; vertical lines 0.7 lengths; thickness 0.01; centered around x = 0, y = 0, RGB color words 0,0,0; RGB color lines 47,79,79).

Table 3 Comparisons of negative and neutral words by One-Way-ANOVAs.

Procedure

Participants were instructed to avoid eye-movements and blinks during the stimulus presentation. While participants were prepared for the EEG, they responded to a demographic questionnaire. They started with either the perceptual decision, grammatical decision, or emotion decision task. Each task contained a block of 120 trials, with all 60 negative and 60 neutral words. The trial structure and presentation were identical (see Fig. 5). In each trial, a word was presented for 100 ms. Afterwards, a variable fixation cross was presented for 2300–2500 ms.

Figure 5
figure 5

Experiment and trial overview. (a) Overview of the three attention tasks. (b) Trial structure for each of the three attention tasks. Please note that screen proportions were increased to increase visibility.

Responses were recorded within the first 1500 ms. Task order and response buttons (x and m) were counterbalanced. In each task, participants always had to decide in a two-alternative forced-choice task: (1) if line orientation was horizontal or vertical, or (2) if the word was a noun or adjective, or (3) if the word was negative or neutral. All 60 negative and 60 neutral words were presented in each task, summing up to a total of 360 trials.

EEG recording and preprocessing

EEG signals were recorded from 64 BioSemi active electrodes using Biosemis Actiview software (www.biosemi.com). Four additional electrodes measured horizontal and vertical eye-movement. The recording sampling rate was 512 Hz. Offline data were re-referenced to average reference, and band-pass filtered from 0.01 to 40 Hz. Recorded eye-movement was corrected using the automatic eye-artefact correction method implemented in BESA73. The remaining artifacts were rejected based on an absolute threshold (< 120 µV), signal gradient (< 75 µV/∂T), and low signal (i.e., the SD of the gradient, > 0.01 µV/∂T). Noisy EEG sensors were interpolated using a spline interpolation procedure. A delay of the LCD screen for stimulus presentation of 15 ms, measured by a photodiode, was corrected during epoching. We included only participants with at least 15 correct trials in each condition (see https://osf.io/nrmsb). On average, 47 trials were kept per single condition (Ms = 45 to 49, SDs = 7 to 8; Min = 29 to 32), with no differences between emotion or task conditions and with no interaction (Fs < 2.33, ps > 0.122). Filtered data were segmented from 100 ms before stimulus onset until 1000 ms after stimulus presentation. Baseline correction used the 100 ms before stimulus onset.

Data analyses

All data were statistically analyzed with two (Emotion: negative, neutral) by three (Task: perceptual, grammatical, emotion decision) repeated measure ANOVAs. We investigated the main effects of task and emotion and their interaction. Partial eta-squared (partial η2) were used to describe effect sizes, where ηP2 = 0.02 describes a small, ηP2 = 0.13 a medium and ηP2 = 0.26 a large effect74. Behavioral data were analyzed with JASP (https://jasp-stats.org) for both reaction time and accuracy. We predefined expected time windows and sensor clusters, with a validation based on inspection of the waveforms. Using these time windows and scalp regions as priors, the P1, N1, and P2 were identified using a collapsed localizer across all conditions, for which we used the predefined sensors and time windows. Similarly, negative-neutral differences were used for the EPN and LPP (see Fig. 6 for the ERP identification results). Based on the scalp topography, we visually examined whether sensors adequately capture the collapsed positivity/negativity, where the scalp differences for LPP led to a modification of the included sensors (see below). In a second step, we averaged the ERP waveform to visually judge whether the time windows were symmetrically around the positive or negative peak or captured the emotion differences in cases of the EPN and LPP. This led to slight deviations from the registration for the used time windows and sensors. We identified the P1 from 90 to 110 ms (registered 80–100 ms), the N1 from 140 to 180 ms (registered 140–190 ms), the P2 from 150 to 200 ms (registered), the EPN from 200 to 350 ms (registered), and the LPP from 380 to 800 ms (registered 400–650 ms). We could not clearly identify a centro-parietal P3. We averaged ERPs from all examined sensors in the above-defined time windows. We used occipital sensors for the P1, and occipito-temporal sensors for the N1 and the EPN (P1: P9, P7, PO7, P10, P8, PO8; N1 and EPN: TP7, P9, P7, PO7, TP8, P10, P8, PO8). The central cluster (P2, LPP) was examined over an extended sensor cluster (registered C1, Cz, C2, CP1, CPz, CP2, additionally including F1, Fz, F2, FC1, FCz, FC2, P1, Pz, P2). In addition, unregistered analyses tested for hemispheric differences in early emotion effects and possible lateralized interactions between emotion effects and the attention task (P1, N1, EPN).

Figure 6
figure 6

ERP identification validation. (a) Collapsed conditions to identify the P1, N1, and P2 components. (b) Respective waveforms averaged across highlighted selected sensors. (c) Collapsed task conditions to identify the EPN and LPP components. (d) Respective waveforms averaged across highlighted selected sensors.