United States politicians’ tone became more negative with 2016 primary campaigns

There is a widespread belief that the tone of political debate in the US has become more negative recently, in particular when Donald Trump entered politics. At the same time, there is disagreement as to whether Trump changed or merely continued previous trends. To date, data-driven evidence regarding these questions is scarce, partly due to the difficulty of obtaining a comprehensive, longitudinal record of politicians’ utterances. Here we apply psycholinguistic tools to a novel, comprehensive corpus of 24 million quotes from online news attributed to 18,627 US politicians in order to analyze how the tone of US politicians’ language as reported in online media evolved between 2008 and 2020. We show that, whereas the frequency of negative emotion words had decreased continuously during Obama’s tenure, it suddenly and lastingly increased with the 2016 primary campaigns, by 1.6 pre-campaign standard deviations, or 8% of the pre-campaign mean, in a pattern that emerges across parties. The effect size drops by 40% when omitting Trump’s quotes, and by 50% when averaging over speakers rather than quotes, implying that prominent speakers, and Trump in particular, have disproportionately, though not exclusively, contributed to the rise in negative language. This work provides the first large-scale data-driven evidence of a drastic shift toward a more negative political tone following Trump’s campaign start as a catalyst. The findings have important implications for the debate about the state of US politics.

in recent years".Many see a cause in Donald Trump, who a majority (55%) think "has changed the tone and nature of political debate [. . . ] for the worse", whereas only 24% think he "has changed it for the better" (Pew Research Center 2019).The purpose of the present article is to investigate whether these subjective impressions reflect the true state of US political discourse.The answer to this question comes with tangible societal implications: as politics impacts nearly every aspect of our personal lives (Smith, Hibbing, and Hibbing 2019), changes in political climate can directly affect not only politics itself, but also the well-being of all citizens.
Although longitudinally conducted voter surveys, notably the American National Election Studies, have shown that negative affect towards members of the other party (Abramowitz and Webster 2016;Iyengar, Sood, and Lelkes 2012;Iyengar et al. 2019), polarization (Abramowitz and Saunders 2008), and partisan voting (Bafumi and Shapiro 2009) have steadily increased over the last decades, such survey-based studies do not directly measure the tone of political debate-a linguistic phenomenon-and thus cannot answer whether Americans' impression of increased negativity is accurate.In contrast, we analyze US politicians' language directly, in an objective, data-driven manner, asking: First, is it true that US politicians' tone has become more negative in recent years?Second, if so, did Donald Trump's entering the political arena bring about an abrupt shift (Costello 2016;Crandall, Miller, and White 2018), or did it merely continue a previously existing trend (Fritze and Jackson 2021;James 2021)?
Our methodology draws on a long history of research on the language of politics and its function in democracies (Dunmire 2012;Farrelly 2010;Hidalgo Tenorio 2002;Zheng 2000).For instance, prior work has used records of spoken and written political language to establish the prevalence of negative language among political extremists (Frimer et al. 2019); to quantify growing partisanship and polarization (Gentzkow, Shapiro, and Taddy 2019), as well as displayed happiness (Wojcik et al. 2015), among US Congress members; to analyze political leaders' psychological attributes such as certainty and analytical thinking (Jordan et al. 2019); to quantify the turbulence of Trump's presidency (Dodds et al. 2021); or to measure the effect of linguistic features on the success of US presidential candidates (Thoemmes and Conway III 2007) and on public approval of US Congress (Frimer et al. 2015).A combination of political discourse analysis and psychological measurement tools has further been applied to obtain insights into the personality traits and sentiments of politicians in general (Kangas 2014;Tumasjan et al. 2010), as well Donald Trump in particular (Kreis 2017;Ahmadian, Azarshahi, and Paulhus 2017;Pryck and Gemenne 2017;Lewandowsky, Jetter, and Ecker 2020).
One limiting factor in the above-cited works is the representativeness and completeness of the underlying data, since subtle social or political behavior may only reveal itself in sufficiently big and rigorously processed data (Monroe et al. 2015;Lazer et al. 2021).On the one hand, Congressional records (Gentzkow, Shapiro, and Taddy 2019;Wojcik et al. 2015;Frimer et al. 2015) and transcripts of public speeches (Thoemmes and Conway III 2007;Jordan et al. 2019) record scarce events and do not mirror political discourse as perceived by the average American, whose subjective impression of growing negativity we aim to compare with objective measurements of politicians' language.On the other hand, news text, despite being a better proxy for the average American's exposure to political discourse, for the most part does not capture politicians' utterances directly, but largely reports events (laced with occasional direct quotes) and frequently focuses on a commentator's perspective.Additionally, most of the above-cited linguistic analyses are not longitudinal in nature, but focus on specific points in time.
Transcending these shortcomings, we take a novel approach leveraging Quotebank (Vaucher et al. 2021a), a recently released corpus of nearly a quarter-billion (235 million) unique quotes extracted from 127 million online news articles published by a comprehensive set of online news sources over the course of nearly 12 years (September 2008 to April 2020) and automatically attributed to the speakers who likely uttered them by a machine learning algorithm.By focusing on US politicians, we derived a subset of 24 million quotes by 18,627 speakers, enriched with biographic information from the Wikidata knowledge base (Vrandecic and Krötzsch 2014).(Details about data in Materials and Methods.)As no comparable dataset of speakerattributed quotes was available before, Quotebank enables us to analyze the tone of US politicians' public language (as seen through the lens of online news media) at a level of representativeness and completeness that was previously impossible, without confounding politicians' direct utterances with the surrounding news text.
In order to quantify the prevalence of negative language over time, we use established psycholinguistic tools (Pennebaker, Booth, and Francis 2007) to score each quote with respect to its emotional content, aggregate quotes by month, and work with the resulting time series.Anchored in the average American's subjective perception that Donald Trump has changed the tone and nature of political debate (see above), we hypothesized the start of his primary campaign in June 2015 as an incision point and fitted linear regression models (Hausman and Rapson 2018) with a discontinuity in June 2015 to the time series, as illustrated in Fig. 1.
The results (cf.Fig. 1) provide clear evidence of a sudden shift coinciding with the hypothesized discontinuity at the start of Trump's primary campaign in June 2015, when the overall political tone became abruptly more negative.The effect was large and highly significant: the frequency of negative emotion words soared by 1.6 pre-campaign standard deviations, or by 8% of the pre-campaign mean.Moreover, whereas negative language had decreased continuously during (at least) the first 6.5 years of Obama's tenure, this trend came to a halt with the June 2015 discontinuity, and the level of negative language remained high afterwards.Similar effects were observed for specific subtypes of negative language (anger, anxiety, and sadness), as well as for swearing terms.Qualitatively, the patterns are universal, emerging within each party and within strata of speaker prominence.Quantitatively, the language of Democrats, more prominent speakers, Congress members, and members of the opposition party is, ceteris paribus, overall more negative, but the shift at the discontinuity remains (and more strongly so for Republicans) when adjusting for these biographic attributes.
These population-level effects are not the results of systematic shifts in the distribution of quoted politicians, but are mirrored at the individual level for a majority of the most-quoted politicians.Moreover, the effect size drops by 40% when omitting Donald Trump's quotes from the analysis, and by 50% when weighing each speaker equally, rather than by the number of their quoted utterances (but the effect remains highly significant in both cases).Both findings imply that prominent speakers-and particularly Donald Trump-have disproportionately, though not exclusively, contributed to the rise in negative language.
Taken together, these results objectively confirm the subjective impression held by most Americans (Pew Research Center 2019): recent years have indeed seen a profound and lasting change toward a more negative tone in US politicians' language as reflected in online news, with the 2016 primary campaigns acting as a turning point.Moreover, contrary to some commentators' assessment (Fritze and Jackson 2021;James 2021), Donald Trump's appearance in the political arena was linked to a directional change, rather than a continuation of previously existing trends in political tone.Whether these effects are fully driven by changes in politicians' behavior or whether they are exacerbated by a shifting selection bias on behalf of the media remains an important open question (see Discussion).Either way, the results presented here have implications for how we see both the past and the future of US politics.Regarding the past, they emphasize the symptoms of growing toxicity in US politics from a new angle.Regarding the future, they highlight the danger of a positive feedback loop of negativity.

Results
To quantify negative language in quotes, we used Linguistic Inquiry and Word Count (LIWC) (Pennebaker, Booth, and Francis 2007), which provides a dictionary of words belonging to various linguistic, psychological, and topical categories, and whose validity has been established by numerous studies in different contexts and domains (Alpers et al. 2005;Kahn et al. 2007;Settanni and Marengo 2015).We computed a negative-emotion score for each of the 24 million quotes via the percentage of the quote's constituent words that belong to LIWC's negative emotion category.Analogously, we computed scores for LIWC's three subcategories of the negative emotion category-anger, anxiety, and sadness-as well as for the swear words category.We collectively refer to these five categories as "negative language". 1 In order to obtain a time series for each word category, we averaged individual quote scores by month.
By giving each quote the same weight when averaging, we obtain quote-level aggregates; by giving each speaker the same weight, speaker-level aggregates.(Formal definitions in Materials and Methods.)Quotelevel aggregates give more weight to more frequently quoted politicians, and thus better capture the overall tone of political language as reflected in the news.Whenever, on the contrary, we reason about politicians, rather than about the overall media climate created by all politicians' joint output, we use speaker-level aggregates as the more appropriate aggregation.
Table 1 summarizes the prevalence of the above five word categories via the means and standard deviations of the respective quote-level aggregates during the pre-campaign period (September 2008through May 2015).We observe that one in 54 words expresses a negative emotion; one in 155, anger; one in 339, anxiety; and one in 285, sadness.Swear words, at a rate of one in 2,329, are least common.Given this wide range of frequencies, and in order to make effect sizes comparable across word categories, we standardize all monthly scores category-wise by subtracting the respective pre-campaign mean and dividing by the respective pre-campaign standard deviation. 2The resulting effect sizes, in units of pre-campaign standard deviations, become more palpable when expressed as multiples of the corresponding pre-campaign 1.As a robustness check, we replicated the analysis using the dictionary provided by Empath (Fast, Chen, and Bernstein 2016) instead of LIWC.In contrast to LIWC, the categories in Empath's dictionary were not hand-crafted but generated by a deep-learning model based on a small set of seed words.The results, which are consistent with those from the LIWC-based analysis, are provided in Supplementary Fig. S1, S2, S3, and S4, and in Supplementary Table S1.
2. In order to facilitate the comparability and interpretability of results, standardization always uses the means and standard deviations computed on quote-level aggregates involving all speakers, even in analyses of speaker-level aggregates or of quotelevel aggregates when omitting individual speakers.means.For this purpose, the rightmost column of Table 1 lists coefficients of variation, i.e., ratios of standard deviations and means.
Temporal evolution of negative language.Fig. 2 visualizes the evolution of negative language between September 2008 and April 2020, with one row per aggregation method, one column per word category, and one point per monthly aggregate score.In order to quantify the shape of the curves, we fitted ordinary least squares linear regression models with a discontinuity in June 2015, the starting month of Donald Trump's primary campaign.For a given word category, we model the aggregate score y t for month t as where t ∈ {−81, . . ., 58}, with t = 0 corresponding to June 2015; i t indicates whether t is located before vs.
after the campaign start (i.e., i t = 0 for t < 0, and i t = 1 for t ≥ 0); and ε t is the residual error.As illustrated in Fig. 1(b), the coefficient α captures the immediate jump coinciding with the campaign start, and β, the change in slope, such that β 0 and β 0 + β describe the slopes of the regression lines before and after the campaign start, respectively.Fig. 2 plots the fitted regression lines, alongside 95% confidence intervals.
We first focus on the time series of quote-level aggregates (Fig. 2(a-e)).The regression coefficients indicate a significant, sudden increase in the relative frequency of negative emotion words (Fig. 2(a)) in June 2015, by α = 1.6 (p = 4.1 × 10 −6 ) pre-campaign standard deviations (SD), translating to a relative increase of 8% over the pre-campaign mean (cf.Table 1).All three subcategories of negative emotion words, as well as swear words, also saw significant jumps in June 2015: anger, by 1.3 SD (+11% of the pre-campaign mean, p = 0.0015, Fig. 2 We emphasize that the June 2015 discontinuity was chosen ex ante based on incoming hypotheses grounded in the general public's subjective impression that Donald Trump's entering the political scene had changed the tone of US politics (Pew Research Center 2019).A data-driven analysis, conducted ex post, revealed that June 2015 is in fact the optimal discontinuity for modeling the data: out of 140 regression models analogous to Eq. 1, but each using another one of the 140 months of our analysis period as the discontinuity, the model with the June 2015 discontinuity yielded the best fit for negative emotion words and anger (Supplementary Fig. S5).(For anxiety and sadness, slightly better fits were obtained when using different discontinuities; for swear words, a 2009 discontinuity led to a better fit due to outliers around that time; see Supplementary Fig. S5.) Role of speaker prominence.When repeating the analysis using speaker-level, rather than quote-level, aggregates, i.e., weighing all speakers equally when averaging, all of the above effects persisted qualitatively, but were reduced quantitatively (Fig. 2(f-j)): in each of the five word categories, the immediate increase α at the June 2015 discontinuity dropped by between one-third and one-half, remaining significant for negative emotions, anger, and swear words (p = 9.1 × 10 −5 , 0.0025, and 0.030, respectively), but becoming nonsignificant for anxiety and sadness (p = 0.11 and 0.43, respectively).The change of slope observed in the quote-level analysis also persisted in the speaker-level analysis.
The fact that speaker-level effects are weaker than quote-level effects indicates that prominent, highly quoted speakers contribute disproportionally to the increase in negative language.To confirm this conclusion more directly, we divided the speakers into four quartiles with respect to the number of quotes attributed to them, and repeated the speaker-level analysis for each quartile individually.Fig. 3 shows (1) that the abrupt increase in negative emotion words emerges in all strata of speaker prominence except the least prominent stratum, and (2) that quotes by more prominent speakers generally contain more negative emotion words.
(Results for other word categories in Supplementary Fig. S3.) Biographic correlates of negative language.The patterns identified above-a sudden increase in negative language in June 2015 followed by a change in slope-hold across party lines, 5 but are more pronounced for Republicans, as seen in Fig. 2(k-o), which tracks the (quote-level) evolution of negative language over time, analogously to Fig. 2(a-e), but separately for quotes by Democrats vs. Republicans.For instance, the 5.All α coefficients of Fig. 2(k-o) are positive.For Republicans, all are significant (p < 0.05); for Democrats, the coefficients for anger (α = 0.57) and sadness (α = 0.43) are non-significant (p > 0.05).
We further considered the possibility that the distribution of speaker characteristics may have changed over time; e.g., members of one party or gender may have become more frequently quoted over time.
To account for potential confounding due to such factors, we repeated the regression analysis with added control terms for four biographic attributes: party affiliation (Republican, Democrat), the party's federal role (Opposition, Government), Congress membership (Non-Congress, Congress), and gender (Male, Female). 6  For a given month, party affiliation fully determines the party's federal role, so for each month, the set of speakers can be partitioned into 2 3 = 8 speaker groups, one per valid attribute combination.We computed monthly aggregate scores y gt separately for each speaker group g, obtaining eight aggregate data points per month, and modeled them jointly using the following extended version of Eq. 1: where Democrat g = 1 (or 0) if group g contains Democrat (or Republican) speakers, and analogously for Government g , Congress g , and Female g .
Inspecting the fitted coefficients (shown in Fig. 4 for quote-level aggregation; for speaker-level aggregation, see Supplementary Fig. S6), we make two observations: First, the sudden June 2015 jump α in negative language (Fig. 4(b)) appears even after adjusting for the four biographic attributes.Second, we observe systematic correlations of negative language with biographic attributes, whereby quotes by members of the opposition party, Congress members, and Democrats contain, ceteris paribus, significantly more negative language.Moreover, quotes by females contained more anxiety and sadness, and quotes by males, more negative emotion, anger, and swear words.The fact that quotes by members of the opposition party contain more negative language may reflect their role as corrective agents in the democratic process.The fact that quotes by Congress members contain more negative language echoes previous work that has highlighted high-ranked politicians' role not only as deputies of their own party, but also as antagonists of the opposing party, epitomized in the finding that "ideological moderates won't run" (Thomsen 2014) for Congress.
Role of individual politicians.In order to determine to what extent the above population-level effects mirror individual-level effects, we fitted regression models (cf.Eq. 1) to individual speakers' time series 6. Due to small sample size, we discarded speakers of a non-binary gender according to Wikidata.
and analyzed the corresponding α coefficients (i.e., the size of the June 2015 jump; for completeness, β in Supplementary Fig. S7).In order to avoid data sparsity issues, this analysis focuses on the 200 mostquoted speakers, with α plotted in Fig. 5(a-e).Additionally, Fig. 5(f-j) plots the fraction of speakers with positive α among the speakers with at least n quotes, as a function of n.We observe that, although many individual α coefficients are non-significant (p > 0.05, gray confidence intervals in Fig. 5(a-e)), the majority of coefficients are positive, particularly among the top most-quoted speakers (as manifested in the increasing curves of Fig. 5(f-j)).For instance, for negative emotion words (Fig. 5(f)), all of the top four most-quoted politicians have positive α.Among the top 50, 74% have positive α; among the top 100, 63%; and among the top 200, 59%.In other words, the June 2015 discontinuity emerges not only by aggregating at the population level, but mirrors a disruption that can also be perceived in the most-quoted politicians' individual language.
We further illustrate (Supplementary Fig. S8) using as examples the two presidents (Barack Obama, Donald Trump), vice presidents (Joe Biden, Mike Pence), and runners-up (Mitt Romney, Hillary Clinton) from the study period.Four of these six speakers were associated with a significant (p < 0.05) positive α for negative emotion words, and none with a significant negative α. 7 Next, we sought evidence whether individual politicians contributed particularly strongly to the overall increase in negative language.We proceeded in an ablation study: for each of the 50 most quoted speakers, we repeated the quote-level regression (Fig. 2(a-e)) on a dataset consisting of all quotes except those from the respective speaker.If α is particularly low when removing a given speaker, that speaker contributed particularly strongly to the overall June 2015 increase in negative language.Fig. 6 7. Interestingly, although both Donald Trump (α = 3.6, p = 0.026) and Barack Obama (α = 3.7, p = 0.0094) followed the population-wide pattern by increasing their frequency of negative emotion words in June 2015, their pre-and post-discontinuity slopes were opposite to the population-wide pattern: their language first became more negative, then less negative.
Positive emotion words.Complementary to the five word categories related to negative language analyzed above, we also conducted an exploratory analysis of positive language, as captured by LIWC's positive emotion category.As seen in Fig. 7(a), the time series of positive emotion words does not simply mirror that of negative emotion words.In particular, no particular changes are observed in June 2015.Rather, positive emotion words appear stable well into Trump's term, and then decline.(The best regression fit is achieved when using July 2018 as the discontinuity, with a sharp drop in quality of fit after February 2019; Supplementary Fig. S10.)As in the case of negative emotion words, the same pattern emerges within each party (Fig. 7(c)), with a more pronounced drop in positive emotion words for Democrats than for Republicans.

Discussion
The goal of this work has been to determine the accuracy of the average American's subjective impression (as of mid 2019) that "the tone and nature of political debate in the United States has become more negative in recent years", and that Donald Trump "has changed the tone and nature of political debate [. . .] for the worse" (Pew Research Center 2019).Based on an analysis of 24 million quotes uttered by 18,627 US politicians between 2008 and 2020, we conclude that both of the above impressions are largely correct.
US politicians' tone indeed became suddenly and significantly more negative with the start of the 2016 primary campaigns in June 2015, and the frequency of negative language has remained elevated ever since.
Intriguingly, the shift at this incision point coincides with a similar abrupt shift in political polarization on online platforms (Waller and Anderson 2021;Frimer et al. 2022).The sudden increase in negative language reported here was not only significant, but also strong; e.g., the frequency of negative emotion words jumped up by 1.6 pre-campaign standard deviations, or by 8% of the pre-campaign mean.The disruption becomes particularly stark when contrasted with the first 6.5 years of Obama's tenure, during which negative language had decreased steadily-at odds with a commonly held belief that Trump merely continued an older trend (Fritze and Jackson 2021;James 2021).The potential of negativity, incivility, and fear as tools to support political campaigns has been long known (Brooks and Geer 2007;Gerstlé and Nai 2019) and might explain the increase of negative language during the campaigns.It cannot, however, explain the fact that the boost in negative language has continued for years after the campaigns had ended.
Rather, our results show that political debate during Donald Trump's entire term was characterized by a negative tone, and they specifically point to Trump as a key driver of this development: when removing Trump's quotes from the corpus, the magnitude of the June 2015 jump in the frequency of negative emotion words drops by 40%.Interestingly, Trump's own negative tone (Supplementary Fig. S8(z)) followed longterm trends opposite to the population-wide trends, with an initial increase and a subsequent decrease (and with a sudden June 2015 increase akin to that of the population).But as his language was overall far more negative than the average (Trump's mean monthly average was 8.0 pre-campaign standard deviations larger than the overall mean monthly average), he strongly skewed the overall tone toward the negative end when he moved to the center of the media's attention (Patterson 2017).Despite Trump's disproportionate impact, the increase in negative language was, however, not due to Trump alone.It remained significant in various complementary analyses: when removing Trump's quotes from the analysis, when giving equal weight to all speakers, when analyzing each party separately, and when analyzing the most-quoted speakers individually.The negative tone of others might be partly provoked by Trump's statements and actions, but as we do not have access to a counterfactual world without Trump, our analysis cannot speak to this possibility.
Our analysis also cannot disambiguate to what extent the observed shift in negative language is caused by a real shift in what politicians say vs. a shift in what the media choose to report.It is well known that media outlets are biased in what they report (Gentzkow and Shapiro 2010;Puglisi and Jr. 2011;Schiffer 2006), typically towards negative news that tend to increase engagement on readers' behalf (Grabe and Kamhawi 2006), and this bias may have drifted during the 12 years analyzed here.Future work may investigate this possibility by comparing quotes reported in the media to complete records of certain politicians' utterances in certain contexts, e.g., via Congressional records or public speeches (Niculae et al. 2015).(But whether the shift was caused by politicians or by the media, the effect would be identical: a more negative tone as perceived by news-reading citizens.)Moreover, although word-counting is a powerful tool for detecting broad trends in large textual data (Pennebaker and Graybeal 2001;Kahn et al. 2007), and although our results are robust to the specific choice of dictionary (Empath (Fast, Chen, and Bernstein 2016) yields qualitatively identical results to LIWC (Pennebaker, Booth, and Francis 2007); Supplementary Fig. S1, S2, S3, S4, and S11, and Supplementary Table S1), word-counting is crude and insensitive to nuances in context (Beasley and Mason 2015).Our findings should thus be considered a starting point and hypothesis generator for more detailed analyses based on a closer inspection of the text and context of quotes, e.g., using a combination of advanced natural language processing tools and human annotation in order to shed light on the relation between negative language and polarization and partisanship: Has politicians' tone become more negative specifically when they talk about opponents?
The present study is also limited with respect to its time frame.Although Quotebank (Vaucher et al. 2021a) is the largest existing corpus of speaker-attributed quotes in terms of size and temporal extent, the 12 years it spans are but a short sliver of the United States' long political history.It is thus unclear when the decrease in negative language at the beginning of the corpus (during the initial 6.5 years of Obama's tenure) started.Gentzkow et al. (Gentzkow, Shapiro, and Taddy 2019) observed a concurrent decrease in partisan language in Congressional speeches during Obama's tenure (2009)(2010)(2011)(2012)(2013)(2014)(2015)(2016), which might be yet another manifestation of the processes that underlie the initial decrease in negative language observed here.Gentzkow et al.'s corpus of Congressional speeches, however, spans a much longer time (144 years, 1873-2016), during which overall partisanship increased much more-especially from the 1980s onward-than it eventually decreased during Obama's tenure.We must therefore consider the possibility that, analogously, the 2009-2015 decrease in negative language may have been merely a short anomaly in a longer increasing trend in negative language.(Unfortunately, Gentzkow et al.'s corpus ends with Obama's tenure, so we cannot compare trends in negative language to trends in partisan language during Trump's tenure.) We saw that the June 2015 rise in negative language was not accompanied by a simultaneous drop in positive language.Rather, maybe in line with a general positivity bias in human language (Dodds et al. 2015), positive language remained stable until it eventually dropped during Trump's term.What happened at this point is an open question that lies beyond the scope of this work.The phenomenon highlights, however, that positive and negative emotion words are not necessarily complementary.From the start of the 2016 primary campaigns through the first half of Donald Trump's term, political tone was both highly positive and highly negative-akin to Trump's own style, characterized by typical features of affective polarization such as positive self-representation and negative other-presentation (Ahmadian, Azarshahi, and Paulhus 2017;Kreis 2017;Lewandowsky, Jetter, and Ecker 2020): "Sorry losers and haters, but my I.Q. is one of the highest -and you all know it!Please don't feel so stupid or insecure,it's not your fault". 99. https://web.archive.org/web/20151013150633/https://twitter.com/realDonaldTrump/status/332308211321425920This said, a key contribution of this work is the conclusion that, despite Donald Trump's key role in setting the tone of political debate, the shift towards a more negative tone permeates all of US politics.The consequences are tangible, as shown by research that highlights the detrimental effects of affective polarization on altruism (Whitt et al. 2020), trust (Mutz and Reeves 2005)

Materials and Methods
US politicians.We considered as politicians all people for whom the Wikidata knowledge base (Vrandecic and Krötzsch 2014) (version of 27 October 2021) lists "politician" (Wikidata item Q82955) or a subclass thereof as an occupation (P106).Given our focus on the United States, we included only those politicians whose party affiliation (P102) was listed as Democrat (Q29552) or Republican (Q29468).We considered as members of Congress those for whom Wikidata listed a US Congress Bio ID 10 (P1157), making no distinction between former and active members of Congress.Due to the small number of speakers of a nonbinary gender, we included only speakers whose gender (P21) was listed in Wikidata as male (Q6581097) or female (Q6581072).
Quotebank.The analyzed quotes were obtained from Quotebank (Vaucher et al. 2021a), a publicly available (Vaucher et al. 2021b) corpus of 235 million unique speaker-attributed quotes extracted from 127 million English news articles published between September 2008 and April 2020, provided by the large-scale online media aggregation service Spinn3r.com.While Spinn3r.com collects and supplies content from a comprehensive set of news domains (West, Leskovec, and Potts 2021), it also includes much content beyond news alone, including "social media, weblogs, news, video, and live web content" (Spinn3r 2017).Therefore, Quotebank was extracted from a filtered data set consisting only of content from a set of about 17,000 online news domains, defined as the set of domains appearing at least once in the large News on the Web corpus (News on the Web Corpus), which has been collecting large numbers of news articles from Google News 10.For instance, https://bioguide.congress.gov/search/bio/S000148and Bing News since 2010 and may thus be considered to provide a comprehensive list of English-language media outlets.We emphasize that the News on the Web corpus was only used for defining the set of news domains.It was not used for obtaining the news articles themselves, which originated from Spinn3r.com only.
We use the quote-centric (as opposed to the article-centric) version of Quotebank, which contains one entry per unique quote and aggregates information from all news articles in which the quote occurs.In constructing Quotebank, a machine learning algorithm (based on the large pre-trained BERT language model (Devlin et al. 2019)) was used to infer, for each quote, a probability distribution over all speaker names mentioned in the text surrounding the quote (and an additional "no speaker" option), specifying each speaker's estimated probability of having uttered the quote.For a given quote, we maintained only the name with the highest probability and consider it to indicate the speaker of the quote (a heuristic that was shown to have an accuracy of around 87% (Vaucher et al. 2021a)).A speaker name may be ambiguous.In such cases, Quotebank does not attempt to disambiguate the name, but rather provides a list of all speakers to whom the name may refer, where speakers are identified by their unique ID from the Wikidata knowledge base (Vrandecic and Krötzsch 2014).In our analysis, we attributed quotes that were linked to ambiguous speaker names (less than 6% of all quotes, see below) to each speaker to whom the respective name may refer.
To further clean the data set, we discarded quotes that were clearly non-verbal (e.g., consisting of URLs, HTML tags, or dates only).Moreover, on some days, Spinn3r.com, which provided the news articles for Quotebank, failed to deliver content due to technical problems (Vaucher et al. 2021a).We therefore identified missing days as those having less than 10% of the median number of unique quotes and dropped eight Quotes by US politicians.Keeping only quotes attributed to US politicians (see above), we obtained 24 million quotes attributed to 18,627 unique US politicians.Out of these, 4,487 were female, 14,140 male; 9,390 were Democrats, 9,237 Republicans; 11 and 1,790 were labeled as members of Congress.The most 11.Out of an original set of 18,954 US politicians appearing in Quotebank, Wikidata listed 327 as (former) members of both parties, usually because they switched membership during their careers.Out of these 327, we manually checked the 21 politicians with over 10,000 quotes, of whom 16 could be unambiguously assigned to one party for the study period.The remaining 311 politicians were dropped from the analysis.prolific speakers were Barack Obama (1.5m quotes), 12 Donald Trump (763k quotes), Mitt Romney (281k quotes), Hillary Clinton (230k quotes), George W. Bush (200k quotes), John McCain (161k quotes), and Joe Biden (127k quotes).For a list of the 30 most frequently quoted politicians, see Supplementary Table S17.Although, as mentioned, ambiguous names led to some quotes being attributed to multiple speakers, this happened rarely: the vast majority (94.3%) of quotes were attributed to a single politician, 4.3% to two politicians, 1.2% to three politicians, and 0.13% to four politicians.Supplementary Fig. S12 shows the number of quotes and the number of unique speakers per month.
Aggregation methods.Consider a fixed LIWC word category c and a fixed month t.Let S be the set of speakers with at least one quote during month t.Let Q s be the set of quotes attributed to speaker s ∈ S during month t, and let Q = s∈S Q s be the set of all quotes from month t. 13 Let ψ(q) be quote q's score for word category c.
Then, the quote-level aggregate score for word category c in month t is defined as and the speaker-level aggregate score, as That is, in quote-level aggregation, every speaker contributes with weight proportional to their number of quotes, whereas in speaker-level aggregation, all speakers contribute with equal weight.
12. It seems unreasonable that any single person could utter over 300 different quote-worthy statements a day.Instead, the large number can be explained by news outlets attending to different parts of a politician's spoken output.For instance, a long sentence can be quoted in various ways.
13.The (few) quotes attributed to multiple speakers are included in Q once per speaker.  of the full y-range).Whereas the frequency of negative emotion words had decreased continuously during the first 6.5 years of Obama's tenure, it suddenly and lastingly increased in June 2015, when Trump's primary campaign started and his quote share began to surpass Obama's.(b) Regression analysis: The black points again show the fraction of negative emotion words, but now as z-scores (i.e., after subtracting the pre-campaign mean and dividing by the pre-campaign standard deviation).In red, we plot regression lines for the periods before and after June 2015.The coefficients of the ordinary least squares regression model y t = α 0 + β 0 t + α i t + β i t t + ε t (where t is the number of months since June 2015, and i t indicates whether t ≥ 0; cf.Eq. 1) quantify the slopes of both lines, as well as the sudden increase of α = 1.6 pre-campaign standard deviations coinciding with the discontinuity in June 2015 (t = 0), as visualized.
Table 1: Overall statistics of LIWC negative-language word categories.Means µ and standard deviations σ were calculated over monthly quote-level aggregates from the pre-campaign period (September 2008 through May 2015).The value n = 1/µ in the fourth column implies that, in an average quote, on average every n-th word belongs to the respective category.The coefficients of variation, σ/µ, shown in the fifth column, allow to easily translate pre-campaign standard deviations (as shown on the y-axes of time series plots) into fractions of pre-campaign means.The most frequent words per category are listed in Supplementary Tables S2 and S3.Swear words

(o)
Democrats Republicans Figure 2: Temporal evolution of negative language.Columns correspond to negative-language word categories from LIWC; rows correspond to aggregation methods for computing monthly averages.Points show monthly averages, expressed as pre-campaign z-scores (i.e., subtracting pre-campaign mean from raw frequency values, and dividing by pre-campaign standard deviation).Lines (with 95% confidence intervals) were obtained via ordinary least squares regression, with coefficients shown in legends (cf.Eq. 1 and Fig. 1(b) for interpretation of coefficients; tabular summary in Supplementary Tables S4, S6, S8, S9).
(a-e) Quote-level aggregation micro-averages over all quotes per month, i.e., speakers have weight proportional to their number of quotes in the respective month.Panel (a) shows the same data as Fig. 1.
(f-j) Speaker-level aggregation macro-averages by speaker, i.e., all speakers with at least one quote in a given month have equal weight in that month.(k-o) Quote-level aggregation by party performs the analysis of (a-e), but separately for quotes from Democrats vs. Republicans (coefficients omitted for clarity; cf.Supplementary Tables S8 and S9).Significance of regression coefficients: *** p < 0.001, ** p < 0.01, * p < 0.05.We observe drastic shifts toward a more negative tone at the modeled June 2015 discontinuity (Trump's campaign start).

(d)
Figure 3: Role of speaker prominence.The set of 18,627 US politicians was split into four evenly-sized quartiles with respect to their total number of quotes (i.e., prominence); each panel shows the time series for negative emotion words obtained by performing monthly speaker-level aggregation on the respective quartile separately.That is, the figure shows the data of Fig. 2(f) after stratifying speakers by prominence.Lines (with 95% confidence intervals) were obtained via ordinary least squares regression, with coefficients in legends (cf.Eq. 1 and Fig. 1(b) for interpretation of coefficients).(Tabular summary of regression coefficients in Supplementary Tables S11, S12, S13, and S14.) Significance of regression coefficients: *** p < 0.001, ** p < 0.01, * p < 0.05.We observe that the abrupt increase in negative emotion words emerges in all strata of speaker prominence except the least prominent stratum, and that quotes by more prominent speakers overall contain more negative emotion words.The figure focuses on one category of negative-language words (negative emotion); for the other four categories, see Supplementary Fig. S3.Coefficients (with 95% confidence intervals) of ordinary least squares regression (Eq.2) for modeling time series of word categories (quote-level aggregation; speaker-level aggregation in Supplementary Fig. S6) while adjusting for party affiliation (γ), the party's federal role (δ), Congress membership (ζ), and gender (η) (tabular summary in Supplementary Tables S15  and S16).Positive coefficients mark word categories that are, ceteris paribus, used more frequently by Democrats than by Republicans, by members of the governing than by members of the opposition party, by Congress members than by others, or by females than by males (and vice versa for negative coefficients).We observe that quotes by members of the opposition party, Congress members, and Democrats contain significantly more negative language.Importantly, the sudden June 2015 jump in negative language (α) remains significant after adjusting for biographic attributes.(f-j) Fraction of speakers with positive α among the speakers with at least n quotes, as a function of n.We observe that, although many individual α coefficients are non-significant (a-e), the majority of coefficients are positive, particularly among the most-quoted speakers (as manifested in the increasing curves of (f-j)).That is, the June 2015 jump in negative language emerges even at the individual level for a majority of the most-quoted politicians.
(b)); anxiety, by 1.5 SD (+11%, p = 5.6 × 10 −4 , Fig.2(c)); sadness, by 0.86 SD (+4%, p = 0.021, Fig.2(d)); and swear words, by 0.92 SD (+31%, p = 2.9 × 10 −6 , Fig.2(e)).3The June 2015 discontinuity was not only associated with a sudden increase in negative language, but also with a change in slope: whereas the frequency of negative emotion words (Fig.2(a)) had steadily and significantly decreased over the first 6.5 years of Obama's tenure by β 0 = −0.023SD per month (p = 9.5 × 10 −7 ), i.e., by about a quarter SD per year, this trend came to a halt in June 2015, with a (non-significantly) positive slope of β 0 + β = 0.0076 (compound p = 0.21) from June 2015 onward.4 , and opinion formation(Druckman   et al., n.d.;Forgette and Morris 2006), and by polls showing that politics has become a stressful experience for Americans (Pew Research Center 2019; American Psychological Association 2017), exacting an ever increasing toll on their physical, emotional, and social well-being(Smith, Hibbing, and Hibbing 2019;   American Psychological Association 2018).Finding ways to break out of this cycle of negativity is one of the big challenges faced by the United States today.

Figure 1 :
Figure1: Quantifying the evolution of negative language in US politics(2008)(2009)(2010)(2011)(2012)(2013)(2014)(2015)(2016)(2017)(2018)(2019)(2020).(a) The black points show the fraction of negative emotion words, averaged monthly over all quotes from all 18,627 quoted politicians.The red vs. blue background shows the quote share of Trump vs. Obama (if Trump had T quotes and Obama had O quotes in a given month, the respective red bar covers a fraction T /(T + O) of the full y-range).Whereas the frequency of negative emotion words had decreased continuously during the first 6.5 years of Obama's tenure, it suddenly and lastingly increased in June 2015, when Trump's primary campaign started and his quote share began to surpass Obama's.(b) Regression analysis: The black points again show the fraction of negative emotion words, but now as z-scores (i.e., after subtracting the pre-campaign mean and dividing by the pre-campaign standard deviation).In red, we plot regression lines for the periods before and after June 2015.The coefficients of the ordinary least squares regression model y t = α 0 + β 0 t + α i t + β i t t + ε t (where t is the number of months since June 2015, and i t indicates whether t ≥ 0; cf.Eq. 1) quantify the slopes of both lines, as well as the sudden increase of α = 1.6 pre-campaign standard deviations coinciding with the discontinuity in June 2015 (t = 0), as visualized.

Figure 4 :
Figure4: Biographic correlates of negative language.Coefficients (with 95% confidence intervals) of ordinary least squares regression (Eq.2) for modeling time series of word categories (quote-level aggregation; speaker-level aggregation in Supplementary Fig.S6) while adjusting for party affiliation (γ), the party's federal role (δ), Congress membership (ζ), and gender (η) (tabular summary in Supplementary TablesS15 and S16).Positive coefficients mark word categories that are, ceteris paribus, used more frequently by Democrats than by Republicans, by members of the governing than by members of the opposition party, by Congress members than by others, or by females than by males (and vice versa for negative coefficients).We observe that quotes by members of the opposition party, Congress members, and Democrats contain significantly more negative language.Importantly, the sudden June 2015 jump in negative language (α) remains significant after adjusting for biographic attributes.

Figure 5 :
Figure5: Role of individual politicians: single-speaker study.Results of ordinary least squares regressions (Eq. 1) fitted separately to the time series of each of the 200 most quoted speakers.(a-e) Each speaker's α coefficient (capturing the size of the June 2015 jump, with 95% confidence intervals) as a function of the speaker's number of quotes.Significant coefficients (p < 0.05) in color, others in gray.(f-j) Fraction of speakers with positive α among the speakers with at least n quotes, as a function of n.We observe that, although many individual α coefficients are non-significant (a-e), the majority of coefficients are positive, particularly among the most-quoted speakers (as manifested in the increasing curves of (f-j)).That is, the June 2015 jump in negative language emerges even at the individual level for a majority of the most-quoted politicians.

Figure 7 :
Figure 7: Regression parameter β resulting from an ordinary least squares regressions fitted separately to the time series of each of the 200 most quoted speakers.For detailed description of plot format, see caption of Fig. 5 in main text.

Figure 9 :Figure 10 :Figure 11 :
Figure 9: Regression parameter β obtained by removing all quotes by one target speaker, using the 50 most quoted speakers as target speakers

Figure 12 :
Figure12: Monthly number of speakers and unique quotations in Quotebank (restricted to the 18,627 US politicians considered in the analysis).Months that were removed due missing data (cf.Materials and Methods in main text) are plotted in gray.

Positive emotion Quote-level aggregation (a)
Figure6: Role of individual politicians: ablation study.Results from ordinary least squares regression (Eq.1; quote-level aggregation) on data sets obtained by removing all quotes by one target speaker and retaining all quotes from all 18,626 other speakers, using each of the 50 most quoted speakers as a target speaker.Each point shows the α coefficient (capturing the size of the June 2015 jump, with 95% confidence intervals) obtained after removing the target speaker's quotes from the analysis, as a function of the target speaker's number of quotes.Dashed horizontal lines mark the coefficients obtained on the full data set without removing any speaker (cf.Fig.2(a-e)).We observe that no single speaker's removal leads to a notable change in α, except Donald Trump: e.g., by removing Trump's quotes, α drops by 40% for negative emotion words (a).

emotion Quote-level aggregation by party (c)
Temporal evolution of positive emotion words.Points show monthly averages, expressed as pre-campaign z-scores.One panel per aggregation method for computing monthly averages.We observe that positive emotion words appear stable well into Trump's term, and then decline.

Table 1 :
(Extended version of Table1in the main text) Key metrics for all word categories, including results for empath word categories.Means µ and standard deviations σ were calculated over monthly quote-level aggregates from the pre-campaign period(September 2008 through May 2015).The value n = 1/µ in the fourth column implies that, in an average quote, on average every n-th word belongs to the respective category.The coefficients of variation, σ/µ, shown in the fifth column, allow to easily translate pre-campaign standard deviations (as shown on the y-axes of time series plots) into fractions of pre-campaign means.The most frequent words per category are listed in Supplementary TablesS2 and S3.

Table 2 :
The most common 30 words for each word category, before 15 June 2015.Bold values deviate from the second period (after 15 June 2015) by more than one percentage point

Table 3 :
The most common 30 words for each word category, after 15 June 2015.Bold values deviate from the first period (before 15 June 2015) by more than one percentage point

Table 4 :
OLS regression parameters for quote-level aggregation shown in Figure 2 in the main text.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05

Table 6 :
OLS regression parameters for speaker-level aggregation shown in Figure 2 in the main text.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05

Table 8 :
OLS regression parameters for quote-level aggregation for Democrats shown in Figure 2 in the main text.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05

Table 9 :
OLS regression parameters for quote-level aggregation for Republicans shown in Figure 2 in the main text.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05

Table 15 :
OLS regression parameters for quote-level aggregation on speaker groups based on party affiliation, party's federal role, Congress membership, and gender.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05

Table 16 :
OLS regression parameters for speaker-level aggregation on speaker groups based on party affiliation, party's federal role, Congress membership, and gender.SEs of coefficients are in parantheses.***p < 0.001, **p < 0.01 and *p < 0.05