Abstract
People rely on data-driven AI technologies nearly every time they go online, whether they are shopping, scrolling through news feeds, or looking for entertainment. Yet despite their ubiquity, personalization algorithms and the associated large-scale collection of personal data have largely escaped public scrutiny. Policy makers who wish to introduce regulations that respect people’s attitudes towards privacy and algorithmic personalization on the Internet would greatly benefit from knowing how people perceive personalization and personal data collection. To contribute to an empirical foundation for this knowledge, we surveyed public attitudes towards key aspects of algorithmic personalization and people’s data privacy concerns and behavior using representative online samples in Germany (N = 1065), Great Britain (N = 1092), and the United States (N = 1059). Our findings show that people object to the collection and use of sensitive personal information and to the personalization of political campaigning and, in Germany and Great Britain, to the personalization of news sources. Encouragingly, attitudes are independent of political preferences: People across the political spectrum share the same concerns about their data privacy and show similar levels of acceptance regarding personalized digital services and the use of private data for personalization. We also found an acceptability gap: People are more accepting of personalized services than of the collection of personal data and information required for these services. A large majority of respondents rated, on average, personalized services as more acceptable than the collection of personal information or data. The acceptability gap can be observed at both the aggregate and the individual level. Across countries, between 64% and 75% of respondents showed an acceptability gap. Our findings suggest a need for transparent algorithmic personalization that minimizes use of personal data, respects people’s preferences on personalization, is easy to adjust, and does not extend to political advertising.
Introduction
The online experience of billions of people is shaped by machine-learning algorithms and other types of artificial intelligence (AI) technologies. These self-learning programs include a variety of algorithmic tools that harvest and process people’s personal data in order to customize and mediate information online, in, for example, personalized social media feeds, targeted advertising, recommender systems, and algorithmic filtering in search engines (for more examples see Table B1 in Appendix B). Although many personalized services might be innocuous (e.g., music or movie suggestions), others challenge the existence of a transparent and open democratic marketplace of ideas and, ultimately, a collectively shared reality (Mazarr et al., 2019). For instance, there is substantial concern that personalized political messages containing false claims influenced both the Brexit referendum and the U.S. presidential election in 2016 (Digital, Culture, Media and Sport Committee, 2019; Persily, 2017). Furthermore, algorithms can amplify conspiracy theories, false or misleading information, and extremist content, which in turn may contribute to radicalization, the rise of political extremism (Baumann et al., 2020; Horwitz and Seetharaman, 2020; Kaiser and Rauchfleisch, 2018; Rauchfleisch and Kaiser, 2017), and increasing distrust in the media (Newman et al., 2020). Most recently, there have been growing concerns that the combination of algorithmic filtering and opinion dynamics on social media networks have fostered the spread of false information about the COVID-19 pandemic and governments’ responses to it, thereby reinforcing dangerous beliefs and conspiracy narratives (Cinelli et al., 2020; Thompson and Warzel, 2021; Zarocostas, 2020) and potentially hampering an efficient public response.
Data privacy and transparency are also causes for concern. People’s data are at the heart of the online ecosystem, where service providers monetize behavioral traces collected directly or by third-party trackers (Zuboff, 2019). This widespread collection of behavioral data enables AI algorithms to infer more information than people intend to share (e.g., information on sexual orientation, personality traits, and political views; Hinds and Joinson, 2019, 2018; Kosinski et al., 2016; Matz et al., 2017; Youyou et al., 2015). Using demographic and behavioral data in targeted advertising may also result in discrimination—for instance, by preferentially targeting (i.e., including or excluding) users belonging to disadvantaged social groups (e.g., according to race, religion, or sexual orientation; Speicher et al., 2018; see also Ali et al., 2019; Datta et al., 2018). For example, the 2016 Trump presidential election campaign has been accused of attempting to deter 3.5 million Black Americans from voting by deliberately targeting them on Facebook with negative Hillary Clinton ads (Sabbagh, 2020). Similarly, the Russian Internet Research Agency ran ads featuring socially divisive topics (e.g., immigration, race-based policing) on Facebook with the goal of creating social discord prior to the 2016 U.S. presidential elections; these ads were targeted at people based on their ethnicity (Ribeiro et al., 2019).
But how aware are people of the influence that algorithms exert on their online experience? And how acceptable do people find the use of their information for personalization?
Investigating these questions has become particularly urgent with the growing number of Internet users who rely on social media or search engines to find and access political news (Newman et al., 2020). Social media news feeds (e.g., on Facebook), video suggestions (e.g., on YouTube), and online advertising (on most platforms) have become highly personalized environments governed by nontransparent algorithms, and users have little control over how the information they see is curated.
There have been multiple calls to regulate online political advertising (e.g., Jaursch, 2020; Zuiderveen Borgesius et al., 2018) and align it with existing strict rules for offline political campaigning (for a discussion see Lewandowsky et al., 2020). To some extent, these calls have been heeded. At the end of 2020, the European Commission launched the Digital Services Act, which aims to upgrade the rules governing digital services in the European Union (European Commission, 2020a). This act complements the European Democracy Action Plan, which proposes legislation on, among other things, transparency of sponsored political content (European Commission, 2020b). Yet for now, platforms pursue their own, often divergent, approaches to targeting and personalization. For instance, Twitter now prohibits the promotion of political content (Twitter, 2021b), and Facebook recently tightened its advertising policies to, among other things, allow it to restrict certain electoral or political ads (e.g., in the lead-up to an election; Facebook, 2021)—its precise targeting mechanisms, however, remain nontransparent. Recent research has shown that Facebook’s ad delivery algorithms favor relevance (alignment with user interests) when it comes to pricing of ads and thereby indirectly incentivize political ads that align with users’ political preferences, thus potentially amplifying polarization dynamics (Ali et al., 2021). The significant power that digital platforms exercise over political discourse—whether through algorithms or human judgment (e.g., unilaterally banning Trump; Twitter, 2021a)—despite a lack of public accountability or scrutiny highlights the urgent need for public-centered social media regulation. Any approach to regulating the digital sphere should take into account people’s attitudes towards these key issues of data privacy and personalization of online content in order to secure public support and capture public ethical concerns. It is therefore puzzling that there has been little public involvement in monitoring and shaping the design of algorithms and the collection of data used for personalization.
We aimed to address this dearth of knowledge by focusing on people’s attitudes towards personalized online services and towards the use of their personal data and information in order to offer those services. Our goal is to contribute to a better understanding of people’s attitudes towards various aspects of online personalization.
Previous studies in the US and the UK have shown that attitudes towards personalization are context dependent: Attitudes are generally more positive towards commercial applications than towards personalized political information (Ipsos Mori, 2020; Smith, 2018). People in the US and Europe feel they have little control over their personal data and have general concerns about their digital privacy (Auxier et al., 2019; Directorate-General for Communication, 2019). Yet although people profess to care a great deal about their data privacy, their actual behavior does not necessarily reflect this concern. The inconsistency between people’s privacy attitudes and privacy behaviors has been coined the “privacy paradox” (e.g., Acquisti et al., 2015; Barth and de Jong, 2017; Kokolakis, 2017; Norberg et al., 2007; but see Dienlin and Trepte, 2015 and meta-analysis by Baruh et al., 2017). For example, even people with high privacy concerns do not display adequate privacy-protecting behavior (e.g., limiting profile visibility on social networks or controlling privacy settings on online platforms), although there is a modest positive relation between high privacy concerns and behavior (Baruh et al., 2017).
Attitudes towards privacy are not homogeneous: they may vary across different types of personalized services and across different types of personal data and information that make personalized services possible. Most studies have looked separately at either attitudes towards personalized services or attitudes towards data privacy. However, personal data is essential for personalized services, and so attitudes on data collection have implications for personalization, and possibly vice versa.
We therefore contrasted people’s attitudes towards different aspects of personalization, including services and collection of data and information, in order to draw a more comprehensive picture of people’s attitudes and the extent to which they cohere or conflict with each other. We included questions about the acceptability of personalization in various kinds of digital services and of collecting and processing people’s data for the purpose of personalization (see Hinds and Joinson, 2018 for a systematic review of demographic characteristics that can be inferred from people’s digital footprints). An awareness of these heterogeneous and fine-grained attitudes is crucial for regulatory interventions or guidelines, as well as for platforms’ efforts to self-regulate in a way that respects people’s preferences, concerns, and values.
In an online survey (using representative quota sampling) in Germany (N = 1065), Great Britain (N = 1092), and the United States (N = 1059), we inquired into three main aspects of public attitudes and behavior: (1) people’s awareness of the use of AI algorithms in online environments; (2) people’s attitudes towards three key components of algorithmic personalization online: personalized services (e.g., recommendations for music and movies, political campaigning), personal data collected online and used for personalization (e.g., location history, likes and shares on social media), and personal information that can be provided by users directly or inferred from data (e.g., age, gender, political leaning, sexual orientation); and (3) people’s concerns about the use of their personal data and how they protect their own personal information. We also investigated the extent to which people’s attitudes and concerns are moderated by political leaning and demographic characteristics.
Our findings (for details see the Results section) show that although people were willing to accept some personalized services (e.g., for shopping and entertainment), they objected to personalization in political campaigning and, in Germany and Great Britain, to the personalization of news and news sources. For instance, most respondents in Germany (61%) and Great Britain (61%) and approximately half in the US (51%) said personalized political advertising was unacceptable. In all three countries (but more so in Germany and Great Britain), people also objected to the use of most personal data and sensitive information that could be collected for personalization, including data related to their online interactions, such as with whom and how often they communicate (GER: 77%; GB: 66%; US: 60%); their location history (GER: 69%; GB: 57%; US: 55%); and their browsing and search history (GER: 63%; GB: 58%; US: 53%). Moreover, a large majority of respondents rated, on average, personalized services as more acceptable than the collection of personal information or data for the purposes of personalization. This acceptability gap between personalized services and the collection and use of data they require can be observed at both the aggregate and the individual level. Across countries, between 64% and 75% of respondents showed an acceptability gap. Furthermore, respondents in all three countries reported high levels of concern about data privacy, with 82% of participants in Germany, 81% in Great Britain, and 82% in the US saying they were somewhat or very concerned. Despite this widespread concern, respondents reported taking few steps to protect their privacy online—although those who were more concerned about privacy were more likely to change privacy settings and use privacy tools. Privacy concerns and attitudes are similar across the political spectrum, indicating that regulation related to data privacy protection and political advertising will likely be met with approval from voters regardless of their political leaning.
Methods
Sample and data collection
Dalia Research conducted the survey for the Max Planck Institute of Human Development in September (Germany) and November (Great Britain and US) 2019. Online samples were obtained in Germany (N = 1065), Great Britain (N = 1092), and the United States (N = 1059), using quota sampling and applying post-stratification weights to account for current population distributions with regard to age (18–65 years), gender, and education. The Institutional Review Board of the Max Planck Institute for Human Development approved the surveys. See Table 1 for demographic information about the three samples (weighted based on post-stratification survey weights; for both weighted and unweighted demographic information see Table B2 in Appendix B). Some preliminary results for the German sample, not including the acceptability gap, were made available in a technical report in English (Kozyreva et al., 2020a) and in German (Kozyreva et al., 2020b).
Study design
The survey was conducted online in German and English. The survey questions covered three topics: Public awareness of the use of AI and personalization algorithms on the Internet, public attitudes towards algorithmic personalization, and public attitudes and behavior regarding online privacy. We also collected information about participants’ demographics and political leanings. In Fig. 1 and in the following we provide an overview of the study design and summarize the gist of the survey questions; for the full questionnaire in English and in German, see Appendix C.
-
(1)
Public awareness of the use of AI and personalization algorithms on the Internet. We defined “artificial intelligence (AI) technologies” as self-learning computer programs (“machine learning”) that analyze people’s personal data in order to customize their online experience. We asked people whether they thought that AI technologies are used in a variety of online situations, including news feeds, advertising on social media, and product recommendations in online shops (see Fig. A1 and Study questionnaire in Appendix C for full list).
-
(2)
Attitudes towards algorithmic personalization. In order to gain a more complete understanding of how acceptable people find algorithmic personalization online, we asked about three key components of personalization: services, information, and data. All three are necessary for a full picture of attitudinal heterogeneity, both within individuals and across individuals. The set of questions for all three dimensions (services, information, and data) represents common personalization practices. To elicit attitudes towards personalizing services, we asked respondents “How acceptable do you think it is for social media and other websites to collect and use data about you and your past online activities to [personalize different online services, e.g., search results or friend suggestions]?” (see Fig. 2 and Study questionnaire in Appendix C for full list). To elicit attitudes towards information we asked respondents “How acceptable do you think it is for online web platforms to use any of the following information about you to create personalized advertising?” (e.g., gender, age, political views, sexual orientation; see Fig. 2 and Study questionnaire in Appendix C for full list). To elicit attitudes towards data collection, we asked respondents “How acceptable do you think it is for web services and applications to record and use the following types of information that they collect about you on their platform?” (e.g., browsing and search history, location history, content of emails and online messages; see Fig. 2 and Study questionnaire in Appendix C for full list). Respondents could answer “not acceptable at all”, “not very acceptable”, “somewhat acceptable”, or “very acceptable”.
-
(3)
Public attitudes and behavior regarding online privacy: To elicit respondents’ concerns about their data privacy online, we asked “How concerned are you about your data privacy when using the Internet?” Respondents could answer “not concerned at all”, “not very concerned”, “somewhat concerned”, or “very concerned”. To elicit people’s self-reported privacy-protecting behavior online, we asked “Which of the following [privacy settings] have you used in the last year to check and/or adjust what kind of data on you can be used by Internet companies?” (e.g., activity controls on Google) and “Which of the following measures and tools do you currently use to protect your data privacy online?” (e.g., ad blockers, VPN; see Fig. 4 and study questionnaire in Appendix C for full lists of settings and tools).
-
(4)
Demographics and political leanings: We collected respondents’ demographics (age, gender, education level, and location: urban/rural) and political leaning (on a scale ranging from “1 (left-wing)” to “7 (right-wing)”; see Figs. A2 and A3 in Appendix A and Table B2 in Appendix B for demographic information).
Data analysis
Anonymized data and reproducible R code are available at Open Science Framework: https://osf.io/7nj8h. Unless explicitly noted, all numbers and figures reported incorporate post-stratification survey weights provided by Dalia Research (based on age, gender, and education level) to increase the representativeness of the reported results.
For binary responses (or binary categorizations of rating-scale responses), the worst-case margin of error (i.e., the 95% confidence interval of a true proportion of 50%) is ≈ ±3 percentage points for a sample size of N = 1000 and ≈ ±10 percentage points for a sample size of N = 100.
Results
Public awareness of AI technologies online
Respondents were partially familiar with AI-related concepts and key entities: They knew that algorithms are employed online, and that algorithms are used to curate social media feeds (see Fig. A1 in Appendix A). For example, in all three countries, the majority of participants were familiar with the term “artificial intelligence” (GER: 86%; GB: 74%; US: 67%) and, more specifically, with targeted/personalized advertising (GER: 70%; GB: 58%; US: 50%). However, significantly fewer participants were familiar with recommender systems (GER: 34%; GB: 12%; US: 12%) and machine learning (GER: 42%; GB: 31%; US: 33%). Respondents were also aware that AI algorithms are employed in smart assistants (e.g., Siri or Alexa; GER: 70%; GB: 66%; US: 63%), search engine results ranking (GER: 59%; GB: 52%; US: 48%), and advertising on social media (GER: 57%; GB: 56%; US: 55%). They were less aware of AI used to recommend partners on dating websites (GER: 38%; GB: 41%; US: 40%) or curate social media news feeds (GER: 44%; GB: 43%; US: 44%). A clear majority of respondents correctly identified environments with little or no personalization (e.g., Wikipedia or a local restaurant’s website).
Public attitudes towards personalization and the collection and use of information and data
We found heterogeneity in respondents’ attitudes towards three key components of algorithmic personalization online (Fig. 2).
Percentage of respondents indicating levels of acceptability for (a) personalizing services, (b) using information for personalization, and (c) collecting and using data in online services in general. White numbers show percentages per rating category; black numbers show total percentages for the two sides of the rating scale. Within panels, items are ordered by their average rating pooled across all three countries (in ascending order of acceptability).
Personalized services
Most respondents in Germany (61%) and Great Britain (61%) and approximately half in the US (51%) said personalized political advertising was unacceptable. Approximately half of respondents in Germany and Great Britain opposed personalized news, including on front pages of online newspapers (GER: 52%; GB: 54%) and in news feeds on social media (GER: 57%; GB: 51%). In contrast, 60% of respondents in the US found personalized online newspapers acceptable and 62% approved of personalized social media news feeds. At the same time, a majority in all three countries approved of personalized recommendations for entertainment (movies or music: GER: 77%; GB: 84%; US: 88%), shopping (GER: 77%; GB: 82%; US: 89%), and search results (GER: 63%; GB: 60%; US: 71%).
Information
A majority of respondents found the collection and use of their personal information unacceptable. They clearly opposed personalization based on sensitive information (e.g., tragic and personal events, household income, sexual orientation, religious or political views). This opposition was highest in Germany and Great Britain; for instance, 71% and 59%, respectively, found it unacceptable to use political views for personalization. Respondents in Germany (71%) and Great Britain (62%) also found the use of information about their sexual orientation unacceptable. In the US, approximately half of respondents objected to the use of information about their political views (49%) and sexual orientation (51%), while a majority opposed the use of information about their household income or personal tragedies. Only age and gender were considered acceptable for personalization in all three countries by the majority of respondents. Respondents in the US were more accepting of information such as personal events (55%), their ethnicity (57%), their marital status (62%), and their personality traits (68%) being used for personalization online.
Personal data
In all three countries, most people objected to the collection and use of their personal data, including data related to their online interactions (e.g., with whom and how often they communicate; GER: 77%; GB: 66%; US: 60%); their location history (GER: 69%; GB: 57%; US: 55%); and their browsing and search history (GER: 63%; GB: 58%; US: 53%). Among the types of data that approximately half of respondents found acceptable were purchasing history (GER: 44%; GB: 47%; US: 51%), videos watched (GER: 44%; GB: 52%; US: 62%), and likes and shares on social media (GER: 43%; GB: 54%; US: 65%).
Acceptability gap in attitudes towards personalization
In all three countries a seemingly paradoxical result emerged: Respondents found personalized services (e.g., customized search results, online advertising, entertainment recommendations) more acceptable than the use of personal information and data (e.g., personal interests or location history), even though this information is currently used in personalized services. This constitutes what we call an “acceptability gap”, which we define as the difference between how acceptable people find personalized online services (e.g., social media news feeds, video suggestions) and how acceptable they find the collection and use of their personal data and information for such personalization. The gap exists on both the aggregate and the individual level (see Fig. 3).
A respondent’s acceptance level is defined as the arithmetic mean of their ratings (mapped into the [0, 1] range), ranging from 0 (not acceptable at all) to 1 (very acceptable). a Country subpanels show kernel-smoothed densities of the population distributions of acceptance levels for services, information, and data, respectively. Vertical dashed lines show the median values for each distribution; decimal values indicate how much lower the median value for information and data is compared to the median value for services. b Respondent-level differences between the acceptability level for services versus information (x-axis) and services versus data (y-axis). Positive values indicate that a respondent rated services as, on average, more acceptable than collecting information (upper half of each subpanel) or data (right half of each subpanel). Bold values show percentages of respondents falling into each of the four quadrants. Respondents in upper-right quadrants (blue) reported higher acceptability levels for both information and data; respondents in lower-left quadrants (gray) reported lower acceptability levels for both. Respondents in upper-left and lower-right quadrants showed an acceptability gap for only data (green) or only information (brown), respectively.
At the aggregate-level, the acceptability gap refers to the finding that the population medians of respondents’ average acceptability rating for services are greater than those for collecting information or data (Fig. 3a). Across comparisons and countries, the size of this gap ranges between one-sixth and one quarter of the full range of the response scale (“not acceptable at all”, “not very acceptable”, “somewhat acceptable”, “very acceptable”). That is, the size of the gap equalled as much as one step on the four-step rating scale. The gap was most pronounced in Germany (one quarter of the rating scale), and somewhat less pronounced, but still notable, for Great Britain (one-fifth of the rating scale) and the US (one-sixth of the rating scale for information and one-fifth for data).
At the individual level, the acceptability gap refers to the finding that a large majority of respondents rated, on average, personalized services as more acceptable than the collection of personal information or data (Fig. 3b). Across countries, 84%–89% of respondents showed at least one acceptability gap (for information and/or data). Between 64% and 75% of respondents showed an acceptability gap for both information and data. Only 13–16% showed no gap. Mirroring the aggregate-level results, the individual level acceptability gap is somewhat more pronounced in Germany than in Great Britain and the US.
In summary, for all three dimensions of online personalization (services, information, and data), we found that preferences were heterogeneous: Some services and data types were judged acceptable, others were not. On average, services were judged more acceptable than information and data.
Data privacy concerns and behavior
People in all three countries reported high levels of concern about their data privacy (Fig. 4 top panel): 82% of participants in Germany, 81% in Great Britain, and 82% in the US said they were somewhat or very concerned. Only a small fraction of respondents were not at all concerned (GER: 4%; GB: 4%; US: 6%), indicating that lower levels of concern do not explain the more pronounced laissez-faire attitudes to algorithmic personalization found in the US (see Figs. 2 and 3).
a Answers to the question “How concerned are you about your data privacy when using the Internet?” White numbers show percentages per rating category; black numbers show total percentages for the two sides of the rating scale. b Percentage of respondents indicating that they currently use a privacy tool, separately for respondents who indicated that they were not or not very (black), somewhat (orange), or very (red) concerned about their privacy online. The size of the points indicates the number of respondents contributing to a percentage value (see legend at the bottom of the figure); because only very few respondents said they were not concerned at all (see top panel), these respondents and those who said they were not very concerned were pooled into one category. Bold values show the percentage of respondents using a tool, irrespective of their level of concern. Items are ordered by this overall percentage of use, pooled across all three countries (in descending order except for “None of the above”, which is always last). c Percentage of respondents who indicated that they had checked or adjusted privacy settings within the last year (only considering respondents who indicated having used the respective service within the last year).
Despite the high levels of concern, respondents reported taking few steps to protect their privacy online (Fig. 4, panels b and c). Popular measures included changing privacy settings on Facebook (GER: 59%; GB: 60%; US: 63%) and Google (GER: 47%; GB: 44%; US: 53%) and using ad blockers (GER: 38%; GB: 34%; US: 36%). About 20% of respondents in Germany, 24% in Great Britain, and 17% in the US indicated that they did not use any privacy-protecting tools; results were similar for privacy-protecting settings (GER: 20%; GB: 24%; US: 19%). Respondents who were more concerned about privacy were more likely to change privacy settings and use privacy tools (see Fig. 4, panels b and c).
Role of demographics and political attitudes
With the exception of male respondents in the US, there was a general decline in acceptability for all three aspects of personalization (services, information, data) across age (Fig. A2 in Appendix A). For men in the US, there was an indication of a slight inverted U-shape, where acceptance increased slightly up to age 40 then declined; men in the US were thus overall slightly more accepting of all three aspects of personalization. No noteworthy gender effects emerged for Germany and Great Britain. Age and gender did not moderate our finding of a lower acceptance of information and data compared to services (see panel a in Fig. A2, Appendix A). In general, older respondents were more concerned about data privacy than were younger respondents and male respondents were slightly less concerned than were female respondents (see panel b in Fig. A2, Appendix A). We found no noteworthy associations between personalization attitudes and privacy concerns on the one hand and education or location (urban/rural) on the other hand (see Fig. A3 in Appendix A).
Importantly for public policy makers, we found no political polarization in attitudes towards personalization and privacy in all three countries (Fig. 5). Respondents across the political spectrum agreed on the acceptability of personalized services, the use of people’s information and data for personalized services, and the collection and use of sensitive information. They were also equally concerned about data privacy.
a Boxplots of respondents’ acceptance level towards services, information, and data (panel columns) broken down by political leaning (x-axis in subpanels) and country (panel rows). A respondent’s acceptance level is defined as the arithmetic mean of their ratings (mapped into the [0, 1] range), ranging from 0 (not acceptable at all) to 1 (very acceptable). The width of the boxplots is proportional to the square root of the weighted number of respondents per distribution; these rounded weighted counts are also shown as gray numbers below the x-axis. b Answers to the question “How concerned are you about your data privacy when using the Internet?’’ broken down by political leaning (x-axis in subpanels) and country (panel rows). White numbers show percentages per rating category; black numbers show total percentages for the two sides of the rating scale. The width of the stacked bars is proportional to the square root of the weighted number of respondents per distribution; these rounded weighted counts are also shown as gray numbers above the stacked bars.
Discussion and conclusion
The public perceives clear ethical boundaries in the use of algorithmic personalization online. Although people accept personalized commercial services (e.g., shopping and entertainment), they object to the use of the personal data and sensitive information that is currently collected for personalization. They consistently oppose advertising that is customized based on sensitive information, and find personalization in commercial services more acceptable than personalization in political and informational (e.g., news) content: People in all three countries oppose personalization in political campaigning, and people in Germany and Great Britain also oppose personalized news sources and social media feeds. This is an important finding with potentially far-reaching implications, given that social media feeds and political advertisement, like entertainment recommendations, can be highly personalized. People across the political spectrum are equally concerned about their data privacy and the effects of personalization on news and politics. This consensus, unusual in the current polarized political environment (especially in the US), raises the hope that policies for protecting online privacy and regulating the personalization of political news and advertising would receive broad support.
A clear tendency towards higher acceptability rates for all three categories—services, information, and data—can be observed in the US. Germany lies on the other side of the spectrum, with the lowest acceptability rating. Yet in all three countries we observed an acceptability gap: Even though most people accept personalized services, they generally oppose the collection and use of the personal, and specifically sensitive, information that personalized services collect. The reasons behind this gap are unclear. One possibility is that people have incommensurable values—that is, they value their data privacy but they also value the use of personalized services. Thus, people cannot help but to act as if they had found an acceptable trade-off between the immediate advantages of personalized services and future risks to their data privacy. When asked about their attitudes, however, they can acknowledge that data privacy and personalized services are to some extent in conflict, thus leading to the emergence of the acceptability gap. Trade-off processes also appear to play a role in the privacy paradox (Acquisti et al., 2015; Barth and de Jong, 2017). Another possible explanation for the acceptability gap is that it results from a lack of transparency in online services. Users might not be aware that companies such as Facebook or Google Maps need to collect information about their online behavior in order to customize news feeds or optimize suggestions. If this were the case, the trade-off people make between convenient personalized services and maintaining privacy online might not accurately reflect their preferences, since they may underestimate the extent to which the efficiency of personalized services hinges on data collection. This lack-of-awareness hypothesis is supported by the finding that 74% of Americans did not know that Facebook maintained a list of their interests and traits (Hitlin and Rainie, 2019).
Unlike the disconnect between people’s concerns about privacy and their approval of personalized services, people’s concerns about data privacy were moderately related to their privacy-protection behavior. This is consistent with the findings of a meta-analysis by Baruh et al. (2017), which demonstrated that privacy concerns are associated with the extent to which individuals engage in privacy management, although the magnitude of the association was modest. The positive relation between concerns and behavior could be another indication that the observed acceptability gap and privacy paradox are rooted in the current online environment, which does not offer users simple tools to keep their data safe and, consequently, does not support attitude-consistent privacy behavior. If this explanation is correct, then in order for privacy concerns and behavior to match more closely, the data privacy functions of online services should be more accessible, explained in simpler terms, and easy to use. Behavioral interventions (e.g., digital nudging and boosting; see Kozyreva et al., 2020c; Lorenz-Spreen et al., 2020) can also be employed to empower users to align their privacy protective measures to their level of privacy concern. New transparency measures could enable people to exercise their preferences in a more nuanced way, which would constitute an important next step towards regaining autonomy in the online world.
Finally, our results highlight the importance of personalization that respects people’s preferences for data privacy and their preference, shared across the political spectrum, that personalization not be used in political campaigning, and, at least in Europe, in news sources. To respect user preferences and concerns, platforms should shift towards using less personal data (e.g., the data minimization principle in Article 5 of the European Union’s General Data Protection Regulation; European Parliament, 2016). Collecting less data for personalization has been shown to be feasible without loss of quality in the recommendations based on minimized data (Biega et al., 2020; Wen et al., 2018). Furthermore, it is important to conceptualize data privacy and its protection in AI-assisted information environments as a public, as well as an individual, right and good (Fairfield and Engel, 2015). As algorithmic inferences from data collected from users can be harnessed to predict personal information of non-users (“shadow profiles”; Garcia, 2017), an individual’s privacy may be at risk through no fault of their own. Instead, the risk may arise from other users who are unconcerned about their data or were “nudged” by online choice architectures towards privacy-threatening options (Utz et al., 2019). Protecting user privacy must, therefore, encompass the privacy protection of citizens as a whole in what is known as a “networked privacy” model (Garcia, 2019)—a challenging but urgent task both for future research and policy making.
Understanding people’s general attitudes is crucial for defining the goals and values that inform regulations on networked data privacy and algorithmic personalization online. While our study contributed to this discussion and uncovered important links between people’s attitudes to online services and pertinent personal data, they do not shed light on what drives these attitudes. Further research is required to better understand the reasons behind both the acceptability gap and the privacy paradox. This should be complemented by research that investigates how people represent, understand, and resolve the trade-offs between using specific services and revealing the data that are needed to personalize those services, using people’s self reports and, crucially, their actual behavior.
Data availability
Anonymized data and code are available at Open Science Framework: https://osf.io/7nj8h/.
References
Acquisti A, Brandimarte L, Loewenstein G (2015) Privacy and human behavior in the age of information. Science 347:509–514
Ali M, Sapiezynski P, Korolova A, Mislove A, Rieke A (2021) Ad delivery algorithms: the hidden arbiters of political messaging. Proceedings of the 14th ACM International Conference on Web Search and Data Mining, 13–21
Ali M, Sapiezynski P, Bogen M, Korolova A, Mislove A, Rieke A (2019) Discrimination through optimization:How Facebook’s Ad delivery can lead to biased outcomes. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW), 1–30
Auxier B, Rainie L, Anderson M, Perrin A, Kumar M, Turner E (2019) Americans and privacy: concerned, confused and feeling lack of control over their personal information. Research report, Pew Research Center
Barth S, de Jong MDT (2017) The privacy paradox–investigating discrepancies between expressed privacy concerns and actual online behavior–a systematic literature review. Telemat Inform 34(7):1038–1058
Baruh L, Secinti E, Cemalcilar Z (2017) Online privacy concerns and privacy management: a meta-analytical review. J Commun 67(1):26–53
Baumann F, Lorenz-Spreen P, Sokolov IM, Starnini M (2020) Modeling echo chambers and polarization dynamics in social networks. Phys Rev Lett 124(4):048301
Biega AJ, Potash P, Daumé H, Diaz, F, Finck M (2020) Operationalizing the legal principle of data minimization for personalization. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, pp. 399–408
Cinelli M, Quattrociocchi W, Galeazzi A, Valensise CM, Brugnoli E, Schmidt AL, Zola P, Zollo F, Scala A (2020) The COVID-19 social media infodemic. Sci Rep 10(1):16598
Datta A, Datta A, Makagon J, Mulligan DK, Tschantz MC (2018) Discrimination in online personalization: a multidisciplinary inquiry. In Proceedings of the 1st Conference on Fairness, Accountability and Transparency, vol 81, pp. 20–34
Dienlin T, Trepte S (2015) Is the privacy paradox a relic of the past? An in-depth analysis of privacy attitudes and privacy behaviors. Eu J Soc Psychol 45(3):285–297
Digital, Culture, Media and Sport Committee (2019) Disinformation and ‘fake news’: Final report. Research report, House of Commons, U.K. Parliament
Directorate-General for Communication (2019) Special Eurobarometer 487a: The General Data Protection Regulation. Research report, European Commission
European Commission (2020a) Europe fit for the digital age: commission proposes new rules for digital platforms
European Commission (2020b) European democracy action plan: making EU democracies stronger
European Parliament (2016) Regulation (eu) 2016/679 of the european parliament and of the council of 27 april 2016 on the protection of natural persons with regard to the processing of personal data and on the free movement of such data, and repealing directive 95/46/ec (general data protection regulation)
Facebook (2021) Advertising policies. 9.a: Ads about social issues, elections or politics
Fairfield JAT, Engel C (2015) Privacy as a public good. Duke Law J 65(3):385–457
Garcia D (2017) Leaking privacy and shadow profiles in online social networks. Sci Adv 3(8):e1701172
Garcia D (2019) Privacy beyond the individual. Nat Human Behav 3(2):112–113
Hinds J, Joinson A (2019) Human and computer personality prediction from digital footprints. Curr Direct Psychol Sci 28(2):204–211
Hinds J, Joinson AN (2018) What demographic attributes do our digital footprints reveal? A systematic review. PLoS ONE, 13(11):e0207112
Hitlin P, Rainie L (2019) Facebook algorithms and personal data. Research report, Pew Research Center
Horwitz J, Seetharaman D (2020) Facebook executives shut down efforts to make the site less divisive. Facebook
Ipsos Mori (2020) Public attitudes towards online targeting-a report by Ipsos MORI for the Centre for Data Ethics and Innovation and Sciencewise. Research report, Ipsos Mori
Jaursch J (2020) Why EU must limit political micro-targeting. https://euobserver.com/opinion/148981
Kaiser J, Rauchfleisch A (2018) Unite the Right? How YouTube’s recommendation algorithm connects the U.S. far-right
Kokolakis S (2017) Privacy attitudes and privacy behaviour: a review of current research on the privacy paradox phenomenon. Comput Secur 64:122–134
Kosinski M, Wang Y, Lakkaraju H, Leskovec J (2016) Mining big data to extract patterns and predict real-life outcomes. Psychol Methods 21(4):493–506
Kozyreva A, Herzog S, Lorenz-Spreen P, Hertwig R, Lewandowsky, S (2020a) Artificial intelligence in online environments: representative survey of public attitudes in Germany. Research report. Max Planck Institute for Human Development. https://pure.mpg.de/pubman/faces/ViewItemOverviewPage.jsp?itemId=item_3188061
Kozyreva A, Herzog S, Lorenz-Spreen P, Hertwig R, Lewandowsky S (2020b) Künstliche intelligenz in online-Umgebungen: Repräsentative Umfrage zur öffentlichen Meinung in Deutschland [Artificial intelligence in online environments: representative survey of public attitudes in Germany]. Research report. Max Planck Institute for Human Development
Kozyreva A, Lewandowsky S, Hertwig R (2020c) Citizens versus the internet: Confronting digital challenges with cognitive tools. Psychol Sci Public Interest 21(3):103–156
Lewandowsky S, Smillie L, Garcia D, Hertwig R, Weatherall J, Egidy S, Robertson R, O’Connor C, Kozyreva A, Lorenz-Spreen P, Blaschke Y, Leiser M (2020) Technology and democracy: Understanding the influence of online technologies on political behaviour and decision-making. Research report. Publications Office of the European Union
Lorenz-Spreen P, Lewandowsky S, Sunstein CR, Hertwig R (2020) How behavioural sciences can promote truth, autonomy and democratic discourse online. Nat Human Behav 4(11):1102–1109
Matz SC, Kosinski M, Nave G, Stillwell DJ (2017) Psychological targeting as an effective approach to digital mass persuasion. Proc Natl Acad Sci USA 114(48):12714–12719
Mazarr MJ, Bauer RM, Casey A, Heintz SA, Matthews LJ (2019) The emerging risk of virtual societal warfare: social manipulation in a changing information environment. Research report, RAND Corporation
Newman N, Fletcher R, Schulz A, Andı S, Nielsen RK (2020) Reuters Institute digital news report 2020. Research report, University of Oxford, Reuters Institute for the Study of Journalism
Norberg PA, Horne DR, Horne DA (2007) The privacy paradox: personal information disclosure intentions versus behaviors. J Consum Aff 41(1):100–126
Persily N (2017) Can democracy survive the internet? J Democr 28(2):63–76
Rauchfleisch A, Kaiser J (2017) YouTubes Algorithmen sorgen dafür, dass AfD-Fans unter sich bleiben
Ribeiro FN, Saha K, Babaei M, Henrique L, Messias J, Benevenuto F, Goga O, Gummadi KP, Redmiles EM (2019) On microtargeting socially divisive ads: a case study of Russia-linked ad campaigns on Facebook. In FAT* ’19: Proceedings of the Conference on Fairness, Accountability, and Transparency, pp. 140–149
Sabbagh D (2020) Trump 2016 campaign ‘targeted 3.5m black Americans to deter them from voting’. The Guardian
Smith A (2018) Public attitudes toward computer algorithms. Research report, Pew Research Center
Speicher T, Ali M, Venkatadri G, Ribeiro FN, Arvanitakis G, Benevenuto F, Gummadi KP, Loiseau P, Mislove A (2018) Potential for discrimination in online targeted advertising. In Proceedings of the 1st Conference on Fairness, Accountability, and Transparency, in PMLR, vol. 81, pp. 1–15
Thompson SA, Warzel C (2021) They used to post selfies. Now they’re trying to reverse the election. The New York Times
Twitter (2021a) Permanent suspension of @realdonaldtrump. Twitter
Twitter (2021b) Political content. Twitter
Utz C, Degeling M, Fahl S, Schaub, F, Holz T (2019) (Un)informed consent: studying GDPR consent notices in the field. In 2019 ACM SIGSAC Conference on Computer and Communications Security, pp. 973–990
Wen H, Yang L, Sobolev M, Estrin D (2018) Exploring recommendations under user-controlled data filtering. In Proceedings of the 12th ACM Conference on Recommender Systems, pp. 72–76
Youyou W, Kosinski M, Stillwell D (2015) Computer-based personality judgments are more accurate than those made by humans. Proc Natl Acad Sci USA 112(4):1036–1040
Zarocostas J (2020) How to fight an infodemic. Lancet 395(10225):676
Zuboff S (2019) The age of surveillance capitalism. Profile Books
Zuiderveen Borgesius FJ, Möller J, Kruikemeier S, Fathaigh Ó, Irion R, Dobber K, Bodo T, de Vreese B (2018) Online political microtargeting: promises and threats for democracy. Utrecht Law Rev 14(1):82–96
Acknowledgements
The study was funded by the grant of the Volkswagen Foundation to RH, SL, and SH (Initiative “Artificial Intelligence and the Society of the Future”). SL was supported by a Research Award from the Humboldt Foundation in Germany while this research was conducted. The authors thank the Volkswagen Foundation for providing financial support and Dalia Research for conducting the survey. We are also very thankful to Deb Ain for editing the manuscript, to the two anonymous reviewers for their helpful comments, and to our colleagues at the Center for Adaptive Rationality for their feedback and productive discussions.
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
AK, PLS, SH, SL, and RH designed the study; AK, PLS, and SH managed and conducted research; SH analyzed data; AK, SH, PLS, SL, and RH wrote the paper. Correspondence concerning this article should be addressed to Anastasia Kozyreva, Center for Adaptive Rationality, Max Planck Institute for Human Development, Berlin. E-mail: kozyreva@mpib-berlin.mpg.de
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
41599_2021_787_MOESM1_ESM.pdf
Supplementary Information for: Public attitudes towards algorithmic personalization and use of personal data online: Evidence from Germany, Great Britain, and the United States
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kozyreva, A., Lorenz-Spreen, P., Hertwig, R. et al. Public attitudes towards algorithmic personalization and use of personal data online: evidence from Germany, Great Britain, and the United States. Humanit Soc Sci Commun 8, 117 (2021). https://doi.org/10.1057/s41599-021-00787-w
Received:
Accepted:
Published:
DOI: https://doi.org/10.1057/s41599-021-00787-w
This article is cited by
-
Overperception of moral outrage in online social networks inflates beliefs about intergroup hostility
Nature Human Behaviour (2023)
-
Personalization and localization as key expectations of digital health intervention in women pre- to post-pregnancy
npj Digital Medicine (2023)
-
A manifesto for applying behavioural science
Nature Human Behaviour (2023)
-
The effects of targeted political advertising on user privacy concerns and digital product acceptance: A preference-based approach
Electronic Markets (2023)
-
Segmentation of ethics, legal, and social issues (ELSI) related to AI in Japan, the United States, and Germany
AI and Ethics (2023)