Abstract
The study of urban and local politics in the United States has long been hindered by a lack of centralized sources of election data. We introduce a new database of about 78,000 candidates in 57,000 electoral contests that encompasses races for seven distinct local political offices in most medium and large cities and counties in the U.S. over the last three decades. This is the most comprehensive publicly-available source of information on local elections across the country. We provide partisan and demographic information about candidates in these races as well as electoral outcomes. This new database will facilitate a myriad of new research on representation and elections in local governments.
Similar content being viewed by others
Background & Summary
One of the most persistent challenges in the study of urban and local politics in the United States is the lack of information about local elections, candidates, and elected officials1,2. As a result, studies on local elections tend to focus on a single time period3, geographic unit4, or office5, rather than holistically examining variation across time, geography, and offices.
In this paper, we describe a new database of election returns from about 78,000 unique candidates in about 57,000 contests in 1,747 cities, counties, and school districts from 1989–2021. Our database is the most comprehensive publicly-available source of information on local elections across the entire country. It includes information about elections for mayors, city councils, county executives, county legislatures, sheriffs, prosecutors, and school boards. It also includes a host of supplemental data, including estimates of candidate partisanship, gender, race/ethnicity, and incumbency status. For many elections, it also includes information on the political characteristics of constituencies, such as their ideology and presidential voting patterns.
This new database will enable scholars to study a wide variety of research questions. It enables examination of whether politicians represent the demographic, partisan, and ideological characteristics of their constituents6,7,8. It also enables expanded work on the factors that affect local elections9,10,11. Moreover, it facilitates study of the incumbency advantage across election types, institutional contexts, and candidate characteristics12,13. Finally, this database enables scholars to expand the study of how elections shape a host of political outcomes such as policy5,14,15,16,17,18, political communication19, interest group activity20,21, and intergovernmental lobbying22.
Methods
This section describes how we constructed our database. Our target universe was all cities and counties with more than 50,000 people in the 2020 Census. There are 1,005 counties and 877 cities in our target universe. But many of these cities, especially in California, do not elect mayors, and most counties do not elect executives. Our data collection for school boards was more opportunistic. We also included district attorney (prosecutor) elections that had districts spanning multiple counties which we were not able to match to Census data. The database includes information on the vast majority of the cities and counties in our target universe. First, we describe how we assembled the raw election returns. Next, we describe how we appended supplemental data on candidates race/ethnicity, gender, and partisanship. Lastly, we discuss how we assembled supplemental data on the constituencies of many of the candidates in our database.
Election returns
The foundation for our data on election returns is previous work on mayoral elections5,14,15, county legislative elections16, sheriff elections18,23, prosecutor elections24,25, the MIT Election and Data Science’s Lab’s data on recent elections26, and the California statewide election database27. We built upon these datasets using several approaches. First, we expanded both the types of offices covered and the temporal coverage of these datasets. We worked with a team of research assistants who coded results from thousands of local elections based on city and county websites. In addition, we scraped data from the crowdsourced website OurCampaigns.com, statewide election websites where available, and some unofficial returns from newspaper archives. Where data sources overlapped in their coverage and conflicted, we prioritized administrative government records as the canonical source, then previously published datasets, then OurCampaigns.com, and then newspapers. These conflicts most likely occurred due to discrepancies between incomplete or provisional and official or final election results that are sometimes released at different times.
The resulting dataset of local election returns includes information on 57,139 contests and 77,853 unique candidates in 1,747 cities, counties, prosecutor districts, and school districts from 1989–2021 (Table 1 and Figs. 1, 2). It includes information about elections for mayors, city councils, county executives, county legislatures, sheriffs, prosecutors, and school boards. In many cases, we verified the validity of the election returns by cross-checking them across sources.
Supplemental information on candidates
We augmented the raw election returns with an array of supplementary information about individual candidates, including their partisanship (even in officially nonpartisan elections), gender, race/ethnicity, and incumbency status. In order to do this, we matched the election returns with a wide range of auxiliary data that enables us to estimate candidates’ partisanship, race/ethnicity, and gender. First, we sought to match each candidate to a record in two national voter files by name and location. Second, we sought to match each candidate with campaign finance-based ideology scores28. Third, we matched candidates that served in Congress or state legislatures to determine their party and roll-call based ideal points. We also matched many candidates in recent elections with their Ballotpedia profiles and Reflective Democracy Campaign data (https://wholeads.us/datasets/). Finally, we matched candidates with pictures we obtained from the Internet where possible.
Based on these data, we use Random Forests to predict the race/ethnicity, gender, and party identification of candidates in the data29. Random Forest is a collection of identically distributed classification trees, where each tree is comprised of a bootstrap sample from the training set and is grown using a recursive splitting rule that minimizes prediction error. To further reduce the correlation among decision trees, only a fraction of randomly selected covariates are used in each tree during the recursive splitting. Once the set of decision trees has been grown on the bootstrapped samples, the unsampled cases from the test set (“out-of-bag” sample) are used to generate predictions. In particular, the predicted probabilities for each class is generated based on the classification from the collection of decision trees and the final predicted class is generated based on a majority vote-i.e., the most frequent class.
Relative to other statistical learning approaches, Random Forests yield several desirable properties including high accuracy, robustness to outliers and noise, internally unbiased estimate of the generalization error, efficient computation, and the ability to handle many predictors. For example, comparing Random Forest with different versions of logistic regressions in class-imbalanced data, previous work finds that Random Forest provides significantly more accurate predictions in out-of-sample data than any of the logistic models30. We also ran several alternative models including ridge and lasso regularizations of generalized linear models, gradient boosting machines (GBM), as well as a super-learner, an ensemble method that accounts for both generalized linear and tree-based models31. We confirmed that Random Forest generates much higher accuracy rates than any of the generalized linear models and yields near-identical accuracy rates as GBM and super-learner.
For partisanship, we collapsed our election data by name, geography, and office, which resulted in approximately 61,894 unique observations (only including data with some information on partisanship). Roughly 49% of them (N = 30,115) had true party identification categorized as either Democrat or Republican (coded as a 0 or 1 for prob_democrat). These are based on ballot returns for either the office in our data or, in a few cases, another office. We use this variable as our target outcome to train a Random Forest for binary classification with 10-fold cross-validation scheme. For model features, we use potentially noisy indicators from the voter file, campaign finance data32, the Reflective Democracy Campaign, and partisanship from other sources (e.g., Ballotpedia).
For race/ethnicity, collapsing our data by name, city, and office resulted in 75,591 unique observations (only including data with some information on race/ethnicity). Roughly 18% of them (N = 13,671) had true racial identification collected from various external sources, including official lists from non-profit organizations and human-labeled data from Amazon’s Mechanical Turk (coded as a 0 or 1 for prob_black and other race/ethnicity categories). Similar to the racial groups used in the Census Bureau and other commonly used prediction methods33, we divide race into five categories: White, Black, Latino, Asian, and Other. We use this variable as our target outcome to train a Random Forest for multi-class classification with 10-fold cross-validation scheme. For model features, we use information from the voter files, surname-based Bayesian racial predictions33, first and last name-based Bayesian predictions34, and predictions from a pre-trained convolutional neural network model based on images of public officials35. This approach produces more accurate predictions of race and ethnicity in our sample than existing methods that rely on names and/or geography alone36.
For gender, collapsing our data by name, city, and office resulted in 77,352 unique observations (only including data with some information on gender). Roughly 13% of them (N = 10,438) had true gender identification collected by the Reflective Democracy Campaign (coded as a 0 or 1 for prob_female). We use this variable as our target outcome to train a Random Forest for binary classification with 10-fold cross-validation scheme. For model features, we use information from the voter files, first-name-based gender predictions37, and gender estimates in the DIME database32.
Overall, our approach generated a probabilistic estimate of most candidates’ partisanship, race/ethnicity, and gender that can be used to study representation, elections, and policymaking.
Constituency-level data
We augmented the election returns with a variety of information about many candidates’ constituencies. We included information about the ideological preferences of each city and county in our dataset8. We also included recent presidential election results for most cities and counties38,39,40. In addition, we assembled a new collection of shapefiles for many city council and county legislative districts. This enabled us to estimate presidential election returns in many local governments’ district-level constituencies by overlaying precinct-level presidential returns on top of the district shapefiles41.
Data Records
The complete American Local Government Elections Database is available for download on OSF42. The dataset can be accessed in different formats such as comma-separated files (.csv, for easy access in programs such as Stata, R, Python, or Excel) and in compressed R data files (.rds, for easy access using the R programming language).
The dataset includes two sets of files. First, we include candidate-level data (ledb_candidatelevel.rds and ledb_candidatelevel.csv), in which each observation is a candidate running in a discrete contest along with associated information about that candidate and electoral contest as well as its results. We provide a number of variables at the candidate level (Table 2).
Second, we include constituency-level data, in which each observation is at the level of a government jurisdiction. These include data for cities (cities_constituency_data.csv), counties (counties_constituency_data.csv), school districts (schools_districts_constituency_data.csv), city council districts (city_council_districts_constituency_data.csv), and county legislative districts (county_leg_districts_constituency_data.csv). These are available for nearly all cities and counties as a whole. We also have them available at the city council district-level in about 150 cities and the county legislative district level in about 130 counties (Table 3).
Technical Validation
In this section, we discuss a number of technical validations of our data. We validate a number of aspects of the candidate-level data, including the vote totals for individual candidates and our imputations for candidates’ partisanship, race/ethnicity, and gender.
Validation of election data
We validate our elections data by comparing the consistency of candidate’s vote totals across secondary sources, and obtaining administrative data where possible.
Validation of estimates of partisanship, race, and gender
A key contribution of our database is a set of estimates for candidates’ partisanship, race/ethnicity, and gender. In some cases, these are based on observed data on these variables. But in other cases, we use Random Forest algorithms for these variables based on potentially noisy indicators as described above. Thus, it is important to validate our estimates of these variables.
We first validate our Random Forest model of race/ethnicity using 20% of the data with outcomes as a test set. Table 4 shows the sensitivity (true positive rate), specificity (true negative rate), precision (positive predictive value), and F-1 score. Note that F-1 score is the harmonic mean of precision and sensitivity-i.e., \(2\cdot \frac{precision\cdot sensitivity}{precision+sensitivity}\).
The F-1 score for our model of race/ethnicity is 0.929, which is significantly higher than other racial classification methods commonly used in academic research33,34,35. Our results also show substantially high sensitivity scores across all racial categories, particularly among Asians and Blacks that are generally associated with high false negative rates33. We also examine the Receiver Operating Characteristic (ROC) curves and find the area under the ROC curve (AUC) to be at least as high as 0.97 across all racial categories, indicating an outstanding classification success.
Next, we perform the same validation check for our Random Forest model of gender using 20% of the data with outcome as a test set. Table 5 shows the results of our model performance. The F-1 score for our model of gender is approximately 0.993 and the AUC score is above 0.99, indicating a very high predictive ability. We also check whether our gender predictions for the subset of our candidate data that are female mayors (i.e. winning mayoral candidates) against recent estimates of mayoral gender from the Center for American Women and Politics (CAWP)43,44. We find that out of 103 candidates in our data that matched to CAWP’s list of women mayors, our random forests models estimate that two are men. This is similar to the accuracy rate reported in Table 5. Through a manual check of CAWP’s lists of 661 unique mayors in 2021 and 2022 who are designated as women, we also find that seven of these mayors were actually men, yielding an accuracy rate of 0.989 - similar to our overall accuracy rate.
Lastly, we perform the same validation check for our Random Forest model of partisanship using 20% of the data with outcome as a test set. Table 6 shows the results of our model performance. The F-1 score for our model of partisanship is approximately 0.903 and the AUC score is above 0.95, indicating a very high predictive ability.
The data files we make available include the final estimated probabilities that each candidate falls in each category of partisanship, gender, and race/ethnicity. This enables scholars to make their own decisions about how to use the results of our imputation models in downstream analyses.
Usage Notes
As the most comprehensive record of local election outcomes and information about local political candidates, our database presents myriad opportunities for researchers looking to expand knowledge about democracy in subnational politics. There are several general points scholars should consider as they use the data.
First, the data is much more comprehensive than previous datasets on local elections. But the target universe is not totally comprehensive. It only includes cities, counties, and school districts with a population of at least 50,000 in 2020. Moreover, it is missing some elections in this target universe where we were unable to find data. In some cases, there may be raw election data for smaller geographic units available from the sources we discussed earlier5,18,23,24,25,26,27.
Second, researchers should note that many prosecutor districts span counties. In addition, one of our raw sources of prosecutor election data included unique numeric district identifiers, but no district numbers or names25. This makes it more difficult to determine incumbency status and other characteristics of prosecutor candidates. In cases where we were unable to properly match a contest to its county or district, we include the original district code from the source data and a synthetic fips code constructed by concatenating state codes with the district code from the source data.
Third, as described above, we imputed the partisanship and race of many candidates. This enables scholars to study a wide variety of research questions related to elections and representation. In the data, we provide the predicted probability of these imputations. For applications that study the causes and consequences of individual elections, scholars may choose to only include candidates with high predicted probabilities of being in particular racial, partisan, or gender categories45. For example, recent research has examined the close link between partisan voting patterns in national and subnational elections to argue that local races are increasingly nationalized46,47. Our database allows scholars to test both the causes and consequences of this nationalization across many different types of elected office. For this analysis, researchers may decide to classify candidates into the partisan group with the highest probability (e.g., only including candidates with known partisanship or ones with >90% in our imputation model).
Scholars could also use our database to examine the aggregate characteristics of candidates and election officials. For instance, an important question in the study of American elections is the under-representation of women and non-white racial groups at various levels of government7,48,49,50,51,52,53,54,55. While prior work suggests that women’s under-representation in local governments mirrors their underrepresentation in Congress and state legislatures, and that the representation of women in local government appears to have plateaued over the past two decades, the limits of previous data have prevented researchers from examining the barriers to both gender and racial parity in local politics6. Our local elections data enable researchers to assess questions about barriers to both women and racial minorities in politics at the subnational level by vastly expanding the temporal and geographic scope of previous data on local candidates and officeholders, as well as expanding the scope of previous datasets. When estimating the racial, gender, or partisan composition of a particular geographic unit, recent research finds that it is more accurate to average up the predicted probabilities for all individuals within a geographic unit rather than aggregating after assigning each individual to a single category based on the highest predicted probability56. Because our data provides the full distribution of predicted probabilities across each category for partisanship, race/ethnicity, and gender, researchers are able to examine topics like representation and turnout while avoiding substantial error in estimating the demographic composition.
A brief examination of over-time patterns of candidates’ race and gender in our data illustrates how it could be used to assess the institutional and contextual determinants of descriptive representation. Figure 3 shows the relative representation of women, Blacks, Hispanics, and Asian-Americans based on the ratio between the share of local officeholders of each type and their fraction of the population. This brief demonstration indicates that women remain under-represented in the majority of local offices, with the percentage of winning candidates under their percentage in the population for all offices except school boards. The breadth of our data allows us to expand this usage beyond single offices or small time spans, and shows that there are dramatic differences in patterns across offices in women’s representation. Women are most under-represented in sheriff elections, and tend to be best represented in school board elections, in line with recent work on California57. Similarly, our data on multiple racial groups allows usage of our data to examine under-representation of multiple groups across offices. Figure 3 suggests that Hispanics and Asian-Americans are particularly under-represented and that descriptive representation is especially poor among sheriffs, but that city councils, in particular, consist of Black legislators at rates roughly proportional to population demographics. White officeholders, meanwhile, are overrepresented in every local office across the last three decades.
Researchers may conduct many other analyses using variables described earlier in the manuscript, as well as by combining our data with additional institutional or contextual data. We encourage researchers to fully harness our data to both further describe the heterogeneity in these over-time and between-office trends, as well as further examine the causes and consequences of local elections.
Code availability
The replication code for the two demonstrations of our data is publicly available on OSF42, and can be used under a CC-BY license.
References
Trounstine, J. Representation and accountability in cities. Annual Review of Political Science 13, 407–423 (2010).
Warshaw, C. Local elections and representation in the United States. Annual Review of Political Science 22, 461–479 (2019).
Sances, M. W. Ideology and vote choice in US mayoral elections: Evidence from facebook surveys. Political Behavior 40, 737–762 (2018).
Arnold, R. D. & Carnes, N. Holding mayors accountable: New York’s executives from Koch to Bloomberg. American Journal of Political Science 56, 949–963 (2012).
Ferreira, F. & Gyourko, J. Do political parties matter? evidence from US cities. Quarterly Journal of Economics 124, 399–422 (2009).
Holman, M. R. Women in local government: What we know and where we go from here. State and Local Government Review 49, 285–296 (2017).
Schaffner, B. F., Rhodes, J. H. & La Raja, R. J. Hometown Inequality: Race, Class, and Representation in American Local Politics (Cambridge University Press, New York, 2020).
Tausanovitch, C. & Warshaw, C. Representation in municipal government. American Political Science Review 108, 605–641 (2014).
Hajnal, Z. & Trounstine, J. Where turnout matters: The consequences of uneven turnout in city politics. The Journal of Politics 67, 515–535 (2005).
de Benedictis-Kessner, J. & Warshaw, C. Accountability for the local economy at all levels of government in United States elections. American Political Science Review 114, 660–676 (2020).
Hopkins, D. J. & Pettingill, L. M. Retrospective voting in big-city US mayoral elections. Political Science Research and Methods 6, 697–714 (2018).
de Benedictis-Kessner, J. Off-cycle and out of office: Election timing and the incumbency advantage. Journal of Politics 80, 119–132 (2018).
Trounstine, J. Evidence of a local incumbency advantage. Legislative Studies Quarterly 36, 255–280 (2011).
Gerber, E. R. & Hopkins, D. J. When mayors matter: Estimating the impact of mayoral partisanship on city policy. American Journal of Political Science 55, 326–339 (2011).
de Benedictis-Kessner, J. & Warshaw, C. Mayoral partisanship and municipal fiscal policy. Journal of Politics 78, 1124–1138 (2016).
de Benedictis-Kessner, J. & Warshaw, C. Politics in forgotten governments: The partisan composition of county legislatures and county fiscal policies. Journal of Politics 82, 460–475 (2020).
Holman, M. R., Farris, E. M. & Sumner, J. L. Local political institutions and first-mover policy responses to COVID-19. Journal of Political Institutions and Political Economy 1, 523–541 (2020).
Thompson, D. M. How partisan is local law enforcement? evidence from sheriff cooperation with immigration authorities. American Political Science Review 114, 222–236 (2020).
de Benedictis-Kessner, J. Strategic partisans: Electoral motivations and partisanship in local government communication. Journal of Political Institutions and Political Economy 2, 227–248 (2021).
Anzia, S. F. Looking for influence in all the wrong places: How studying subnational policy can revive research on interest groups. Journal of Politics 81, 343–351 (2019).
Anzia, S. F. Local Interests: Politics, Policy, and Interest Groups in US City Governments (University of Chicago Press, Chicago, 2022).
Payson, J. A. The partisan logic of city mobilization: Evidence from state lobbying disclosures. American Political Science Review 114, 677–690 (2020).
Zoorob, M. There’s (rarely) a new sheriff in town: The incumbency advantage for local law enforcement. Electoral Studies 80, 102550 (2022).
Hessick, C. B. & Morse, M. Picking prosecutors. Iowa L. Rev. 105, 1537 (2019).
Krumholz, S. The effect of district attorneys on local criminal justice outcomes. SSRN Working Paper. https://doi.org/10.2139/ssrn.3243162 (2019).
Baltz, S. et al. American election results at the precinct level. Scientific Data 9, 651 (2022).
CEDA. California elections data archive (CEDA). https://csu-csus.esploro.exlibrisgroup.com/esploro/outputs/99257830890201671 (2020).
Bonica, A. Mapping the ideological marketplace. American Journal of Political Science 58, 367–386 (2014).
Breiman, L. Random forests. Machine Learning 45, 5–32 (2001).
Muchlinski, D., Siroky, D., He, J. & Kocher, M. Comparing random forest with logistic regression for predicting class-imbalanced civil war onset data. Political Analysis 24, 87–103 (2016).
van der Laan, M., Polley, E. & Hubbard, A. Super learner. Statistical Applications of Genetcs and Microbiology 6 (2007).
Bonica, A. Database on ideology, money in politics, and elections: pre-release version 3.0. Stanford, CA: Stanford University Libraries. https://data.stanford.edu/dime (2023).
Imai, K. & Khanna, K. Improving ecological inference by predicting individual ethnicity from voter registration records. Political Analysis 24, 263–272 (2016).
Imai, K., Olivella, S. & Rosenman, E. T. Addressing census data problems in race imputation via fully bayesian improved surname geocoding and name supplements. Science Advances 8, eadc9824 (2022).
Parkhi, O. M., Vedaldi, A. & Zisserman, A. Deep face recognition. In Xie, X., Jones, M. W. & Tam, G. K. L. (eds.) Proceedings of the British Machine Vision Conference, 41.1–41.12 (2015).
Lee, D. D. I. & Velez, Y. R. Measuring descriptive representation at scale: Methods for predicting the race and ethnicity of public officials. https://doi.org/10.31219/osf.io/tpsv6 (2023).
Mullen, L. gender: Predict gender from names using historical data. R package version 0.6.0, https://github.com/lmullen/gender (2021).
Ansolabehere, S. & Rodden, J. Harvard election data archive. http://projects.iq.harvard.edu/eda (2012).
Voting and Election Science Team. 2016 precinct-level election results. Harvard Dataverse https://doi.org/10.7910/DVN/NH5S2I (2018).
Voting and Election Science Team. 2020 precinct-level election results. Harvard Dataverse https://doi.org/10.7910/DVN/K7760H (2020).
Lee, D. D. I. & Velez, Y. R. Rising tides or political ripcurrents? gentrification and minority representation in 166 cities. Urban Affairs Review (2023).
de Benedictis-Kessner, J., Lee, D. D. I., Velez, Y. R. & Warshaw, C. American local government elections database. OSF. https://doi.org/10.17605/OSF.IO/MV5E6 (2023).
Center for American Women and Politics. Women mayors in U.S. cities 2021. https://cawp.rutgers.edu/facts/levels-office/local/women-mayors-us-cities-2021 (2021).
Center for American Women and Politics. Women mayors in U.S. cities 2022. https://cawp.rutgers.edu/facts/levels-office/local/women-mayors-us-cities-2022 (2022).
Li, Y., Hyun, M. & Alvarez, R. M. Why do election results change after election day? the “Blue Shift” in California elections. Political Research Quarterly 75, 860–874 (2022).
Hopkins, D. J. The Increasingly United States: How and Why American Political Behavior Nationalized (University of Chicago Press, Chicago, IL, 2018).
Kuriwaki, S. The swing voter paradox: Electoral politics in a nationalized era. Doctoral dissertation, Harvard University Graduate School of Arts and Sciences. https://dash.harvard.edu/handle/1/37368520 (2021).
Barnes, T. D. & Holman, M. R. Gender quotas, women’s representation, and legislative diversity. Journal of Politics 82, 1271–1286, https://doi.org/10.1086/708336 (2020).
Bernhard, R. & de Benedictis-Kessner, J. Men and women candidates are similarly persistent after losing elections. Proceedings of the National Academy of Sciences 118 (2021).
Bernhard, R., Shames, S. & Teele, D. L. To emerge? Breadwinning, motherhood, and women’s decisions to run for office. American Political Science Review 115, 379–394 (2021).
Clayton, A., O’Brien, D. Z. & Piscopo, J. M. All male panels? representation and democratic legitimacy. American Journal of Political Science 63, 113–129 (2019).
Grofman, B. & Davidson, C. The effect of municipal election structure on black representation in eight southern states. In Davidson, C. & Grofman, B. (eds.) Quiet Revolution in the South: The Impact of the Voting Rights Act, 1965-1990, 301–334 (Princeton University Press, Princeton, NJ, 1994).
Hajnal, Z. L. Dangerously Divided: How Race and Class Shape Winning and Losing in American Politics (Cambridge University Press, New York, 2020).
Lawless, J. L. Female candidates and legislators. Annual Review of Political Science 18, 349–366 (2015).
Shah, P. R., Marschall, M. J. & Ruhil, A. V. Are we there yet? the Voting Rights Act and black representation on city councils, 1981–2006. Journal of Politics 75, 993–1008 (2013).
DeLuca, K. & Curiel, J. A. Validating the applicability of bayesian inference with surname and geocoding to congressional redistricting. Political Analysis 1–7 (2022).
Anzia, S. F. & Bernhard, R. Gender stereotyping and the electoral success of women candidates: New evidence from local elections in California. British Journal of Political Science (2021).
Enamorado, T., Fifield, B. & Imai, K. Using a probabilistic model to assist merging of large-scale administrative records. American Political Science Review 113, 353–371 (2019).
Tausanovitch, C. & Warshaw, C. Measuring constituent policy preferences in congress, state legislatures, and cities. Journal of Politics 75, 330–342 (2013).
Warshaw, C. & Tausanovitch, C. Subnational ideology and presidential vote estimates (v2022)., Harvard Dataverse, https://doi.org/10.7910/DVN/BQKU4M (2022).
Acknowledgements
We are grateful for many conversations about this project with Dan Thompson. We appreciate excellent research assistance on this project from Camila Alvarez Bisbe, Caitlin Berg, Noam Brenner, Tom Cawley, Cole Dushin, Alexander Hupp, Josh Koppel, Jace Knie, Cory Maks-Solomon, Jeremy Marsh, Daniel Perez, Rob Pressel, John Ramsey, Annie Salyers, Anmol Sapru, Josiah Selagea, Mikaela Rose Tajo, Ariel Wexler, and Yiling Yao. We are also grateful to Fernando Ferreira, Joseph Gyourko, Elisabeth Gerber, Dan Hopkins, Dan Thompson, Carissa Byrne Hessick, Michael Zoorob, and Sam Krumholz for generously sharing data on elections they had collected on various local offices. We appreciate feedback on earlier versions of this manuscript from audiences at the 2021 Conference on Election Sciences, Reform, and Administration. This work was supported by funding from the MIT Election Data and Science Lab, the Russell Sage Foundation, and the Boston University Initiative on Cities.
Author information
Authors and Affiliations
Contributions
C.W. and J.D.B.K. conceived the overall project. C.W. and J.D.B.K. gathered most of the raw elections data and auxiliary data. Y.V. gathered data on the race/ethnicity of city councilors. D.D.I.L. and Y.V. developed the underlying approach for imputing race/ethnicity. D.D.I.L. ran the Random Forests imputation models for both race/ethnicity and partisanship. C.W., D.D.I.L. and Y.V. gathered the GIS files of city council and county legislative maps that were used to estimate the constituency-level datasets. All authors contributed equally to the drafting of the manuscript and reviewed the final draft.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
de Benedictis-Kessner, J., Lee, D.D.I., Velez, Y.R. et al. American local government elections database. Sci Data 10, 912 (2023). https://doi.org/10.1038/s41597-023-02792-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-023-02792-x