Citizens cast votes in a second-round primary of the 2017 French presidential election. Credit: Sebastien Nogier/EPA

For the second time since 2000 and at least the fourth in the history of the United States, the winner of the presidential election did not win the popular vote. The framers of the US Constitution established the electoral-college system deliberately. To prevent a few large states from dominating outcomes, they blunted the voting power of people within them.

The United Kingdom has a similar system. In 2005, for example, Britain's Labour Party won 55% of the seats in the House of Commons with only 35% of the votes cast.

In the French presidential race, by contrast, the winner always receives a majority of the votes cast. If neither of the two top vote recipients from a first-round election receives a majority, voters choose between the two candidates in a final 'run-off' election. But such methods can lead to a winner who may not be the most preferred candidate overall, depending on how voters pick candidates in the preliminary round.

Which is the best voting system? It depends on the circumstances. The study of choices by groups — social-choice theory — seeks to understand how different rules affect outcomes, and how well those rules perform against dozens of criteria. But there have been few practical and systematic studies — especially head-to-head comparisons of different rules and how they influence collective decisions, from governments to scientific societies.

Poor understanding

Voting, weighting, ranking, scoring and grading mechanisms are poorly understood by most people who apply them. The biases and limitations of each method are underappreciated. We pick a good wine with the help of our social network, sommeliers, magazine reviews, point scales and star ratings. But each person weighs attributes differently, leading to different scores for the same wines. Similar issues arise at every level of society, from choosing houses to picking spouses. Science, too, faces collective-choice challenges. Which country should host the proposed International Linear Collider, for instance? Which space missions should NASA undertake? Which medications, foods and consumer goods should be approved?

Bad group decisions harm research, the economy and society. For example, the way in which honorific and professional societies shortlist candidates for elections preserves the current membership structure rather than diversifying it. Voting rules can aid or discourage public participation — the cornerstone of a democratic process. Some voting methods favour centrists, others extremes.

Improving collective decision-making requires a new type of research: direct comparisons of various voting rules in different settings. Some of this can be done using computer simulations. Much will be learned by testing how decisions are made in real situations.

Picking winners

Even in this age of machine learning and social media we rely on a handful of basic choice mechanisms. A group may want to select one winner, such as for Pope, or pick the top few of many candidates, for say, a city council or a corporate board. Students are graded to evaluate excellence; apps, hotels, films and cars are rated by users to provide information to consumers.

The United Nations General Assembly: collective decision-making in action. Credit: Albin Lohr-Jones/Pacific Press/Getty

All mechanisms have two key features: what type of voter input they use (say, affirmation, or letter or numeric grades) and how those inputs are aggregated (mostly totals, averages or medians). Some methods are more expressive than others. With, say, five candidates, first-past-the-post voters have only six different options or 'words' to use, including a blank ballot. A 'rank order list' (wherein voters rank options in a numerical order of preference) has 120 different arrangements for the same five candidates. A system with six grades (such as the standard academic A, B, C, D, E, F) provides more than 7,776 combinations.

In the latest US election, the inability of voters to choose between two seemingly unpopular presidential candidates — Hillary Clinton and Donald Trump — may have lowered turnout to less than 60% of eligible voters, and less than in 2008 or 2004 (other factors may have also contributed to the reduced vote totals). Neither of the two major candidates received 17 million votes in the primary elections out of a total of 232 million eligible voters. Thus, only about 7% of eligible voters determined the final candidate slate.

Weighing up methods

Social-choice theorists evaluate selection methods using dozens of criteria. For example, 'join consistency' requires that an election result is unchanged if two or more districts are merged. As the differing winner of the popular vote in the 2016 US election shows, the US presidential-election system fails this test, as do many other voting systems.

The 'rank reversal rule' states that the ranking of any two candidates, A and B, should not change if C enters or leaves the race1. A notable violation is from 2000, when Green Party candidate Ralph Nader captured a few per cent of the vote in Florida, giving the election to George W. Bush (over Al Gore). Gore would have won either if Nader was not in the race or if Florida had used a run-off election. But a run-off approach can also lead to rank reversal.

Some of these criteria conflict — meeting one guarantees failing another or they confound the electorate's intention. Many situations invite strategic voting, wherein blocs of voters can alter the outcome by voting for someone other than their true preference. In a run-off ballotage, voters preferring one candidate may elevate a weaker opponent into the final round, assuring the victory of their favourite.

Voting systems that sum or average the points assigned to candidates — such as the rank-order ballots used in Australian elections — exacerbate these risks. Major Australian political parties hand out strategies so voters may game such opportunities.

Yet there have been few attempts to compare different methods for reaching a group decision. Voting methods can be tested in real elections by getting voters to fill out two types of ballot — an official one using established voting rules and a test ballot with alternative ways for voters to express themselves. One can then ask the voters which method they found easiest to use, which they understood, trusted and possibly preferred to use in future.

Voting methods can be tested in real elections by getting voters to fill out two types of ballot.

Such comparisons in wider political contexts are rare. In one example, members of the Society for Social Choice and Welfare were asked to rank three candidates for their president as well as voting for one using the official voting rule. Later analysis revealed that the different voting rules elected different candidates2,3.

In a larger experiment, voters from several districts in the 2007 French presidential election were asked by researchers Michel Balinski and Rida Laraki to fill out a separate ballot in which they graded each candidate. The researchers tested the time needed to vote (about one minute per ballot), preferences (voters found this 'majority judgement' system easy to use and appreciated the increased vocabulary), and how the official vote, majority judgement and other outcomes compared4.

The real French presidential election results showed two finalists, who ranked second and third in the polls, in a run-off against each other. The test system would have selected the candidates most preferred in the polls. The difference was not a defect of polling accuracy but rather the result of widespread strategic voting in the first round that was widely discussed before election day in the French media4.

We — an engineer, an economist and a biologist — have confronted similar issues in our own work. We developed a decision-support tool for public officials, corporate executives and donors to prioritize potential vaccines for research-and-development investment5,6,7.

Our software assigns a weight to each of a list of vaccine candidates based on a range of factors, from cost-effectiveness to target population and disease characteristics. The final scores are based on the attributes that a decision-maker values most in domains such as health, economic, demographic, business, operations and policy. And, like most such aids, it is written for individuals, not groups. Almost all its users have asked us for guidance on how to create prioritized lists by combining individual preferences in, say, an advisory committee.

To learn more about how groups use our software, we carried out a trial (see ‘Prioritization pilot’). We compared some common voting methods to see whether the resulting recommendations differed (they did) and which approach people preferred. Our participants found it easier and more expressive of their views to evaluate the merits of the options using letter grades rather than numerical ranks. Although this is but one example, and not definitive, more such experiments would help to guide groups in other scenarios.

boxed-text

Field trials

We cannot help groups and society to choose better selection mechanisms by continuing to study which voting systems pass or fail theoretical criteria. Social-choice research needs to include human factors such as ease and preference of use, and effects on voter participation — a key element in any democratic society. First, it should focus on practical ways to assess the performance and impact of various mechanisms under different conditions.

Field-testing different selection methods in parallel with real voting situations should determine, for example, how often they produce different results and how much voters understand each approach, as well as what the potential consequences are. Such information can help groups to pick which rule would be best for them. Computer simulations could be used to analyse distributions of public preferences, and also situations in which conflicting voting problems actually occur.

Two good places to start in science, engineering and public health are: peer-review panels that prioritize research proposals for funding decisions; and professional societies choosing members and officers. Testing in other organizations — from corporate boards to religious congregations and performance awards in sports, literature, arts, and films — would broaden understanding of the human factors at play.

We urge government agencies and philanthropic organizations to fund such practical research. Learning more about choice mechanisms will have widespread social and economic benefits, from better presidents to better science.

Table 7.41541 Table