Accuracy and Security in Voting Systems - princeton.edu · Accuracy and Security in Voting ... They...

43
Accuracy and Security in Voting Systems Henry E. Brady Professor of Political Science and Public Policy Iris Hui Department of Political Science University of California, Berkeley May 1, 2008 Note: This paper will appear in the forthcoming volume, Mobilizing Democracy: A Comparative Perspective on Institutional Barriers and Political Obstacles, James Johnson, Jack Knight, Margaret Levi and Susan Stokes (editors), Russell Sage Publications, New York. Disclosure and Acknowledgements: The senior author was involved as a consultant in several court cases mentioned in this paper, and he also so testified before the Carter-Baker Commission on these issues. Our thanks to James Dickson, Jim Johnson, Jan Leighley, Margaret Levi, Deborah Markowitz, Ted Selker, Regina Smyth, Daniel Tokaji, David Wagner, and the participants in “Practicing Democracy” on April 27, 2007 at the Ash Institute at Harvard University, and two anonymous referees for very helpful comments.

Transcript of Accuracy and Security in Voting Systems - princeton.edu · Accuracy and Security in Voting ... They...

Accuracy and Security in Voting Systems

Henry E. Brady

Professor of Political Science and Public Policy

Iris Hui

Department of Political Science

University of California, Berkeley

May 1, 2008

Note: This paper will appear in the forthcoming volume, Mobilizing Democracy: A Comparative Perspective on Institutional Barriers and Political Obstacles, James Johnson, Jack Knight, Margaret Levi and Susan Stokes (editors), Russell Sage Publications, New York. Disclosure and Acknowledgements: The senior author was involved as a consultant in several court cases mentioned in this paper, and he also so testified before the Carter-Baker Commission on these issues. Our thanks to James Dickson, Jim Johnson, Jan Leighley, Margaret Levi, Deborah Markowitz, Ted Selker, Regina Smyth, Daniel Tokaji, David Wagner, and the participants in “Practicing Democracy” on April 27, 2007 at the Ash Institute at Harvard University, and two anonymous referees for very helpful comments.

I. Introduction

“Trust in Paper” proclaims the May 5 2007 editorial in the New York Times congratulating Florida for

getting rid of electronic voting machines which last year “somehow lost 18,000 votes” in Sarasota

County. “The new law will eliminate touch-screen voting in favor of the more trustworthy optical-

scanning system. Unlike touch screens, optical-scanning machines are based on paper.” So “by next

summer Floridians should be a lot more confident that whenever they vote, their votes will finally be

counted correctly.” (New York Times, May 5, 2007, page A26).

Reality is more complicated than this rhetoric suggests. Many of the 18,000 votes that were lost

in Sarasota Florida were probably lost because of poor ballot design (Frisina et al 2007; Mebane & Dill

2007) – a problem that could have (and has) occurred with paper ballots—most famously with the

“butterfly ballot” in Palm Beach County Florida—as well as with electronic systems. And while paper

ballots may help with recounts when there are fears that computers could be programmed poorly or

maliciously, paper trails do nothing to protect voters from systems that confuse them or fail to record

votes in the first place.

There are two basic challenges for voting machines: accurately recording people’s votes and

keeping them secure once they have been cast by the voter. Both are equally important, but attention has

shifted since 2000 from accuracy to security, with the result that at least half the problem is being papered

over with simplistic rhetoric. Moreover efforts to improve voting systems have been stymied by

inadequate legislation, the limitations of legal action, and the sclerosis of American federalism. We

describe the sad course of voting reform1, and we provide evidence showing that accuracy must be taken

at least as seriously as security. The next chapter in this book by Charles Stewart provides a

comprehensive overview of how we might do better.

Accuracy and Punch Cards, Security and Electronic Voting – The extraordinarily close

election in Florida in 2000 exposed numerous problems with America’s machinery of democracy,2 but

voting systems—the methods for recording and counting votes—performed especially poorly. Confusing

ballot forms in Palm Beach County and Duval County led to thousands, if not tens of thousands, of

1

mistakes. The butterfly ballot caused at least two thousand Democratic voters in Palm Beach to vote for

Reform Party candidate Pat Buchanan when they meant to vote for Democratic Presidential candidate Al

Gore (Wand et al 2001). A voting booklet instruction that said “vote all pages” and a ballot that spread

presidential candidates over two pages may have caused thousands of voters in Duval County to vote for

two presidential candidates, thus spoiling their ballot for the presidential contest. In many counties in

Florida, especially in minority areas, there were substantially fewer valid votes – sometimes as many as

nine percent fewer – in the presidential contest than the number of voters who went to the polls. These

ballots without valid votes—called “residual votes”—seemed especially prevalent in counties using

Votomatic style punch cards, central-count optical scan systems, and Datavote punchcards.3 Finally,

recounts suggested that Votomatic style punch card systems and central count optical scan systems were

especially prone to varying counts depending upon the interpretation of ambiguous marks such as hanging

chads, smudged ballots, and inadequately marked ballots. Many people concluded from these

experiences that using different ballots and methods for recording votes led to serious problems with the

accuracy and fairness of voting systems. In fact, in Bush v. Gore—the case that decided the 2000

election—the Supreme Court argued explicitly that it was unfair—it violated equal protection

guarantees—to have one standard for counting votes in one county and a different standard in another

county.4

Since 2003, another problem, the security and trustworthiness of voting systems has come to the

fore based upon concerns that Direct Record Electronic (DRE) voting systems—similar to Automatic

Teller Machines (ATMs)—might be subject to computer hacking or to malfunctions. These DRE systems

were initially heralded as “solutions” to the problem of accuracy and fairness: They prevent people from

“overvoting” for more than one candidate or proposition in a contest. By providing reminder that citizens

had not voted in a contest, they reduce “undervoting” from people failing to choose an option in a contest.

They replace fuzzy marks on paper ballots or punchcards with crisp “yes” or “no” computerized bits, and

they provide more accessible voting machines for the disabled. But soon some skeptical computer

scientists began to argue that these systems could be untrustworthy due to poorly written computer code

2

or insecure due to computer hacking or illicit code. Consequently, these experts argued that they should

be supplemented by a “voter verified paper audit trail” (VVPAT) which would allow votes to be checked

by both the voter at the time of the vote and election officials after the voting was finished, thus producing

a trustworthy system.

The change in the public debate from 2001 to 2004 is striking. A search of New York Times

articles in 2001 and 2004 reveals 77 stories on “voting machines” in 2001 and 147 in 2004 (see Table 1).

In 2001, over one-third of these stories discussed accuracy and fairness alone, and only one-twentieth

mentioned security and trustworthiness alone. But by 2004, a majority discussed security and

trustworthiness alone, but only ten percent mentioned accuracy and fairness.5

<Table 1 here>

Another way to characterize the changes in the public debates about voting systems is to consider

mentions of the phrases “punch card voting” or “touch screen voting” in newspaper articles. Figure 1

does this for the New York Times. An average of fifteen punch card voting articles appeared each year

between 2000 and 2003, but the number fell to about one per year in 2004-2006. Moreover, the upsurge

in 2003 was entirely due to coverage of litigation on the 2003 California recall election. Stories on touch

screen voting followed just the opposite pattern. An average of seven touch screen voting articles

appeared each year between 2000 and 2002, but there were forty-four in 2004 and eighteen in 2006—an

average of twenty-three per year from 2004 to 2006. Moreover, the early articles about touch-screen

voting were largely laudatory—speaking to the promise of touch-screen voting, but later articles, as we

shall see, became highly critical.

<Figure 1 here>

It is not unusual for the definitions of a policy problem to change (consider the Iraq War or the

energy crisis), but such transformations suggest that something interesting is going on. Why the

redefinition? Is it the right one?

Consider first the redefinition. How did politics lead to a change in emphasis from accuracy and

fairness to security and trustworthiness, and, in turn, how did this change in emphasis affect public

3

policy? We argue that the presidential election events in Florida in 2000 brought to the fore questions

about accuracy and fairness, largely, but not exclusively, related to punch card voting. The events also

heightened partisan fears about voting systems, and led to a spate of law suits regarding punch card

systems. In subsequent years, a new definition of the “voting systems problem” developed as a result of

several converging trends. The effort to get rid of punch cards through court cases and legislation helped

clear the field of concerns about punch card systems, but it did so without leaving a legacy of case law or

legislation devoted to improving accuracy and securing equal protection by setting standards for reducing

residual vote levels. The availability of Help America Vote Act (HAVA) funds after 2002 made it

possible for localities to consider buying electronic machines which had become cheaper and somewhat

better as a result of the continuing advances in chip-making and electronics. But partisan fears about

stolen elections and growing fears over the role of computers in American life (perhaps further fueled by

the horrific events of 9/11) led to a concern with electronic voting. Finally, a social movement started on

the Internet in 2003 by skeptical computer scientists and others crystallized these concerns and offered a

seemingly simple solution to the problem of security—the use of voter verified paper trails with

electronic voting. Although the solution has proved harder to implement than many initially thought,

advances in technology and worries about computer security—not to mention the populist logic of the

voter verified paper trail idea—may be leading to national standards for paper trails.

Is this redefinition of the problem the right one? Is the emphasis on security and trustworthiness

instead of accuracy and fairness correct, and how far have we gone towards achieving either set of goals?

None of the strategies pursued to improve voting systems – including court cases, federal legislation, and

the verified voting social movement – has solved the system’s problems although some progress has been

made. Part of the problem is a failure to take seriously the need for tough performance standards, and

another part of the problem is the failure to have a balanced approach that considers accuracy and fairness

at least as much as security and trustworthiness.6

4

II. Accuracy and Security Risks

Risk, Threats, and Vulnerabilities – One reason for the paper-trail movement’s success is its

singular focus on the vulnerabilities of electronic systems and its contention that paper trails would

dramatically diminish these vulnerabilities. There is certainly some truth to this notion, but the focus on

machine vulnerabilities looks at only one-half of the risk equation—it ignores threats, and it looks at only

one-half of the vulnerabilities in the American voting system – it ignores vulnerable populations whose

votes are threatened by some voting systems. We summarize this argument in Table 2 and we explicate it

in the next few paragraphs.

<Table 2 here>

Reflections on the “riskiness” of a voting systems should start with the following equation:

Risk = Threats X Vulnerabilities

An unlocked house, for example, is vulnerable, but it may not be threatened if it is in a gated community.

Hence, the owner may face no security risks. Similarly, electronic voting machines may have the

equivalent of unlocked doors, but they may be protected by bureaucratic procedures and other checks. In

short, it is important to consider realistic scenarios as to how someone could actually rob a house or rig an

election,7 and it is sensible to think about how these threats might be controlled in other, perhaps simpler

and cheaper, ways than with a paper trail.8 Except for cryptographic and computer science approaches

that are controversial (Adida 2006; Smith 2005), there has been very little discussion of these

possibilities, perhaps because many computer scientists feel much more comfortable talking about

vulnerabilities which involve the computer software and hardware that they know so well than about

threats involving the actions of people and organizations about which they know very little.9 Social

scientists, on the other hand, may err by focusing on threats which they understand while ignoring

vulnerabilities which they do not. Each group ignores half the risk equation listed in the bottom panel of

Table 2.

But even more than that has been ignored in the past few years. There are different kinds of risk.

Security risks are one kind that leads to doubts about the trustworthiness of a system. Another kind is

5

accuracy risks which stem from human factors problems. When the equation described above is used to

study accuracy risk its components must be reinterpreted. Vulnerabilities in a voting system become the

vulnerable populations who might be affected by the threat of a poorly conceived voting system. And the

experts in these vulnerabilities are the social scientists and lawyers who know about the problems faced

by people when they interact with governmental systems. The experts on threats are human factors

engineers who understand how people can be buffaloed by machines.

What is at Stake – Both kinds of risks must be taken seriously. Security risks have been taken

seriously partly because most people believe both that computers might be vulnerable and that there are

those who want to rig elections. Yet upon reflection it seems just as self-evident that there must be voting

systems that don’t record votes properly and people who have trouble with them. Moreover, there is

strong empirical evidence that voting system accuracy varies with the kind of voting system and that the

likelihood of someone’s vote being counted depends upon what machine is used.

Figure 2 reveals significantly different levels of voting system accuracy as measured by residual

vote rates across two adjoining counties in Southern California.10 In 2004, Los Angeles County

responded to a court decree that outlawed the use of punch-cards by converting their Votomatic-style

punch-card system to an optical scan system called “Ink-A-Vote.” The County did this by simply

replacing the “stylus” which punches out the chad in the punchcard system with a pen that marks a punch-

card inserted inside the same Votomatic style mechanism. As a result, the Ink-A-Vote system has all the

problems of the widely discredited Votomatic system except for the hanging chads, although that problem

is replaced by concerns about whether the marker pen provides an adequate mark on the ballot. Ink-A-

Vote shares with the Votomatic system the difficulty that a voter cannot easily see whether or not a mark

is made in the proper place while the punch-card is inside the machine, and once the punch-card is

removed, the voter has a hard time telling what was done because there are only numbers on the card that

correspond to the various candidates and contests listed on the ballot.

<Figure 2 here>

6

Given the serious problems with Votomatic style punchcards that were noted by the Supreme

Court in Bush v. Gore,11 it seems likely that Ink-A-Vote might not perform very well. This turns out to

be the case. Darker shades in Figure 2 indicate higher residual vote rates in the 2004 presidential race.

The crucial comparison is along the border between Los Angeles and San Bernardino counties which runs

up the center of the picture. Moving from left-to-right in the picture from Los Angeles to San Bernardin

County, the coloring becomes, on average, much lighter just

o

at the border between the counties, and the

effect is especially apparent in the highly populated area near the bottom of the picture. (The 10,000 foot

San Gabriel and San Bernardino Mountains are in the middle part of the picture, and the upper part of the

picture is the relatively sparsely settled Mojave Desert.)

Residual vote rates change dramatically just at the border between Los Angeles and San

Bernardino counties. The most likely explanation for this is that different voting systems—different

ballot forms, methods of election administration, and voting machines—perform differently, and some of

them are more prone to lose votes than others. In this case, compelling evidence shows that in the 2004

election, of 1000 identical people living in the two counties who went to the polls, eight more of them in

Los Angeles County than in San Bernardino County had their votes lost because of the voting system.12

This difference makes a mockery of the American commitment to equal protection for votes – simply

living in San Bernardino County makes it more likely your vote will be counted than living in Los

Angeles County.

III. Accuracy and Fairness: Reports and Debates in 2001

Themes – A flood of task force and academic reports came out in 2001 about voting systems and

other aspects of elections. There were four strands of the discussion, two major ones and two minor ones.

One major strand was represented by reports authored by academics and Civil Rights groups which

tended to blame the election system for inaccurate and unfair voting systems that violated equal

protection. A second major strand was represented by reports by election administrators and other

government officials who were, not surprisingly, somewhat defensive about the system’s performance

(Florida Governor’s Select Task Force 2002; National Task Force 2002). One minor strand, which

7

initially received very little attention, was a concern with the security of DRE systems. A second minor

strand focused on accessibility, but we will treat it as part of the discussion of accuracy and fairness

because it dealt with equal access and equal protection. As we shall see, those concerned with

accessibility, like those concerned with the emphasis upon accuracy and fairness, eventually clashed with

those worried about the security of DRE machines.

The two major strands of thought grappled with a fundamental question: Who is at fault for

residual votes and confusion over ballot forms? On the one side there are those who say that residual

votes and confusion are simply the result of inattentive people who do not follow instructions carefully.

These defenders of the system argue that voters are responsible for most of the problems with the existing

system. Figure 3, “Voting for Dummies” is this group’s satirical take on who is at fault. Bill Daley, a

major figure in Al Gore’s Florida recount effort and the son of Richard Daley and the brother of Richie

Daley—two Democratic mayors of Chicago, is the putative author of this volume.

<Figure 3 here>

Many election officials fall into this camp that blames voters for most of the problems with the

system. Their perspective is mostly simple self-preservation. Although there is a professional ethic of

serving the voter, there is also a posture of avoiding blame for the inevitable problems of an under-funded

and nearly unmanageable system. Consequently, some election officials take refuge in stories about how

voters (especially those with low education) often do not vote in top-of-the-ticket races which produces

large percentages of residual votes.

On the other side are those who say that voting systems must be easy to use and equipped with

numerous check-points to avoid mistakes. These critics of the system argue that confusing systems, such

as the “Butterfly Ballot” in Palm Beach Florida, and hard to use and inaccessible voting machines are the

major problem with America’s electoral machinery. Figure 4, which shows a ballot with an impossible

tangle of lines to be followed by the voter who does not want to vote for George Bush, is this group’s

amusing take on what is at fault. Many good government and Civil Rights groups fall into this camp. For

8

these groups, concerns about historic injustices such as poll taxes and literacy tests lead them to want to

reduce the impediments to voting.

<Figure 4 here>

These two perspectives align with those of the political parties. For Republicans, the ethic of

personal responsibility, conjoined with the fact that many of the voters who are cowed or confused by the

system have low levels of education and are likely to be Democrats, leads them to blame many of the

problems of voting systems on individual mistakes. For them, Democrat Bill Daley’s efforts in Florida

amounted to a “second generation of voting manipulation” by making excuses for “dummies” in Florida.

For Democrats, the desire to help working class and minority groups through governmental action,

conjoined with the fact that African Americans, Hispanics, and working class people are

disproportionately Democrats, leads them to blame the system. For them, George W. Bush obtained an

unfair advantage through a voting system that made it harder for Democrats to vote than Republicans.

Reports – The academic and civil rights reports issued in 2001 focused on the problems of

accuracy and fairness, and they emphasized faults in the system, not in individual voters. These reports

were either written by or substantially informed by social scientists (especially political scientists) (Cal-

Tech/MIT Voting Technology Project 2001a; Cal-Tech/MIT 2001b; Brady et al 2001; United States

Commission on Civil Rights 2001; Wand et al 2001; Minority Staff Special Investigations Division

2001). They used residual vote analysis to investigate the performance of voting machines and statistical

methods to examine the impact of ballot design. The US Commission on Civil Rights report devoted an

entire chapter to “The Machinery of Elections,” and the first sentence was “Florida lacks uniform voting

systems for its 8.4 million voters.” Moreover, the report noted that “voters in poorer, predominantly

people of color communities were more likely to use voting systems with higher spoilage rates—meaning

those voters had a lower chance of having their votes counted accurately (United Commission on Civil

Rights 2001, chapter 8).” The report argued that the use of precinct based counting systems “dramatically

decreases spoilage rates,” and it noted that poor ballot design seemed to have disproportionately affected

the poor and minority communities in Palm Beach and Duval Counties.

9

The governmental reports typically took somewhat more moderate positions on the inaccuracy

and unfairness of voting systems by emphasizing that any system will suffer from some errors, but in

most cases they also mentioned the need to improve accuracy. In its report, the Florida Governor’s Select

Task Force on Election Procedures, Standards, and Technology (2002) noted that “Florida will always

have voters who will mark ballots incorrectly or voters who will choose not to mark their ballots, making

their intent difficult to determine.” Nevertheless, “In statewide or national elections, when different kinds

of voting systems with different error rates are used, every voter does NOT have the same chance to have

his or her vote counted accurately.”

The report of the National Council of State Legislatures (NCSL) (2001) enunciated two principles

in the section on “voting systems”: “Voting systems should be chosen to [1] allow voters to clearly and

easily express their vote; and [2] accurately record such votes.” The report went on to say that “The issue

of uniformity in voting systems is especially important after Bush vs. Gore. Some argue that Bush vs.

Gore mandates that states adopt uniform voting systems.”

Perhaps not surprisingly, the report that placed the least emphasis upon accuracy and fairness and

that was most inclined to blame residual votes on voters was a task force appointed by the Board of

Directors of The Election Center, a non-profit organization that caters to the needs of election and

appointed election officials. The report argues that the problems in 2000 were “created by people, not

machines....” and “Only a very small percentage of the problems were directly related to any failure by

vote tally devices themselves.” Moreover, “Only on rare occasion does an undervote mistakenly occur

due to a voter incompletely marking or punching his/her ballot” and “Overvotes are often due to voter

confusion about the voting instructions but occasionally are intentional statements by voters (National

Task Force 2002).”

None of these reports had much discussion of security issues. The NCSL report’s only mention

of security was with respect to Internet voting: “Many task forces and special committees have been

formed to study the potential of Internet voting, but most have found that today’s technology cannot yet

handle secure voting over the Internet.” The Florida report mentioned that one drawback of DRE

10

machines was that they “produce no paper ballots, making recounts difficult” but security per se was only

mentioned in a discussion of Internet voting “No current technology adequately ensures security and

privacy of the voter (Florida Governor’s Select Task Force 2002).”

The minor line of thought, hardly represented in any reports, and mentioned only in an occasional

newspaper story, was some skeptical computer scientists’ concern with touch-screen voting. A March 11,

2001 New York Times article “Ideas and Trends: Down for the Count: Why Some Numbers are Only

Very Good Guesses” by science writer Gina Kolata summarizes the issues and quotes the man who was

the Godfather of the effort to bring these concerns to the fore:

After the Florida election, cries went out to improve the accuracy of voting by getting rid of the old punch cards with their fickle chads. Less error-prone machines, like touch-screen versions, are expensive—Riverside County in California just spent nearly $14 million on 4,250 machines. But you can’t accidentally vote twice, and there should be no ambiguity about whom you voted for. Over all, accuracy tends to go up.

But as with any computer solution, the problem is that not only can the machines introduce errors but the errors can be much harder to detect, Peter Neumann, a computer scientist at SRI International in Menlo Park, Calif., warned.

What if there is a tiny glitch in the software or hardware? How would you know if 1 out of every 100,000 votes goes to someone other than the intended candidate? What if the misdirected votes were created by programmer sabotage?

‘There is no guarantee that you votes goes into the computer the way it looked on the touch screen,’ Dr. Neumann said. ‘One thing is sure: with touch-screen computers, there is no chance for a recount.’13

But none of the New York Times’ articles on the subject of computer security radiates the kind of

intensity that would be apparent in this debate just a few years later, and it is telling that this story was by

a science reporter, not a political reporter.

Carter-Ford Report – After the 2000 election debacle, the National Commission on Federal

Election Reform was created with former Presidents Jimmy Carter and Gerald Ford as co-chairs. The

Carter-Ford report came out in August 2001, and it became the touch-stone for voting reformers.

Informed by the work of over two dozen political scientists, but by very few computer experts, and only

two, David Jefferson and Peter Neumann, who would later figure as skeptics in the debates regarding

touch-screen voting,14 the report listed among its goals for Federal election reform “equipment that

reliably clarifies and registers the voter’s choice” but nothing about security. The first sentence of the

11

chapter on voting equipment noted that “In the 2000 presidential election more than two million voters

went to the polls but did not have any vote counted in the race for president.” The report went far towards

blaming the system for residual votes and ballot design problems by noting that “Every analyst of voting

equipment agrees that the number of residual votes and the rate of voter error is greatly affected by the

kind of equipment that is used.” Among its recommendations were benchmarks for voting system

performance that “should be expressed as a percentage of residual vote (the combination of overvotes,

spoiled votes, and undervotes) in the contest at the top of the ballot (Carter et al 2002, p.5, 50, 50, 9).”

Specifically, the report suggested the following benchmarks:

Good: Below 1% Adequate: Between 1% and 2% Worrying: Between 2% and 3% Unacceptable: Over 3%

Almost six pages of the report were devoted to explaining, justifying, and applying these standards. One

clause in one recommendation was devoted to security—“The scope of voting system standards should

include security (including a documentary audit for non-ballot systems)….”, but this was very little

compared to the page devoted to accessibility and the observation that “accessible voting technology will

tend to promote the future acquisition of DRE (touch-screen) electronic systems equipped with an audio

feedback device (Carter et al 2002, p.57, 58).”

The next few years would be marked by litigation in the courts over voting systems and

procedures, controversy over the passage of HAVA—the Help America Vote Act which attempted to

implement some of the recommendations in the Carter-Ford report, and a growing political movement

concerned with a somewhat surprising topic – the security of computerized voting systems.

IV. Litigation over Punch-Card Voting Systems – 2001-2003

Rationale for Litigation over Voting Systems – Concerns with accuracy and equal protection, the

ambiguous impetus of Bush v. Gore’s flirtation with equal protection, and the availability of data about

residual votes led to major court cases starting in 2001 in California, Illinois, Florida, and elsewhere. In

12

an “op-ed” piece entitled “Equal Protection for Votes” written just a few hours before Bush v. Gore was

announced on the evening of December 12, 2000, Henry Brady argued that:

But if the Supreme Court decides to base its decision in Bush v. Gore on the need for equal protection of votes, then something good may come of this. Because the differences across voting devices is a much bigger problem than differences across manual recount methods—especially once the manual recount has been extended to the entire state of Florida—the next decade may see a flurry of legislation to equalize voting systems so that the promise of one person, one vote is realized. 15

In both California and Illinois, two of the leading users of Votomatic style punch-card systems, court

cases led to decrees that set schedules for eliminating these systems. Many other cases have been brought

as well, but the case with the highest visibility involved the California Recall of 2003.16

The California Recall Case – On September 15, 2003, a three-judge panel of the Ninth Circuit

Court of Appeals postponed the California Gubernatorial Recall Election scheduled for October 7, 2003

based upon concerns with the high residual vote rates of punch card voting systems to be used in the

recall election. Eventually this decision was overturned by an eleven judge Circuit Court panel, but the

preliminary decision leaned upon the 2000 Supreme Court decision in Bush v. Gore to justify postponing

the recall until better voting systems could be installed in six California counties. The panel’s decision

responded to a lawsuit arguing that Votomatic style prescored punch cards (PPC) perform much worse

than any other voting system, thus violating the equal protection clause of the Fourteenth Amendment to

the U.S. Constitution because votes are less likely to be recorded and counted in counties using these

systems. The suit also argued that Votomatic style punch cards violated the Voting Rights Act of 1973

because “people of color in California are more likely to live in counties that continue to use PPC systems

and because, within those counties, PPC systems lead to high rates of undervotes and overvotes for

people of color and those with lower levels of education” (D-ACLU, 8/7/03, 3).17

The lawsuit brought by the plaintiffs named California Secretary of State Kevin Shelley as the

defendant, but the initiator of the recall effort, Ted Costa, asked to intervene to defend the recall. The

responses by the state and Costa’s attorneys were remarkably different. In the first sentence of its brief,

the state announced that “The Secretary of State concurs with plaintiffs that punch-card voting systems

13

are antiquated” (D-State, 8/15/03, 1). Instead of arguing the facts about PPC systems, the state argued

that postponing an election is an extraordinary remedy that should almost never be exercised.

The Costa brief took a much more aggressive approach focusing on PPC performance. The brief

argued that “there is no constitutional right to an error-free voting system or to a particular voting

system—or even to the best available voting system”(11). The Fourteenth Amendment, the brief argued,

requires only “substantial equality” which punch cards amply meet. The brief questioned whether the

statistical evidence actually demonstrates high residual vote rates for punch card systems. It argued that

even if there are higher residual vote rates for punch card systems, residual vote rates are not errors, they

are often intentional undervotes, “Many voters simply choose to abstain from voting in particular

contests, and others register vote protests by casting votes for competing candidates or positions” (17).

The brief also questioned whether findings of one to one and one-half percentage higher residual vote

rates for punch card systems are substantial enough to invoke equal protection or voting rights concerns

even if this higher figure is due to machine errors. Finally, the brief argued that even if punch cards do

perform poorly, new systems would do no better.

Extensive data have been presented elsewhere disputing these claims (Brady 2004). Figure 5

comes from that analysis, and its panels display the residual vote rates by percent minority in census tracts

in Fresno California in 1996, when the county used Votomatic style punchcards, and in 2000, when the

county used a precinct count optical scan system. The right-hand picture for 2000 shows a dramatic

decline in residual votes for all tracts and especially for the virtually 100% minority tracts at the right of

the picture.

<Figure 5 here>

The steep slope of the line in the first panel of Figure 5 shows that PPC systems disadvantage minorities

because they are even more affected by Votomatic-style punch card systems than non-minorities. In

social science parlance, there is both a direct effect of PPC systems on residual vote rates and an

interactive effect with the characteristics of the voters.18 Interactive effects, however, pose special

problems for those who want to explain them, especially in the highly charged area of race. They invite

14

questions about what it is about minorities that results in higher residual vote rates. Perhaps, some might

argue, minorities are at fault for these high residual vote rates.

The plaintiffs met the claim that minorities are at fault by arguing that the most important fact

about Votomatic-style punch card systems is not that they lead to more errors for some people than

others. The most important fact is that they lead to higher residual vote rates for everybody because they

are hard to use. In Figure 5, the intercept is much higher for punch cards in 1996 than for optical scan in

2000. Even in Marin County California, with one of the highest percentages of college graduates in the

country, residual vote rates went down after the county switched away from PPCs.

In the end, the eleven judge Circuit Court panel ruled that the equal protection claim was not

strong enough for an injunction although they noted that this was an issue over which “reasonable jurists

may differ”. They noted that the Voting Rights claim was stronger and that the plaintiffs were right in

saying that they only had to demonstrate a causal relationship between the voting practice and the

discriminatory result (and not an intention to discriminate as the Costa brief had claimed). But the panel

noted that “There is significant dispute in the record, however, as to the degree and significance of the

disparity” (9). Finally, although the panel thought the plaintiffs “are legitimately concerned that the use

of the punch-card system will deny the right to vote to some voters who must use that system”

nevertheless “At this time, it is merely a speculative possibility, however, that any such denial will

influence the result of the election” (11-12).

What Role Should the Courts Have? – The punch card cases led to settlements in California

and Illinois calling for the eventual discontinuation of punch cards,19 but they led to no case law that

extended the equal protection standard of Bush v. Gore. For a moment the three judge panel decision

postponing the California Recall did rely upon equal protection and Bush v. Gore, but this decision was

overturned by the eleven judge panel. In 2006, a three judge panel of the Sixth Circuit ruled in an Ohio

case (Stewart v. Blackwell) that punch cards and central count optical scan violated equal protection for

voters in general because of their higher residual vote rates, and that they violated Section 2 of the Voting

Rights Act because such systems disproportionately deny the votes of African Americans. But in July

15

the full court granted rehearing en banc and vacated the panel’s opinion. The case is now moot because

none of the systems being challenged were used in the November 2006 Ohio elections. Consequently, as

election expert Dan Tokaji puts it “The doctrinal legacy of Bush v. Gore remains up for grabs.”20

Court cases involving the accuracy and fairness of voting systems seem to founder on several

difficulties. One is that litigation such as the California Recall case brought right before an election must

speculate about effects and deal with the burden of justifying postponement. Cases brought after an

election must prove the effects and deal with the equally daunting burden of justifying a new election.

Cases that simply deal with equipment and that do not ask for postponements or new elections seem more

promising, but they must still deal with the complexities of residual vote measures and proving effects

using messy observational data. They must also cope with the intricacies of Section 2 of the Voting

Rights Act and the ambiguous equal protection legacy of Bush v. Gore. Nevertheless, there has been

some modest success in changing, and perhaps improving, election systems through court cases.

Certainly they speeded up the retirement of Votomatic style punch card systems.

V. The Passage of the Help American Vote Act: 2002

The Help America Vote Act of 2002 (HAVA) was the nation’s legislative response to the

problems in Florida and to the recommendations of the Carter-Ford Commission. After a difficult

struggle, in which the legislation seemed stalled at several points, partly because Democrats were fighting

for nationwide voting standards, whereas Republicans were more concerned about stronger measures to

combat fraud (Mark 2001; Pear 2002), the bill was signed by President Bush on October 29, 2002.

HAVA was a classic creation of American federalism and two-party compromise. It could and would

give out billions of dollars in payments for voting equipment (especially for replacing punch card or lever

voting) and other purposes (Title I), but it had no authority to issue rules or regulations. It established the

Election Assistance Commission (EAC) but “The Commission shall not have any authority to issue any

rule, promulgate any regulation, or take any other action which imposes any requirement on any State or

unit of local government except to the extent permitted under section 9(a) of the National Voter

Registration Act of 1993” (Section 209) It could only produce “voluntary voting system guidelines”

16

(Title II) based upon relatively weak “Voting Systems Standards” for voting for federal offices (Title III).

The standards in Title III certainly included some excellent principles such as voter notification

about overvotes and audit capacities for voting systems. But the voter notification requirement was

weakened considerably by allowing for “voter education” about overvotes instead of machine notification

(or machine prevention) of their occurrence, and the audit capacity, although it mentioned the need for a

“paper record,” did not stipulate when and how this paper record would be produced or reviewed. Only

the accessibility requirement was specific enough to lead to a clear-cut test of whether or not it was

implemented because it required “the use of at least one direct recording electronic voting system or other

voting system equipped for individuals with disabilities at each polling place.” The Carter-Ford

Commission’s notion of residual vote measures of performance and usability standards ended up in a

section allowing for studies on issues such as the “best methods for establishing voting system

performance benchmarks, expressed as a percentage of residual vote in the Federal contest at the top of

the ballot” (Section 241) or human factors research (Section 243).

Unfortunately, in both its formulation and its implementation, HAVA has run up against the

strong strain of American federalism that has traditionally left voting administration, and many other

functions to the states. In the case of voting, the counties (and in some states the townships) have been

the strongholds of voting administration. There are some reasons to believe that this has been a good

thing – it puts administrators close to the voters and it makes it harder to steal a national election by

manipulating 3000 counties, but there are also reasons to believe that it has hampered reform because the

counties have very different ways of doing things and many of them, partly because of their small size,

lack expertise. As a result, there is a lack of statewide, not to mention national standards for voting

systems. Furthermore, county (and in some states township) officials create a powerful constituency

arrayed against interference from the federal government in the form of clear-cut and forcefully

implemented performance standards. HAVA catered to this constituency by stating most standards in

vague language (Title III), and by not giving the Election Assistance Commission (EAC) any rule-making

or enforcement powers (Title II). Indeed, even the “voluntary voting system guidelines” described in

17

Title II must run the gauntlet of a Board of Advisors and an Election Assistance Standards Board

controlled by state and local governments. Despite EAC’s weakness, it was still too threatening for the

National Association of Secretaries of State who, at their 2005 winter conference, voted to encourage

Congress “not to reauthorize or fund the EAC after the conclusion of the 2006 federal general election,

and not to give rulemaking authority to the EAC (NASS 2005).”21 Perhaps the only bright aspect of the

standards process was the involvement of the highly respected and very competent National Institute of

Standards and Technology with the Technical Guidelines Development Committee.

The 2005 Carter-Baker Commission on Federal Election Reform probably summarized HAVA

best in its opening letter from its co-chairs, Jimmy Carter and James A. Baker: “Many Americans

thought that one report—the Carter-Ford Commission—and one law—the Help America Vote Act of

2002 (HAVA)—would be enough to fix the system. It isn’t (Commission on Federal Election Reform

2005).”

VI. Security and Trustworthiness: The Voter Verified Ballot Movement – 2003-2004

Security Concerns and the Voter Verified Ballot – On October 27, 2000, just eleven days before the

2000 presidential election, computer scientist Rebecca Mercuri defended her dissertation at the University

of Pennsylvania on “Electronic Vote Tabulation Checks & Balances”. Since 1992, Mercuri had been

writing about her doubts regarding the security of electronic voting (Mercuri 1992), and one of her

advisors, Peter Neumann, had written about his concerns at least as early as 1985 (Neumann 1985, 1990,

1993). In her dissertation, she concluded that electronic voting could not be trustworthy and secure

without a voter verified ballot (VVB) in which electronic machines print-out a copy of peoples’ votes

which they can inspect and which can be kept as an audit trail in case there is a need for a recount.22

Within three years, Mercuri’s doubts and her idea of a voter verified ballot would become commonplace,

but they were oddities in 2000.

Starting in September 2002, Bev Harris, a freelance writer and publicist, started to surf the web to

learn more about electronic voting. “On Jan. 23, 2003, she hit the mother lode. On an unprotected Web

site, she found 40,000 files of Diebold Election Systems' source code—the guts of software to run touch-

18

screen voting machines (Howland 2004).” She decided to post these files on the web and to pass them on

to computer scientists. She also went on to write a book, Black Box Voting, that presents electronic

voting in the worst possible light—with tales of back-room machinations and darkly hinted at

conspiracies.

At the same time in January 2003, David Dill, a computer scientist at Stanford University

released the “Resolution on Electronic Voting”:

Computerized voting equipment is inherently subject to programming error, equipment malfunction, and malicious tampering. It is therefore crucial that voting equipment provide a voter-verifiable audit trail, by which we mean a permanent record of each vote that can be checked for accuracy by the voter before the vote is submitted, and is difficult or impossible to alter after it has been checked. Many of the electronic voting machines being purchased do not satisfy this requirement. Voting machines should not be purchased or used unless they provide a voter-verifiable audit trail; when such machines are already in use, they should be replaced or modified to provide a voter-verifiable audit trail. Providing a voter-verifiable audit trail should be one of the essential requirements for certification of new voting systems." Rebecca Mercuri was an early signer, and the resolution eventually garnered the support of over 2,000

technologists and 8,000 other people. Thus began the year 2003 which would prove a turning point in the

discussion of electronic voting.

On July 23, 2003, Professor Aviel D. Rubin of the Information Security Institute at Johns

Hopkins University and three co-authors released a technical report entitled “Analysis of an Electronic

Voting System”23 which analyzed the Diebold code found by Bev Harris. They concluded that “this

voting system is far below even the most minimal security standards applicable in other contexts (Kohno

et al 2004, Abstract).” They argued that both “outsider” and “insider” attacks were possible. Voters

could cast unlimited votes by creating their own “smart cards,” and insiders could modify votes and even

identify individual voters.

The New York Times picked up the story the next day, and quoted Professor Rubin exclaiming

that “We found some stunning, stunning flaws”(Schwartz 2003). The paper led the State of Maryland,

which was procuring Diebold equipment, to commission two reports. The SAIC (2003) report concluded

that “The [Diebold] System, as implemented in policy, procedure, and technology, is at high risk of

compromise” although it also noted that “The State of Maryland procedural controls and general voting

19

environment reduce or eliminate many of the vulnerabilities identified in the Rubin report.” The RABA

(2004) report said that “We generally agree with the conclusions of the Hopkins Report on purely

technical issues” but it noted that the Hopkins report posited threats with respect to “unscrupulous voters

and malevolent insiders” without asking “whether e-voting systems introduce additional vulnerabilities.”

Nevertheless, the report ended with the following:

In discussions amongst the team members, there was no single consensus recommendation, except that the introduction of voter-verifiable paper receipts is absolutely necessary in some limited form. The number of software vulnerabilities such receipts mitigate, the amount of savings they introduce by lowering procedural requirements, and the trust they garner are likely to be just as cost-effective in the long run as a fully locked-down all-electronic system.

RABA (2004, p.23) did not think that voter-verified receipts were needed for every DRE, but they did

think that they should be provided for a small number of randomly chosen machines.

The Politics of the Campaign Against Electronic Voting – Web sites and Internet discussion

have been essential parts of the campaign against electronic voting. Bev Harris was the first to make a

splash with her posting of Diebold source code in January 2003 and with her use of the phrase “Black

Box Voting”. In June 2003, computer scientist David Dill started a website called VerifiedVoting.Org

which incorporated as a nonprofit organization in January 2004. The group grew quickly to have a

lobbyist, a nationwide coordinator, and other staff. A companion nonprofit foundation started in February

2004. These groups engage in lobbying, education, litigation, and organizing voters to monitor elections.

In 2004, VerifiedVoting.org worked to limit the implementation of paperless electronic systems before

the 2004 election, and it has continued to do so.

In the second-half of 2003, the verified voter movement and those with fears about computerized

voting found their cause celebre in the chief executive of Diebold, Walden W. O’Dell, who wrote a letter

in mid-August 2003 inviting 100 wealthy friends to a Republican Party fund-raiser. No doubt to

underscore his earnestness, he told them that “I am committed to helping Ohio deliver its electoral votes

to the president next year (Warner 2003).” For Internet browsers with suspicions about voting machine

manufacturers, Mr. O’Dell’s letter was the smoking gun – electronic voting system manufacturers were

trying to fix elections. O’Dell resigned from Diebold in December 2005.

20

A rough sense of the importance of the campaign over the Internet can be obtained by typing in

queries about its protagonists or antagonists on Google. Typing in the phrase [“Bev Harris” OR “Black

Box Voting”] leads to almost 750,000 hits.24 Typing in the phrase [“David Dill” OR “Verified Voting”]

yields about 500,000. Simply entering [“Wally O’Dell”] into Google retrieves over 100,000 hits and

entering [“Wally O’Dell” AND (“fraud” OR “stolen”)] retrieves about 14,300 hits. For comparison,

typing in “butterfly ballot” leads to about 123,000 hits and typing in “CalTech/MIT Voting” leads to

about 40,000. Typing in [“Carter-Ford Commission” OR “National Commission on Electoral Reform”]

leads to about 20,000 hits. The number can be doubled to 40,000 by dropping “National,” but that also

yields hits on the more recent Carter-Baker Commission. In short, the Internet movement against

electronic voting has a substantial public presence – probably an order of magnitude greater than some of

the touchstone events, studies, or reports that focused on accuracy and fairness in the aftermath of the

2000 election.

The net result of this movement has been to put enormous pressure on government officials and

voting system manufacturers. The nature of this impact can be seen in Figure 6 which plots mentions of

“touch screen voting” in the New York Times and mentions of people who became prominent critics of

electronic voting.25 The figure vividly shows that initially (from 2000 to 2002), touch screen voting was

often mentioned without reference to even one of these critics (the solid line is above the dashed line), but

by 2003 and 2004, the critics were almost always mentioned when touch screen voting was discussed.

The result was a sea-change in coverage on electronic voting.

<Figure 6 here>

In addition, the rise of the security issue led to both a split in professional elites and in political

groups. Whereas political scientists had dominated the debate about voting systems in 2000-2002,

computer scientists, especially those skeptical of DRE’s who advocated paper trails, began to dominate

the debate in 2003 and 2004. Moreover, although there were differences of opinion in both groups, the

average opinion of each was quite different. The VerifiedVoting.Org website has all the signers of Dill’s

“Resolution on Electronic Voting” classified into groups such as “Technologists,” “Political Scientists,”

21

and “Lawyers.” Whereas over 2000 technologists signed the resolution, only about 90 political scientists

did so, and none of them were involved with the Carter-Ford Commission except for Robert Pastor.26

Although many political scientists may not have realized that they were welcome to sign a resolution that

emanated from the computer science community, it is still surprising to see only one prominent name.

Furthermore, the senior author has repeatedly attended meetings at which the split between the two

communities was quite clear. Most political scientists and lawyers focus on accuracy and fairness, and

most computer scientists and other technologists focus on security and trustworthiness.27

This elite split is also reflected in a split between groups on the political left. Civil rights and

disability groups focus on accuracy and fairness while more populist movements such as Moveon.org

focus on security and trustworthiness. The split became a political issue for Common Cause and the

League of Women Voters which both initially supported paperless DRE’s, but which switched mid-2004

to opposition (Konrad 2004). As membership groups, Common Cause and the League of Women Voters

have a populist tendency, but as “good government groups” they have traditionally been concerned with

the political participation of the poor, working class, minorities, and the disabled. It is not surprising they

found themselves “in the middle” on this issue, but it is also telling that they eventually opted to support

voter verification.

VII. The Situation in 2006

In 2000, almost one-half of American voters voted with either punchcards (31%) or lever systems (17%).

In 2004, only one-quarter of Americans voted with either punch-cards (12%) or lever machines (13%).

In 2006 less than one-twelfth of Americans voted with either lever machines (7%) or punch cards (.24%).

These two older technologies have been practically eliminated. At the same time, the fraction of those

using DRE systems has increased from 12% in 2000 to almost 40% in 2006 while optical scan systems

have increased from 30% to almost 50% (Election Data Services 2006). The elimination of punch-cards

is probably an unmitigated good, although some opponents of DRE’s have argued that using punch-cards

is perfectly fine.28

22

The increase in DRE’s will be evaluated differently by different people. For those concerned

with security, it is a bad development as long as the machines lack a paper trail, but one of the effects of

the voter verified ballot movement has been to encourage systems with paper trails, although they are still

being used in only a few places and not much is known about their performance. In his observations of

an experiment with paper trails in Nevada, Ted Selker, an MIT engineer, found that the paper trails were

less confusing than he feared they would be but handling printers and paper added another level of

confusion to an already complicated enterprise (Selker 2004). A newspaper article on voting systems

with voter-verified paper audit trails used in Georgia in the 2006 election reports that a survey of 459

voters showed that about nine out of ten people thought the experience was good and the system easy to

use. All but one voter thought the paper trail accurately reflected their choices (Campos 2006). But in

truth, we still do not know much about what the practical difficulties will be with voter-verified paper

trails, and there are some indications that they add another layer of complexity to an already complex

system.

Nevertheless, the movement towards paper trails seems to be gaining steam. The 2005 Carter-

Baker Commission on Federal Election Reform recommended paper trails, and a November 2006 draft

report by the staff of the National Institute of Standards and Technology prepared for the Election

Assistance Commission’s Technical Guidelines Development Committee (TGDC) argues that election

officials should be able to recount ballots independently of a machine’s software which implies the use of

something like VVPAT (Burr et al 2006).

VIII. Studying the Performance of Voting Systems

Even if paper trails solve the security problem (and there are reasons to believe that they will be at best

only a partial solution) (Tokaji 2005), there is still the problem of accuracy and fairness. There are good

reasons to believe that some voting systems perform much worse than others. Figure 2 presented

evidence about the poor performance of Ink-A-Vote in Los Angeles County. Indeed, in the past three

statewide California elections using Ink-A-Vote, its residual vote rates were significantly above the

statewide average: In 2004, Los Angeles had a 2.0% residual vote rate compared to 1.35% statewide; in

23

2005, Los Angeles County had 3.5% compared to 1.6% statewide, and in 2006, it had a 3.1% residual

vote rate compared to 2.3% statewide.29

Of course, the differences between the residual vote rate for Los Angeles and the rest of state

might be affected by its high proportion of minorities and those with a low education. Figure 7

demonstrates that there is a strong negative relationship between residual vote rates for the 2004

presidential race for cities in Los Angeles and the percent high school graduates in those cities– more

education seems to reduce residual vote rates. There is also a positive relationship between percent

minority (defined as percent Hispanic and African American) and residual vote rates in those cities,

presumably the consequence of the low educational levels of minorities in Los Angeles. Yet, these

pictures also suggest that there is an interaction effect, as we found in Fresno County (see Figure 5)

between percent minority and residual vote rates, and this interaction effect suggests that Ink-A-Vote

performs badly.

<Figure 7 here>

Nevertheless, because we recognize the problems of comparing counties with different mixes of

people, we have used both geographic matching methods (as in Figure 2) where we compare precincts at

the borders of counties and we have used sophisticated statistical matching methods to determine the true

impact of voting systems on residual vote rates. The statistical matching methods take precincts from Los

Angeles and match them with precincts in nearby counties that have similar characteristics along many

different dimensions. These methods yield results that similar to our geographic matching methods.

Figure 8 suggests why the geographic methods do pretty well. The map at the top of shows the

education levels of various Census tracts at the Los Angeles-San Bernardino border with darker colors

indicating higher education. The map at the bottom shows residual vote rates for the 2004 election. On

the top map, variations in high school graduation rates have no obvious relationship to the border, but on

the bottom map, variations in residual vote rates stand-out at the border with a big drop-off between Los

Angeles and San Bernardino. Hence, the difference in residual vote rates cannot be traced to differences

in education levels. They seem to be due to Los Angeles Counties’ use of Ink-A-Vote.

24

<Figure 8 here>

These results show that convincing and scientifically defensible studies of voting system

performance can be done using residual vote rates, and the results show that there are significant

differences across systems. Indeed, one more picture demonstrates the need to become even more refined

in our analyses. Typically we lump different classes of voting systems together – all DREs, all optical

scan systems, and so forth. But Figure 9 shows that this may be very unwise. It presents a picture of the

border between Los Angeles County and Orange County which used a DRE system -- the Hart Inter-Civic

e-vote system. One might have expected the DRE system to do much better than Ink-A-Vote, but a

glance at the picture suggests that it did not. There is no obvious break at the border between the two

counties. Consequently, we must start to analyze each voting system model separately, and we must be

prepared for the possibility that some system models do well, and others do well.

<Figure 9 here>

IX. Conclusions

We began with two questions: (1) Why did politics bring about the change in emphasis from

accuracy and fairness to security and trustworthiness, and, in turn, how did this change in emphasis affect

public policy? (2) And is this emphasis the right one?

The presidential election events in Florida in 2000 brought to the fore questions about accuracy

and fairness, largely, but not exclusively, related to punch card voting and ballot design. The events also

heightened partisan fears about voting systems, and led to a spate of law suits regarding punch card

systems. In subsequent years, a new definition of the “voting systems problem” developed as a result of

several converging trends. The effort to get rid of punch cards through court cases and legislation helped

reduce concerns about punch card systems, but it did so without leaving a legacy of case law or

legislation devoted to improving accuracy and securing equal protection by setting standards for reducing

residual vote levels or even by requiring better reporting of residual votes as proposed in the next chapter

in this volume. Things might have turned out differently if some of the “near-miss” court cases had

established residual vote measures as important indicators of voting system performance.

25

The availability of Help America Vote Act funds after 2002 made it possible for localities to

consider buying electronic machines which had become cheaper and somewhat better as a result of the

continuing advances in chip-making and electronics. But partisan fears about stolen elections and

growing fears over the role of computers in American life (perhaps further fueled by the horrific events of

9/11) led to a concern with electronic voting. Finally, a social movement started on the Internet in 2003

by some skeptical computer scientists and others crystallized these concerns and provided a seemingly

simple solution to the problem of security—the use of voter verified paper trails with electronic voting.

The “verified voter” movement and its more extreme cousin “black box voting” brilliantly

touched a nerve in American politics that was exposed by the presidential election fiasco in Florida in

2000. Using what Richard Hofstadter called the “paranoid style” in American politics, some parts of the

movement, especially the “black box wing” have engaged in “heated exaggeration, suspiciousness, and

conspiratorial fantasy” to make its case (Hofstadter 1964). This style, made famous by Joe McCarthy,

attributes defeats, such as the ones the Democrats suffered from 2000 through 2004, to a dark, sinister

conspiracy. In the case of voting systems, some computer scientists have used their specialized

knowledge and prestige to refer to possibilities that only they can fully comprehend, but that, if known to

the general public, would be frightening in the extreme, and when pressed for examples about actual

incidents they have often responded with “How would you know?” The answer, of course, is that you

could generally know by parallel monitoring30 or by analyzing voting statistics,31 but these methods are

seldom mentioned.

At the same time the voting machine manufacturers exacerbated the problem by writing sloppy

code, designing vulnerable machines, and sometimes refusing to release source code – not to mention

writing political fund raising letters. There are, in fact, real problems with voting system security which

need to be addressed. As Hofstadter noted, “nothing really prevents a sound program or demand from

being advocated in the paranoid style.” Is the voter verified ballot a good idea? If it were costless and

easy to implement, it would be an excellent idea. Unfortunately, it is neither, but it is certainly not a

terrible idea, and it may be the best way to quiet fears unleashed regarding electronic voting.32 Advances

26

in technology and worries about computer security—not to mention the populist logic of the voter

verified paper trail idea and the very effective campaign waged by computer scientists—appear to be

leading to national standards for paper trails.

So have we made any progress over the past six years? None of the strategies that have been

pursued to improve voting systems – including court cases, federal legislation, and the verified voting

social movement – has really solved the system’s problems although some progress has been made.

Court cases are slow, incremental, and often poor instruments for changing administrative systems.

National legislation has been vexed by partisan differences over the causes and nature of voting system

failures and by the power of American federalism—especially the power of local election officials. The

“verified-voting” social movement has become enormously powerful, but it has mostly focused on one

thing – security issues for DRE voting systems – and it has both encountered substantial opposition from

election administrators and it has focused almost all of its and the media’s attention on security when

many other issues, such as poor registration systems, inaccurate voting systems, and poorly trained poll

workers, are equally pressing and important (see the next chapter).

The result has been a failure to consider all risks of voting systems – not just security risks but

accuracy risks as well, and to often focus on only one side of the risk equation, such as the vulnerabilities

of electronic systems without considering the nature of the threats. One approach that might get beyond

the current focus on one thing at a time would be to develop risk analyses that considered all risks of

voting systems and that utilized measures of performance. This paper has shown that it is possible to get

a meaningful measure of performance for accuracy by using residual vote analysis and sophisticated

statistical and mapping techniques. Moreover, the mapping techniques have something in common with

the voter verified ballot idea – they are very compelling. The problem becomes immediately clear to

anyone who will look. Perhaps that is the device that proponents of accuracy and fairness need to make

their case.

27

Table 1 – Subjects of Articles in New York Times on Voting Machines

2001 2004 Accuracy and Fairness 35% 10% Security and Trustworthiness 5% 50% Both 7% 9% Neither 53% 31% Totals 77 147 Source: Authors’ coding of articles from a Lexis-Nexis Search.

Table 2: Risks, Threats and Vulnerabilities

Type of Risk Threats (Experts on threats)

Vulnerabilities (Experts on vulnerabilities)

Who is Affected?

Accuracy Risks (Can mean inaccurate system)

Systems that don’t record votes properly (Human factors; engineers and psychologists)

People who have trouble with machines (Social scientists who study vulnerable population)

Disabled, elderly and those with less education

Security Risks (Can mean untrustworthy system)

People who rig elections (Social scientists & risk analysts who study corruption)

Corruptible voting systems (Computer scientists and management experts)

Partisans of either party

Disabled,Elderly, &Those With Less Education

People who have Trouble with machines(Social scientists who study vulnerable populations)

Systems that don’t record votes properly(Human Factors Engineers and Psychologists)

Accuracy RisksCan Mean

InaccurateSystem

Partisans Of Either Party

Corruptible voting systems

(Computer scientists And Management Experts)

People who rig elections

(Social Scientists& Risk Analysts

who study corruption)

Can Mean Untrustworthy

Security Risks

System

Who is Affected

Vulnerabilities(Experts on Vulnerabilities)

Threats(Experts on Threats)

Type of RiskTable 2: Risks, Threats, and Vulnerabilities

28

Figure 1: Mentions of Punchcard and Touchscreen Voting

in New York Times, 2000-2006

YEAR

2006200520042003200220012000

Num

ber o

f Arti

cles

in N

ew Y

ork

Tim

es

50

45

40

35

30

25

20

15

10

5

0

Mention of:

Punchcard

Touchscreen

Source: Author’s coding of articles from a Lexis-Nexis Search.

Figure 2 –Residual Vote Rates by Precinct in 2004 Presidential Race for Los Angeles and San Bernardino Counties – Darker Colors Indicate Higher Residual Vote Rates

Source: Author’s research.

29

Figure 3 – The Voter’s Fault Figure 4 – The System’s Fault

Figure 5—Fresno County Residual Vote Rates in Census Tracts by Percent Minority in 1996 (on the left--Votomatic Style Punchcards) and 2000 (on the right--Precinct count optical scan)

30

Figure 6: Mentions of Touchscreen Voting and Selected

Critics of DRE Voting in New York Times, 2000-2006

YEAR

2006200520042003200220012000

Num

ber o

f Arti

cles

in N

ew Y

ork

Tim

es

50

40

30

20

10

0

Mention of:

Touchscreen

DRE Critic

Source: Authors’ coding from a Lexis-Nexus search.

Figure 7

Pres. Residual Rate by Percent HS Graduation

for Cities in LA County in 2004

Per Cent High School Grad

110%100%

90%80%

70%60%

50%40%

30%20%

Res

idua

l Rat

e fo

r Pre

side

ncy,

Nov

. 200

4 5.0%

4.5%

4.0%

3.5%

3.0%

2.5%

2.0%

1.5%

1.0%

.5%0.0%

City Size

gt 1000K

200-1000K

100-200K

50-100K

30-50K

20-30K

10-20K

5-10K

lt 5000

Total Population

WhittierWestlake Village

West Hollywood

West Covina

WalnutTorranceTemple City

South Pasadena

South Gate

South El Monte

Signal Hill

Sierra Madre

Santa Monica

Santa Fe Springs

Santa Clarita

San Marino

San Gabriel

San Fernando

San Dimas

Rosemead

Rolling Hills Estate

Rolling Hills

Redondo Beach

Rancho Palos Verdes

Pomona

Pico Rivera

Pasadena

Parmount

Palos Verdes Estates

Palmdale

Norwalk

Monterey ParkMontebello

Morovia

Maywood

Manhattan BeachMalibu

Lynwood

Los Angeles

Long Beach

Lomita

Lawndale La Verne

La Puente

Lancaster

La Mirad

Lakewood

La Habra Heights

La Canada/Flintridge

Irwindale

InglewoodHuntington Park

Hidden Hills

Hermosa Beach

Hawthorne

Hawaiian Gardens

GlendoraGlendale

Gardena

El Segundo

El MonteDuarate

Downey

Diamond Bar

Culver City

Cudahy

Covina

ComptonCommerce

Claremont

CerritosCarson

Calabasas

BurbankBeverly Hills

Bell Gardens

Bell Flower

Bell

Baldwin Park Azusa

Avalon

Artesia

Arcadia

Alhambra

Agoura Hills

31

Figure 8—Comparing High School Education Levels with Residual Vote Rates

32

Figure 9 –

Residual Vote Rates in Los Angeles and Orange County

33

Bibliography

Adida, Ben. 2006. “Advances in Cryptographic Voting Systems.” P.h.D. diss., MIT. Barr, Earl, Matt Bishop and Mark Gondree. 2007. “Fixing Federal E-Voting Standards.”

Communications of the ACM 50(3):19-24. Brady, Henry. 2002. “Equal Protection for Votes.” In The Longest Night: Polemics and

Perspectives on Election 2000, edited by Arthur J. Jacobson and Michel Rosenfeld. University of California Press.

Brady, Henry and Iris Hui. 2006. “Is It Worth Going the Extra Mile to Improve Causal

Inference.” Paper presented to the Political Methodology Meeting. Davis, California. Brady, Henry, Justin Buchler, Matt Jarvis, John McNulty. 2001. Counting All the Votes. Survey

Research Center: University of California, Berkeley. Brady, Henry. 2004. “Postponing the California Recall to Protect Voting Rights.” PS: Political

Science and Politics 37(1): 27-32. Burr, William, John Kelsey, Rene Peralta, John Wack. 2006. “Requiring Software Independence

in VVSG 2006: STS Recommendations for the TGDC.” National Institute of Standards and Technology.

Caltech/MIT Voting Technology Project. 2001a. “Residual Votes Attributable to Technology:

An Assessment of the Reliability of Existing Voting Equipment.” Caltech/MIT Voting Technology Project. 2001b.”Voting: What Is, What Could be.” Campos, Carlos. 2006. “Paper Trail a Winner with Voters.” Atlanta Journal Constitution,

November 28, 2006, p.3B. Carter, Jimmy, Robert Michel, Lloyd Cutler, Gerald Ford, National Commission on

Federal Election Reform. 2002. To Assure Pride and Confidence in the Electoral Process: Report of the National Commission on Federal Election Reform. Brookings Institution Press.

Commission on Federal Election Reform. 2005. Building Confidence in U.S. Elections: Report of

the Commission on Federal Election Reform. Center for Democracy and Election Management, American University.

Election Data Services. 2006. “Almost 55 Million, or One-Third of the Nation’s Voters, Will

Face New Voting Equipment in 206 Election.” Press Release. October 2, 2006. Florida Governor’s Select Task Force on Election Procedures, Standards and Technology. 2002.

34

“The Governor’s Select Task Force on Election Procedures, Standards and Technology, Revitalizing Democracy in Florida.” Election Law Journal 1(1): 125-136.

Frisina, Laurin, Michael Herron, James Honaker and Jeffrey Lewis. 2007. “Ballot Formats,

Touchscreens, and Undervotes: A Study of the 2006 Midterm Elections in Florida.” Working Paper.

Gathright, Alan and John Wildermuth. 2003. “Counties Pushing Punch Cards into Extinction, All

Those Remaining Should be Gone by March, Officials Say.” San Francisco Chronicle September 16, 2003. p. A-17

Harris, Beverly. 2003. Black Box Voting: Ballot Tampering in the 21st Century. Elon House/Plan

Nine Press. Hofstadter, Richard. 1964. “The Paranoid Style in American Politics.” Harper’s Magazine

November, 1964, p.77-86. Howland Jr, George. 2004. “Black Box Backlash.” Seattle Weekly, March 10, 2004. Kohno, Tadayoshi, Adam Stubblefield, Aviel Rubin, Dan Wallach. 2004. “Analysis of an

Electronic Voting System.” IEEE Symposium on Security and Privacy Proceedings (2004): 27-40.

Konrad, Rachel. 2004. “League of Women Voters Drops Support for Paperless Voting

Machines.” Associated Press, June 15, 2004. Mark, David. 2001. “With Next Election Only a Year Away, Proponents of Ballot Overhaul

Focus Their Hopes on 2004.” CQ Weekly, October 1, 2001, p.2532. Mebane Jr., Walter and David Dill. 2007. “Factors Associated with the Excessive CD-13

Undervote in the 2006 General Election in Sarasota County, Florida.” Working Paper. Mercuri, Rebecca. 1992. “Voting Machine Risks.” Communications of the ACM 35(11):138.

--- 2002a. “Explanation of Voter-Verified Ballot Systems.” ACM Committee on Computers and Public Policy. The Risks Digest, July 24, 2002.

--2002b. “A Better Ballot Box?” 2002. IEEE Spectrum, October 2002. p.46-50.

Minority Staff Special Investigations Division. 2001. Income and Racial Disparities in the

Undercount in the 2000 Presidential Election. Committee on Governmental Reform, U.S. House of Representatives.

National Association of Secretaries of State (NASS). 2005. “NASS Position on Funding and

Authorization of the U.S. Election Assistance Commission.” Press Release February 6, 2005.

35

National Conference of State Legislatures Elections Reform Task Force. 2001. Voting in

America: Final Report of the NCSL Elections Reform Task Force. National Task Force. 2002. “National Task Force on Election Reform Election 2000: Review

and Recommendations by the Nation’s Elections Administrators.” Election Law Journal 1(1): 91-103.

Neumann, Peter. 1985. ACM Committee on Computers and Public Policy. The Risks Digest,

August 1, 1985.

-- 1990. “Risks in Computerized Elections.” Communications of the Association for Computing Machinery 33(11), p.170.

-- 1993. “Security Criteria for Electronic Voting.” Paper Presented at the 16th National Computer Security Conference. Baltimore, Maryland. (September 20-23, 1993)

Pear, Robert. 2002. “Bill to Overhaul System of Voting is Seen in Danger.” New York Times,

September 7, 2002, p.A1. RABA Innovative Solution Cell. 2004. “Trusted Agent Report: Diebold AccuVote-TS Voting

System.” Department of Legislative Services. Maryland General Assembly. R& G Associates, LLC. 2005. “Parallel Monitoring Program---California Special Statewide

Election Report of Findings.” California Secretary of State Office. Saltman, Roy G. 2006. The History and Politics of Voting Technology: In Quest of Integrity and

Public Confidence. Palgrave Macmillan. Schwartz, John. 2003. “Computer Voting is Open to Easy Fraud, Experts Say.” New York Times,

July 24, 2003, p.16. Science Applications International Corporation (SAIC). 2003. “Risk Assessment Report:

Diebold AccuVote-TS Voting System and Process.” Department of Budget and Management, Office of Information Technology, State of Maryland.

Selker, Ted. 2004. “Processes Can Improve Electronic Voting: A Case Study of an Election.”

Caltech/MIT Voting Technology Project. Shamos, Michael. 2004. “Paper v. Electronic Voting Records---An Assessment.” Working

Paper. Smith, Warren D. 2005. “Cryptography Meets Voting.” Working Paper. Technical Guidelines Development Committee. 2006. “Resolutions Adopted by the Technical

36

Guidelines Development Committee at the December 4-5, 2006 Plenary Session.” [http://vote.nist.gov/AdoptedResolutions12040506.pdf]

Tokaji, Daniel. 2005. “The Paperless Chase: Electronic Voting and Democratic Values.”

Fordham Law Review 73(4):69-84. United States Commission on Civil Rights. 2001. Voting Irregularities in Florida During the

2000 Presidential Election. Wand, Jonathan, Kenneth Shotts, Jasjeet Sekhon, Walter Mebane, Michael Herron and Henry

Brady. 2001. “The Butterfly Did It: The Aberrant Vote for Buchanan in Palm Beach County, Florida.” American Political Science Review 95(4): 793-810.

Warner, Melanie. 2003. “Machine Politics in the Digital Age.” New York Times. November 9,

2003, Section 3.

37

38

FOOTNOTES

1 The best history of where we have been is Saltman (2006).

2 In addition to problems with voting systems, registration systems failed to record people who had legally

registered, and they listed others who should have been eliminated from the rolls. Felons and those thought to be

felons (sometimes only because they shared a name similar to that of a felon) were purged from the rolls, sometimes

correctly and sometimes incorrectly. When people arrived at the polls, some found congestion and others

experienced what they considered harassment and even discrimination. Some found that they could not vote

because there was no provisional balloting procedure. Absentee votes were handled in different ways depending

upon the County, and some absentee votes were counted even though they lacked the proper date and signatures.

3 In Votomatic style punchcard systems, voters put their punchcard into a machine that makes it impossible to see

which chad is being punched out when they mark their ballot. Optical scan systems use paper ballots in which

voters mark candidates as on standardized tests—central count systems count these ballots at a central location

whereas precinct count systems count them at the precinct. Datavote punchcards make it possible for voters to see

the marks they are making as they choose a candidate.

4 The Supreme Court, however, explicitly disavowed the notion that their one-time decision should lead to a general

concern with equal protection for voting systems: “Our consideration is limited to the present circumstances, for the

problem of equal protection in election processes generally presents many complexities.” (Bush v. Gore, Per

Curiam, Section II.B.). Yet many court cases have been launched based upon the presumption that Bush v. Gore set

a precedent in which equal protection in voting systems should be considered. For an argument about why this

should be so see Brady (2002).

5 These results are based upon a search for articles on “Voting Machines” in Lexis-Nexis for both years. Articles

were then coded by the author as to their subject matter. An approach based upon searching for key words such as

“fraud,” “tampering,” “trustworthy,” or “security” in the text as indicators of “security and trustworthiness” and

“accuracy,” “mistakes,” or “fairness” for accuracy and fairness led to broadly similar results for each year from 2000

to 2005 for The New York Times and for the Washington Post newspapers.

6 Tokaji (2005) offers a somewhat similar take on these issues. He argues that four equality norms can be advanced

by electronic technology: racial equality, disability access, multi-language access, and inter-jurisdictional equality.

39

7 Shamos (2004) notes that “It is pointless to discuss the security of a computer system in the absence of a well-

articulated list of threats.”

8 Some of these methods include ensuring a “chain of custody” for machines that makes it difficult for hackers to

corrupt them, using “parallel monitoring” that involves taking a random sample of voting machines on election day

and entering a known script of votes into them while video-taping the process, and using statistical methods after an

election to make sure that the election results make sense. For a description of parallel monitoring see R&G

Associates, LLC, (2005).

9 Exceptions to this include the work of computer scientists Shamos (2004) and Barr et al (2007, p.19) which

criticizes current standards for containing “no system model or threat model.”

10 This picture is taken from Brady & Hui (2006). For a discussion of the strengths and limitations of residual vote

rates see Brady et al (2001) and the next chapter. Residual vote comparisons are meaningful here because they are

being compared across counties within the same state. Residual vote measures still miss people (as many as half a

percent) who make mistakes by marking adjacent races—as in the butterfly ballot in Palm Beach County in 2000.

11 Bush v. Gore, Per Curiam, Section IIA. “This case has shown that punch card balloting machines can produce an

unfortunate number of ballots which are not punched in a clean, complete way by the voter. After the current

counting, it is likely legislative bodies nationwide will examine ways to improve the mechanisms and machinery for

voting.”

12 The raw figures for residual vote rates are 1.35% statewide in California, 1.1% in San Bernardino County, and

2.0% in Los Angeles County. The net difference cited in the text of 0.8% is determined from sophisticated

statistical techniques described in more detail later in the paper.

13 The original article referred to Sequoia County but a subsequent correction changes this to Riverside County. All

in all in 2001, The New York Times mentioned these security issues only four times, and three of the articles cited

Peter Neumann.

14 Carter et al (2002). See Appendix A, “Contributors to the Commission’s Work.” There were some other

engineers or computer scientists listed as contributors such as David Chaum of SureVote, Ron Rivest of MIT, Roy

Saltman (formerly of the National Bureau of Standards), and Ted Selker of MIT, but none of them became

outspoken advocates for paper trails, although Rivest has been instrumental in pushing for the idea in his capacity as

a member of the Election Assistance Commission’s Technical Guidelines Development Committee. See Technical

40

Guidelines Development Committee (2006). At least one other person, Kim Alexander of the California Voter

Foundation, later became an outspoken critic of touch-screen voting and an advocate for voter verified paper trails.

15 The piece was not published as an op-ed, but it was subsequently published, see Brady (2002).

16 Brady was an expert witness in California and Illinois and in the Recall case discussed in the next section.

17 Three sets of documents were produced during the court cases heard by the Federal District Court, the three-judge

panel of the Ninth Circuit Court of Appeals, and the 11-judge panel of the Ninth Circuit. Documents submitted to

these courts are indicated by D, A3, and A11 respectively. There were briefs from the American Civil Liberties

Union (ACLU), the state of California (State), and the lawyers for Ted Costa (Costa). In addition there was a

decision (Court). We reference documents by using the abbreviation for the venue, the author, the date of the

document, and the page number. These documents can be found on FindLaw (http://www.findlaw.com/) by

searching on “Southwest Voter Education Project v. Shelley.” Brady’s declarations can be found at:

http://ucdata.berkeley.edu/new_web/recall.html.

18 In terms of the curves in Figure 5 – the curve on the left has a higher intercept than the curve on the right (a direct

effect of PPC) and a much steeper slope (due to the interaction between race and PPC).

19 The California case was Common Cause v. Jones and the Illinois case was Black v. McGuffage. There were at

least three other cases: NAACP v. Harris in Florida, Andrews v. Cox in Georgia, and Stewart v. Blackwell in Ohio.

20 Daniel Tokaji Blog, December 1st, 2006. http://moritzlaw.osu.edu/blogs/tokaji/

21 Source obtained from Charles Stewart.

22 On her web page, (http://www.notablesoftware.com/evote.html), Mercuri says that she first proposed the idea in

her paper: “Physical Verifiability of Computer Systems" presented at the 5th International Computer Virus and

Security Conference in March 1992, and she apparently refers to this paper in an October 2002 publication: “A

method of voting described by this author over a decade ago, referred to as the Mercuri Method, requires that the

voting system print a paper ballot containing the selections made on the computer. This ballot is then examined for

correctness by the voter through a glass or screen, and deposited mechanically into a ballot box, eliminating the

chance of accidental removal from the premises (Mercuri 2002b).” But the 1992 paper is not available on her

website. Perhaps the first use of the phrase “Voter-Verified Ballot” appeared in Mercuri (2002a).

41

23 Kohno et al 2004. According to VerifiedVoting.org’s website, David Dill was responsible for assembling this

group: “July 2003 Professor Dill assembles a team of computer security experts to study Diebold e-voting software

brought to light by Bev Harris.” http://www.verifiedvoting.org/article.php?id=5604.

24 All searches done in December 2006. Because of a break with one of her collaborators, the original website called

“BlackBoxVoting.Com” displays a banner saying: “Reminder: We are NOT affiliated with Bev Harris”. Harris’

website is now “BlackBoxVoting.Org”.

25 These people are Bev Harris and computer scientists David Dill, David Jefferson, Douglas Jones, Rebecca

Mercuri, Peter Neumann, Avi Rubin, and Barbara Simons. These people were chosen because of their prominence

in the DRE debate and because they were mentioned at least twice in New York Times’articles between January 1,

2000 and December 15, 2006. They were also all signers of Dill’s “Resolution on Electronic Voting,” and many of

them have close professional connections. An article counted as one mention if the person (or concept of touch

screen voting) was mentioned once or several times. The same article could count twice (or more) if two (or more)

critics were mentioned.

26 Pastor was an especially important political scientist because he was a senior advisor to the Carter-Ford

Commission and later organizer of the Carter-Baker Commission.

27 There are some important exceptions such as Ted Selker of MIT and Michael Shamos of Carnegie Mellon.

28 “’I’m deeply concerned about the rush to change voting systems,’ said Kim Alexander, president of the California

Voter Foundation, a nonpartisan group focusing on voting technology. ‘I’m frustrated that the ACLU is doing

everything to ban the (punch-card) systems. They’re not bad systems if they’re used properly (Gathright and

Wildermuth 2003).’”

29 Since Los Angeles County is roughly one-quarter of the total electorate, the statewide figures are inflated by their

inclusion of LA.

30 Parallel monitoring involves taking a random sample of voting machines on election day and entering a known

script of votes into them while video-taping the process. The machine results are then compared with known results

to check for mistakes or fraud. For a description of parallel monitoring see R&G Associates, LLC (2005).

31 See Wand et al (2002) for an example of how to do the analysis by using past voting statistics to predict current

patterns—departures from these patterns may be due to mistakes or fraud.

42

32 Since this is an area where one’s opinion often matters and the senior author has been involved in debates over

voting systems, it may be worth remarking that Brady supports having some kind of voter verification system of

which a paper trail is one possibility. Paper trails might be a useful interim measure to increase trust in electronic

voting systems, but it is not a magic fix for all the problems which ail American voting systems.