<<

review Improving systems’ user-friendliness, reliability, & security Michael D. Byrne

abstract About half of Americans have limited confidence that their vote will be properly counted. These fears have focused attention on voting system reliability, security, and usability. Over the last decade, substantial research on voting systems has demonstrated that many systems are less usable and secure than they should be. Producing truly reliable voting systems demands more than just following the federal guidelines enacted in 2005 (which, although well intentioned, have failed to substantially improve current systems) or simply updating voting systems to computers using monies allocated by the 2002 (HAVA). In fact, HAVA has inadvertently led to the purchase of systems that may have actually increased the vote error rate. Key reforms needed to deliver reliable voting systems include substantial testing for usability, especially regarding the accurate capture of voter intent and the reduction of voter error rates, and measures to ensure the integrity of elections, such as election officials’ ability to secure .

Byrne, M. D. (2017). Improving voting systems’ user-friendliness, reliability, & security. Behavioral Science & Policy, 3(1), 15–24. hroughout the 2016 presidential election best practices for both usability and computer season, dark claims were floated about the security that had been established in the Telection being rigged, and almost half of decades prior.4 all Americans have limited confidence that their vote will be properly counted, according to an However, there were positive consequences as October 2016 survey.1 These fears focus atten- well. The contested 2000 election spurred a tion on the voting procedures and systems used wave of new research on many aspects of voting, in the . Are they, in fact, fair, and do including voting system usability, election admin- they give all citizens a voice, as the Constitution istration practices and procedures, computer requires? And in the wake of the vote recount security, and statistical auditing methods. efforts by Green Party candidate Jill Stein and Core Findings the Clinton campaign, with both camps voicing For example, the Caltech/MIT Voting Tech- concerns of potential computer hacking,2 Amer- nology Project (http://vote.caltech.edu/), an What is the issue? icans may yet wonder: are their votes secure? interdisciplinary research effort focused primarily After the Help America on political science, has produced a substantial Vote Act passed in 2002, The voting process has been questioned before, amount of valuable research on voting, particu- many states transitioned to direct recording electronic particularly following the contested presiden- larly on election administration. For example, the voting machines. While tial election of 2000 and the infamous butterfly idea of a residual vote—the difference between these improved usability (see the sidebar The Butterfly Ballot From the total number of ballots received and the total for disabled voters, they did not improve error rates Palm Beach County, Florida, 2000). Two years number of votes cast in a particular race—came over traditional voting later, with strong bipartisan support, Congress from this research and has now become a stan- methods. Further usability passed legislation called the Help America Vote dard measure of the quality of voting systems. and security testing is Act (HAVA) of 2002 to address election adminis- needed to improve the integrity of U.S. elections. tration problems. The ACCURATE Center (http://www.accurate-­ voting.org/), a 6-year interdisciplinary research How can you act? HAVA allocated billions of dollars to local center funded by the National Science Founda- Selected interventions jurisdictions to replace outdated voting equip- tion, focused instead on both computer security include: 1) Joining electoral ment. But it turned out that many of the voting and voting system usability. The center is respon- machine research and machines those jurisdictions rushed to purchase, sible for the vast majority of the research on design efforts in Travis most often voting computers known as direct voting system usability published since the County (Texas) and LA County (California), and recording electronic machines (DREs), offered center’s inception in 2005. In addition, the across other jurisdictions, little to no improvement. In fact, HAVA likely center’s research has yielded ideas that will likely to pool resources made usability worse for some voters in terms of be incorporated into the security and cryptog- 2) Embedding behavioral preventing voter errors, because some of these raphy architectures of future voting systems. science research and insights in electoral replacement systems were measurably worse design processes than traditional paper ballots, the best alternative In this article, I focus on usability, but usability then available.3 is not by any means the only important consid- Who should take eration. A truly successful voting system must the lead? The fundamental problem with HAVA is that it address multiple factors, such as security, usability, Behavioral science researchers, security put the need for purchases ahead of the science. accessibility, certification, ease of administration, experts, election officials The law imposed substantial pressure on county cost, compliance with election laws, transpar- clerks to purchase new voting systems and ency, and auditability. Clearly, improving and granted them generous budgets to do so, yet it updating the country’s voting methods, practices, offered almost no scientific evidence to guide and administration is no simple task. them on which systems were most usable and most secure. Commercial vendors, hungry for an allotment of the billions about to be spent, Human Factors of rushed in with poorly designed systems. These Voting Systems early systems were primarily DREs. They were Human factors is an academic discipline not only scientifically unproven to enhance concerned with matching engineered systems to voting usability but also failed to follow industry human capabilities. A human factors researcher,

16 behavioral science & policy | volume 3 issue 1 2017 The Butterfly Ballot From Palm Usability and human error are human factors Beach County, Florida, 2000 problems. A 2004 report issued by the National Institutes for Standards and Technology (under This confusing ballot played a key role in the 2000 U.S. presidential race. the direction of the Federal Election Assis- Many voters inadvertently chose Buchanan/Foster when intending to vote for tance Commission) acknowledged that there Gore/Lieberman, and even more voters failed to cast a valid vote in the pres- idential race. A follow-up behavioral study using a paper ballot and Canadian was a dearth of research data on voting system participants (who were unaware of the Palm Beach ballot) showed that they usability, noting that virtually no significant made similar errors on 8% of the ballots and rated the butterfly ballot more human factors studies of voting systems existed.5 confusing than a single-column ballot.A An international standard for usability measure- ment provides a three-component definition for usability: effectiveness (the accuracy or complete- ness users achieve with specific goals in the target environment), efficiency (the resources expended by users in relation to the effect achieved), and satisfaction (a subjective rating of the system).6 Table 1 applies these universal usability definitions specifically to the voting context.

Laboratory Studies on Legacy Voting Systems In the mid-2000s, my colleagues and I did some of the first systematic studies designed to assess 7,8 A. Sinclair, R. C., Mark, M. M., Moore, S. E., Lavis, C. A., & Soldat, A. S. (2000, December 7). An electoral the usability of legacy voting technologies. butterfly effect. Nature, 408, 665–666. (For these studies and other research done by me and my colleagues, voting-age adults were recruited from the Houston, Texas, area through for example, might investigate how to design a a mix of newspaper and Internet advertising. website so its visitors can quickly and easily find These samples were generally close to balanced what they are looking for. Most human factors on gender. They represented age ranges from academic programs are housed in psychology roughly 20 to 80 years and contained people departments, although some are in industrial representative of a broad mix of other demo- engineering programs or schools of information. graphic variables, such as education and ethnicity. Human factors started as a field of study in the For details, consult the individual cited papers.) 1950s, primarily in aviation, where practitioners investigated how to reduce deaths by designing First, we assessed voter error in each of three cockpits and training pilots to match their capa- traditional voting systems: paper ballots, punch bilities with what the engineers who built the cards, and lever machines. Voter error occurs airplanes assumed they were. when a voter casts a vote for someone other than

Table 1. Usability of voting systems Component Definition Note Effectiveness Voter error rate Requires knowing the voter’s actual intent. Therefore, for privacy reasons, the voter error rate cannot be studied in real elections; it can only be studied in the laboratory, where the voter’s intent can be known. Efficiency Time needed to fill out and cast a ballot Satisfaction Subjective user satisfaction Generally measured by a subjective usability questionnaire such as the System Usability Scale (SUS).

a publication of the behavioral science & policy association 17 errors, regardless of the technology used. This “the voting technology used is most likely because it is easier to work directly from a list and not from memory—an important had no effect on how quickly distinction, because many voters do not bring lists into the . In fact, in some juris- voters cast their ballots” dictions, it is illegal to do so.

the person whom the voter intended. Voter error also occurs when a voter fails to cast a vote when Electronic (DRE) she or he intended to. These errors are impos- Voting Problems sible to assess in real elections because ballots As we were conducting our studies on the are secret. In the laboratory, however, we were usability of legacy voting system, other research able to assemble diverse groups of voting-age teams were investigating the new commer- adults and assign them to participate in a mock cial DREs that flooded the market after HAVA vote with each of the three voting systems. become law.

We determined voter intent using one of two A research team led by Paul Herrnson, a professor methods. We gave mock vote participants either of political science now at the University of (a) a list of candidates to vote for and measured Connecticut, conducted a large study comparing errors as deviations from that list or (b) a booklet the most popular commercial DREs available on with names and descriptions of fictional candi- the market.3 They measured voter error by giving dates and then asked them to cast their ballots mock vote participants a list of candidates and in all three systems but vote the same way on measuring how often their actual votes diverged. each ballot. When a voter cast a vote that did not They found that even the best commercial DREs match his or her other ballots, we counted that were no better than paper ballots and most were ballot as having an error. worse, some substantially so.

We determined voter error rate, the measure of HAVA mandated that voters with special needs a voting system’s effectiveness, by comparing be given an accessible way to vote. Commercial each voter’s intent with the actual vote that DREs are more accessible than paper ballots, person cast. We measured a voting system’s effi- punch cards, and lever machines, all of which ciency by tracking the time it took each person are essentially impossible to use by voters to cast a ballot. We also measured voter’s satis- with various disabilities, such as blindness or faction with a system using the System Usability substantial motor impairments. The accessi- Scale (SUS), a standard usability questionnaire.9 bility features (mostly audio-based presentation of the ballot) of these early DREs were quite The key findings were straightforward: The poor by modern standards,10 but they did allow voting technology used had no effect on how jurisdictions to comply with HAVA’s accessi- quickly voters cast their ballots, but it did affect bility mandates. After HAVA, some jurisdictions their error rate and user satisfaction. Error rates combined paper ballots and DREs, whereas with paper ballots averaged 1%–2%, which was others moved entirely to DREs. These changes markedly lower than the error rates produced carried other costs. County clerks essentially had by punch cards and lever machines (typically to become information technology managers, a around 3%–4%, but sometimes even higher than new role for them. Furthermore, in some cases that). Voters also indicated via the SUS question- these changes likely led to more voter errors if naire that they were somewhat more satisfied paper ballots were replaced with DREs. There- when voting with paper ballots than with punch fore, although DREs may make the physical act cards and lever machines. of casting a vote easier for people with certain disabilities, they are not necessarily better for the We also discovered that when voters have a list general voting population, at least compared in hand of whom to vote for, they make fewer with paper ballots.

18 behavioral science & policy | volume 3 issue 1 2017 In other words, voting systems became measur- ably less usable in jurisdictions that moved from “County clerks essentially had to paper voting to early DREs after HAVA became law in 2002. Jurisdictions that moved from become information technology punch cards or lever machines to DREs generally did not take as big a step backward, but overall, managers” the change was not always progress.

Problems With Early DREs group to vote from a list of preferred candidates, To improve commercial DREs, it is first neces- whereas the other half chose from a booklet sary to figure out what makes most of them so with the names and descriptions of fictional difficult to use. Some fail to conform to simple candidates. We instructed those in the booklet guidelines about text size and readability. Some group to choose candidates and vote the same require voters to follow novel and unusual way on multiple ballots. And to control for the procedures. Others have poor touchscreens, order of voting (DRE vs. other system), we had confusing instructions, or other complications. half the voters vote first with the DRE and half Almost every DRE on the market in 2008, when vote first with the traditional system. Herrnson’s team conducted their study, has not one but multiple usability problems. As it turned out, our in-house DRE was no more effective or efficient than the traditional systems. To understand why DREs are difficult to use, my It took at least as long to vote on the DRE as it did colleagues and I constructed a DRE for research on other technologies. (More educated voters purposes. Like all other DREs, this one, called vote slightly faster overall, regardless of tech- VoteBox, is a computer system. Unlike most nology—a result often seen in such studies, and other DREs, VoteBox is not a touchscreen; voters here we found the same.) interact with it while either sitting or standing by clicking on buttons using a standard computer mouse. (See one of the VoteBox screens in Figure 1.) re creens fr e e reserc VoteBox was intended to be a better DRE than rec recrn eecrnc n sse many early commercial DREs, which did not conform to federal usability guidelines issued in 2005.11 These guidelines, The Voluntary Voting System Guidelines, called for voting systems to meet basic criteria for usability and security. They also identified a minimum text size, a minimum contrast, and other features to help make the systems more usable.

The VoteBox DRE met these minimum stan- dards, but it went no further. How did the VoteBox DRE compare with traditional systems? To examine the causes of differences in voting behaviors, we conducted a laboratory study in which we randomly assigned mock vote partic- ipants to vote on a medium-length ballot (27 races) twice: once using the VoteBox DRE and once using a traditional system (either a bubble- style paper ballot, a lever machine, or a punch card).12 We also randomly assigned half of each

a publication of the behavioral science & policy association 19 The VoteBox DRE did not reduce the error rate usability questionnaire my colleagues and I used compared with paper ballots. In fact, when we in all of our studies—our voting experiment compared VoteBox or one of its variants with participants rated VoteBox as substantially more paper ballots in our subsequent studies, the satisfying to use than bubble-type paper ballots, two had a similar error rate of roughly 1.5%. The lever machines, and punch cards. The satis- results demonstrate that simply following basic faction scores are, in fact, unusually high for usability guidelines can help improve usability, engineered systems of any kind, and the results but that alone is not enough to do better than held for young and old voters, computer experts the best legacy technology, paper. and computer novices, rich or poor, and similarly wide ranges on other demographic variables. Advantages of DREs Paper ballots, although not very accessible, More critically, those who used VoteBox were produce a record that is readable by humans, more satisfied than were those who used tradi- less vulnerable to malicious electronic tional systems, regardless of which system tampering, and auditable later. Yet despite their allowed them to vote faster or make fewer errors drawbacks, DREs have some advantages over (see Figure 3). This has two implications. First, paper ballots. Even when voters make errors, just because voters like a system does not mean interpreting the submitted ballot in a DRE is it generates lower error rates or allows people to unambiguous, whereas interpreting a paper vote in less time. Second, election officials who ballot is not. Consider the 2008 Senate elec- move away from DREs may find that their voters tion in Minnesota. A razor-thin margin of victory dislike the change. caused statewide recounts, and the two major political parties spent months contesting ambig- uous paper ballots, such as the ones shown in re er ssfcn Figure 2. (An excellent resource full of exam- n sse ples from this election can be found at http://

minnesota.publicradio.org/features/2008/11/19_ DSE Traditional method challenged_ballots/round1/.) Although DREs 100 might not improve voter error rates, they also do 90 not lead to such complications. 80

The one way that VoteBox differed consistently 70 from legacy systems in our experiments was in 60 satisfaction: Voters repeatedly rated VoteBox as 50 more satisfying to use than traditional systems. 40 For example, using the SUS—the same standard 30 20 10 oter satisfaction (SUS score) ±1 SEM

re s fr e nnes ene rce V 0 Bubble Lever Punch card Traditional voting method

Voters report being significantly more satisfied with the voting process when using direct recording electronic (DRE) voting machines than when using the following traditional methods: paper bubble ballots, lever machines, or punch cards. Voter satisfaction was measured by a standard survey called the System Usability Scale, which runs on a scale from 0 to 100, and the values shown are the mean across the voters. The error bars show the standard error of the mean—a statistical measure of how widely the results varied from voter to voter. Data come from “Electronic Voting Machines Versus Traditional Methods: Improved Preference, Similar Performance,” by S. P. Everett, K. K. Greene, M. D. Byrne, D. S. Wallach, K. Derr, D. Sandler, and T. Torous, 2008, roceedins o the S onerence on uan Factors in outin Sstes: 008, , NY: Association for Computing Machinery.

20 behavioral science & policy | volume 3 issue 1 2017 Ensuring DRE Security & Accuracy high, this can alert election officials that some- Some voters have stated that they like voting on thing went wrong. DREs because after they have navigated through the ballot, they can review those choices on the However, some residual votes do not indicate last screen before submitting their vote. If the voter error. When a voter abstains on purpose— voting machine software was malfunctioning— perhaps because he or she doesn’t like any of the or, worse, maliciously altered—would voters candidates—this also counts toward the residual notice the altered votes on the review screen? vote rate. Say Mary is an avid voter but chooses We have done multiple studies showing that not to vote in races in which she doesn’t like any most of the time, voters do not. In fact, roughly of the candidates or on propositions that she feels two-thirds of voters failed to notice changes, uninformed about. Mary’s intentional even though the study used a permissive would be counted toward the residual vote rate, standard for what counted as noticing the despite not actually being errors. change.13 When voters were asked if they noticed anything amiss on the review screen, There are also errors that do not show up in they got credit for detection if they said that the residual vote rate—for example, if a voter something was wrong, even if they could not meant to choose one candidate and instead articulate what it was. What this suggests is that selected another. Unfortunately, my colleagues 1-2% security measures that depend on voters thor- and I have demonstrated that wrong-choice voting error rates oughly checking their ballots are unlikely to be errors are much more common than other error for paper ballots completely effective. types (for example, see Figure 4). This means that the residual vote rate is not necessarily a One of the earliest proposals for improving good indicator of a bad ballot. It also suggests the security of DREs was to have the machines that voting system designers cannot rely on the also print out a paper record that voters could 55% examine through glass.14 These records are of Americans have only some, or little confidence generally produced by inexpensive thermal Figure 4. Frequency of dierent that their vote will be printers—imagine a low-quality, light purple voting error types properly counted credit card receipt. If voters do not notice anom- alies on the high-resolution display immediately prior to casting their vote, it seems highly unlikely 1.6 that they would notice them under even worse visual conditions. This suggests that other secu- 1.4 3-4% rity measures are necessary. 1.2 voting error rates for punch cards and 1.0 lever machines We tested two other interface manipulations 0.8 in our experiments and found that there was 0.6 little difference in error detection rates based on where the votes were on the ballot or the 0.4 15 number of votes that were altered. Changing Mean error rate (%) ±1 SEM 0.2 the interface to highlight party affiliation and 0.0 Wrong Undervote Extra Overvote missing votes helped a bit, but even in the best choice vote case, this brought detection rates up to just 50%. Error type

Instead of relying on voters to detect their own erote indicates that the voter selected too many candidates. issions indicate that a voter who intended to errors, sometimes errors can be detected using vote instead made no choice. Etra ote means a voter who the residual vote rate.16 Residual votes occur intended to abstain accidentally selected a choice. The error bars refer to the standard error of the mean. Data are from when voters fail to cast a vote or when they “Now Do Voters Notice Review Screen Anomalies? A Look at Voting System Usability,” by B. A. Campbell and M. D. Byrne, invalidate their vote, for example, by selecting roceedins o the 00 SENTE Electronic otin two candidates in a contest where only one is Technolo orshoorsho on Trustorth Elections, https://www.usenix.org/legacy/events/evtwote09/tech/full_- allowed. When the residual vote rate is unusually papers/campbell.pdf.

a publication of the behavioral science & policy association 21 compromise one for the other. This can be a “it is unlikely that a one- difficult balance, but it is critical.

size-fits-all approach • Voting by mail is not an ideal solution. The vast majority of U.S. voters still vote in person will be effective” at their designated , but in some areas of the United States (predominantly on residual vote rate to indicate the true error rate the West Coast), voting by mail has become and instead need to conduct laboratory usability popular. However, this approach is not studies that can verify voter intent. High residual favored by most voting security researchers vote rates can indeed indicate problems, but low because it offers essentially zero resistance to residual vote rates do not necessarily mean that coercion and weak resistance to other forms ballots were cast accurately. of fraud. Voting by mail also usually relies on paper ballots, which can seriously limit accessibility. For these reasons, it is unclear Building Usable Voting Systems whether voting by mail will continue to grow Although there is still a great deal that is not in popularity, and few researchers have inves- known about voting system usability, the last tigated its usability. decade has produced some key lessons: (For additional voting system usability studies, • The most critical measure of a voting system’s see the online Supplemental Material.) usability is the system’s ability to accurately capture voter intent. The time it takes to After more than a decade doing research on cast a ballot is also important, but it is not voting systems in collaboration with election particularly sensitive to design. Acceptable officials, I have learned that elections are dramat- satisfaction with a voting system is relatively ically more complex and challenging to manage easy to achieve. than most people realize. It is no easy task to maintain security and accessibility while also • Almost all changes in the way people vote keeping things manageable for election officials, impact usability, from ballot layout to small who have to navigate a maze of idiosyncratic choices in wording on instructions. So voting laws and customs. although guidelines are a good start and can help prevent certain classes of usability prob- Designing usable voting systems requires more lems, they are insufficient to guarantee usable than just people with expertise in accessibility voting systems. Usability testing, both during and usability. It requires collaboration between the design process (usually multiple times) people with expertise in election administra- and after the design is finalized, is critical. tion; computer security; certification and legal compliance; auditing; and, of course, usability • DREs offer the best avenue to accessibility and accessibility. What’s more, designing an for those with a disability, but most DREs effective system involves many trade-offs. in use today produce untenably high error Because of differences in election laws, local rates. Yet with careful usability testing, they budgets, and demographics, it is unlikely can most likely be made more effective than that a one-size-fits-all approach will be effec- legacy systems (even paper). Usability testing tive. Instead, different jurisdictions will require at multiple stages of development is a key different systems, so designing a usable voting requirement, one that no current commer- system is a problem that will likely need to be cially available system has met. solved multiple times.

• Both security and usability must be consid- Building on the research produced by Caltech/ ered early in the design of the system, and MIT, ACCURATE, and other groups, two collab- it is important to take great care not to orative efforts to build better voting machines

22 behavioral science & policy | volume 3 issue 1 2017 are currently under way: the Los Angeles County author affiliation (California) Voting Systems Assessment Project (VSAP; http://www.lavote.net/vsap/) and the Byrne: Rice University. Corresponding author’s Travis County (Texas) STAR-Vote project. (STAR e-mail: [email protected]. stands for secure, transparent, auditable, and reliable.)17 These two jurisdictions have different supplemental material constraints in terms of , demo- graphics, and resources. Nevertheless, both • https://behavioralpolicy.org/publications/ have brought election and voting system experts • additional references together to share their expertise, and the systems they are building share some major design features. Both will use a DRE user interface similar to the Center for Civic Design’s Anywhere Ballot (http://civicdesign.org/projects/anywhere- ballot/) to support usability and accessibility, and both will produce a paper record to ensure the system is secure and auditable. Both projects are also committed to usability testing. Preliminary usability data from the VSAP project are avail- able at http://www.lavote.net/vsap/research, and usability testing for the STAR-Vote project is under way at Rice University. If these systems ultimately prove successful, other jurisdictions may use Travis County’s and Los Angeles’s collaborative processes as models, and those with circumstances similar to those of Travis County and Los Angeles may adopt the systems themselves, although this is still years away.

Today, many of the DREs purchased in the early 2000s with HAVA funds are only a few years away from the end of their life cycle, and election officials are watching the Los Angeles and Travis County voting system development collaborations with keen interest. Further, many election officials are beginning to understand how behavioral science can help improve voting systems for their constituents.

It is the job of security experts and election administrators to worry about keeping ballots safe; it is partly up to behavioral scientists to ensure that what is recorded on those ballots accurately matches voters’ intent. Without that, all the security machinery in the world does not guarantee the integrity of American elections. For citizens to trust in the elections, they have to be able to trust that voting systems are user- friendly. Behavioral science has a key role to play in ensuring that they are—thus securing the integrity of U.S. elections.

a publication of the behavioral science & policy association 23 references

1. Cooper, B., Cox, D., Lienesch, R., & 10. Runyan, N. H. (2007). Improving access Jones, R. P. (2016). The divide over to voting: A report on the technology America’s future: 1950 or 2050? Findings for accessible voting systems. from the 2016 American Values Survey Retrieved from Dēmos website: http:// [Research report]. Retrieved from PPRI www.demos.org/sites/default/files/ website: http://www.prri.org/research/ publications/improving_access%20(1)_0. poll-1950s-2050-divided-nations- pdf direction-post-election/ 11. United States Election Assistance 2. Davidson, A. (2016, November 28). The Commission. (2005). Voluntary voting recount road to nowhere. New Yorker. system guidelines. Retrieved from Retrieved from http://www.newyorker. http://archives.eac.gov/program-areas/ com voting-systems/docs/vvsgvolumei.pdf/ 3. Herrnson, P. S., Niemi, R. G., Hanmer, attachment_download/file M. J., Bederson, B. B., Conrad, F. 12. Everett, S. P., Greene, K. K., Byrne, M. C., & Traugott, M. W. (2008). Voting D., Wallach, D. S., Derr, K., Sandler, technology: The not-so-simple act D., & Torous, T. (2008). Electronic of casting a ballot. Washington, DC: voting machines versus traditional Brookings Institution Press. methods: Improved preference, 4. California Secretary of State. (2007). similar performance. In Proceedings Top-to-bottom review. Retrieved of the SIGCHI Conference on Human from http://www.sos.ca.gov/ Factors in Computing Systems: CHI elections/voting-systems/oversight/ 2008 (pp. 883–892). New York, NY: top-bottom-review Association for Computing Machinery. doi:10.1145/1357054.1357195 5. Laskowski, S. J., Autry, M., Cugini, J., Killam, W., & Yen, J. (2004). Improving 13. Everett, S. P. (2007). The usability of the usability and accessibility of voting electronic voting machines and how systems and products (NIST Special votes can be changed without detection Publication 500-256). Washington, DC: (Doctoral dissertation). Rice University, U.S. Government Printing Office. Houston, TX. 6. International Committee for Information 14. Mercuri, R. T. (2005). Physical verifiability Technology Standards. (1998). ISO 9241- of computer systems. Retrieved from 11:1998. Ergonomic requirements for http://www.notablesoftware.com/ office work with visual display terminals Papers/PhysVerify.pdf (VDT)s—Part 11: Guidance on usability. 15. Campbell, B. A., & Byrne, M. D. (2009). Geneva, Switzerland: Author. Now do voters notice review screen 7. Greene, K. K., Byrne, M. D., & Everett, anomalies? A look at voting system S. P. (2006). A comparison of usability. In Proceedings of the 2009 usability between voting methods. USENIX/ACCURATE Electronic Voting In Proceedings of the 2006 USENIX/ Technology Workshop/Workshop on ACCURATE Electronic Voting Trustworthy Elections. Retrieved from Technology Workshop. Retrieved from https://www.usenix.org/legacy/events/ https://www.usenix.org/legacy/events/ evtwote09/tech/full_papers/campbell. evt06/tech/full_papers/greene/greene. pdf pdf? 16. Caltech/MIT Voting Technology Project. CFID=907371946&CFTOKEN=91391177 (2001). Residual votes attributable to 8. Byrne, M. D., Greene, K. K., & Everett, technology: An assessment of the S. P. (2007). Usability of voting systems: reliability of existing voting equipment Baseline data for paper, punch cards, [Report, Version 2]. Retrieved from and lever machines. In Proceedings http://www.hss.caltech.edu/~voting/ of the SIGCHI Conference on Human CalTech_MIT_Report_Version2.pdf Factors in Computing Systems: CHI 17. Bell, S., Benaloh, J., Byrne, M. D., 2007 (pp. 171–180). New York, NY: DeBeauvior, D., Eakin, B., Fisher, G., . . Association for Computing Machinery. . Winn, M. (2013). STAR-Vote: A secure, doi:10.1145/1240624.1240653 transparent, auditable, and reliable 9. Brooke, J. (1996). SUS: A “quick and voting system. Journal of Election dirty” usability scale. In P. W. Jordan, B. Technology and Systems, 1, 18–37. Thomas, B. A. Weerdmeester, & A. L. McClelland (Eds.), Usability evaluation in industry (pp. 189–194). London, United Kingdom: Taylor & Francis.

24 behavioral science & policy | volume 3 issue 1 2017