Accessibility Analysis of 50 States’ Mail-In Voting Web Pages

Accessibility Analysis of 50 States’ Mail-In Voting Web Pages

Catharine McNally, Accessibility Lead
#Accessibility | Posted

Mail-in voting is the most accessible, barrier-free way to cast your vote. But, what good is it if the process of starting to request mail-in voting isn’t accessible? My team and I set out to analyze all 50 states’ mail-in voting request web pages to evaluate whether or not the 1 in 5 Americans with a Disabilities would face a barrier in this democratic process.

The “quick & easy” way we tested was using automated testing software, which tests other software against conditions and its expected outcome. Within a few hours, we found that the average accessibility score of all 50 states is 90%.

“Gracefully passable,” I’d call it. I found myself saying, “It’s good enough. There’s not a problem with mail-in voting accessibility.” But I was wrong.

Bar Chart showing ascending count of states with positive scores, averaging 90% pass rate

I quickly reminded myself that automated testing has its limitations. While excellent for monitoring and a quick analysis, tools like Lighthouse (free) can only flag up to 30% of possible accessibility errors on a website. What about the other 70%? How do they get evaluated?

It’s done by us. You and I. Humans. There’s no way around it.

Does a positive automated testing score imply that manual testing will be positive, too? My Phase2 colleagues and I set out to find out.

We spent a couple days performing keyboard and screen-reader manual testing of all 50 states, and tracking the manual testing results against the automated results for an average score. The scores went from a 90% average to 77% average. No longer “gracefully passable.”

Bar Chart showing descending count of states with majority negative scores, averaging 77% pass rate

We found that manual testing exposed a more frictious, frustrating experience interacting with the page using a keyboard or a screen-reader. A story that the automated testing couldn’t tell us.

So what went wrong?

At a high level, users who rely on a keyboard to navigate through the web (vs. a mouse, due to lack of fine motor control) wouldn’t have the following experiences:

  • clear guidance on their “place” on the page.
  • tab order that didn’t flow in a logical manner from left to right, top to bottom, or
  • trapped cursor inside a section of the page.

Also users who use a screen-reader to read aloud the contents of the screen due to low vision or blindness didn’t have appropriate context or understanding of the information. The screen-reader information comes through the foundational code of the site. In our testing, we found these experience issues:

  • Missing form labels (i.e., search),
  • Alt-Text of images were nonexistent,
  • Links didn’t make sense alone.

These are all issues that are basic and easy to fix. Accessibility isn’t rocket science. It’s about beginning to understand the needs and best practices to follow. We’ll share a few examples:

Keyboard Focus

People with a motor control disability such as Parkinson’s or hand tremors may prefer to use a keyboard alone to better control the page navigation. Designers and developers need to ensure that keyboard focus is easily visible especially against different color backgrounds.

The keyboard focus is an outline around each interactive element (links, navigation menu, buttons) that provides feedback on where our current navigation position is on the screen. You could say it’s like having a spotlight on different dancers or actors on stage. Without the spotlight, it’d be more challenging to track the main action.

This is a problem in nearly half (46%) of the states tested. Automated testing cannot detect keyboard focus issues or their potential color contrast issues. Want to test on your state’s website? Go ahead and load it up, and start using the TAB key. Can you see the outline around the different elements and against background colors as you go through?

Connecticut has a great example of this because the keyboard focus color changes against backgrounds that are different colors. Too often do I see one keyboard focus color used and oftentimes it gets “lost” against different background colors.

Screen recording animation shows tab order with changes in outline color based on different backgrounds

To learn more, visit Google’s WebDev:

https://web.dev/focusable-controls/

Tab Order

Here is an example of Tab Order gone awry. In the Utah example, you will see it *starts* off in the header with “Services” and “Agencies”, then the Logo, and it goes to the footer. It skips over the main body altogether.

Screen-Recording shows the main body not receiving keyboard focus after going from header to footer

When the tab order does not match the visual order of the screen, it leads to frustration for the user. We need to ensure that all interactive elements on a page are focusable and accessible to a keyboard user.

To learn more about tab order and ensuring elements receive keyboard focus, visit Google’s Web.dev.

Control focus with tabindex

Form Labels

A common issue we also found surrounding the presence of form labels — such as the

  • search form field.
  • the name and address fields, or
  • a checkbox.

This is an issue on nearly 20% of all states’ mail-in voting webpages. Form labels are really helpful for blind or low vision users who use a screen-reader to communicate back the *purpose* of a form field.

Secondly, users may use voice recognition or talk-to-text applications such as Dragon Naturally-speaking, iOS’s Voice Control, or Android’s Voice Access to direct their cursor to specific fields on the page. Clearly defined labels make this experience smoother.

To learn more, visit Google’s WebDev:

Form elements do not have associated labels

Do Not Use Color Alone to Convey Meaning

Using color alone — such as a font color change — may not be easily distinguishable to users who have low vision or color blindness. As a quick stat, 1 in 12 men have color blindness. The most common is red-green color blindness (deuteranopia).

A prime example of when colors may be used to convey meaning are inline links. 34% of all states evaluated were in violation of this. Not all states used the typical visual affordance of blue links, as in the case of this state pictured below that uses cyan-colored links, which is problematic for those with deuteranopia.

An example of how a link presented with color alone may be perceived to someone with color blindness (right)

A simple fix is to make the type bold. Or add an underline. As you can see in this example below, the current implementation of links uses color alone. An improvement (and to bring up to compliance) would be to bold the text so that it’s more distinguishable by its style as well. Another way that is common on these state websites is use of an underline.

Side by Side examples of use of color alone vs. use of color with additional styling

To learn more about Color Contrast and Use of Color, visit Google’s Web.dev:

Color and contrast accessibility

Context

Another issue we frequently came across was lack of context or information. Any context greatly helps the user make decisions and jump around the site quickly, whether that’s through helpful link context beyond vague “Click Here” links, or the image alt text that is descriptive beyond the file name as the alt text, or having properly structured landmarks to assist with screen-reader navigation.

To learn more about link text and purpose, visit Google’s Web.dev:

Links do not have descriptive text

and for Landmarks:

HTML5 landmark elements are used to improve navigation

Ensuring our Elections are as Accessible as Possible. In-Person or by Mail.

Automated testing is still an excellent way to help us get started, it helps developers track regressions and some “quick wins.” With time and practice, manual testing truly gives us the best picture of how we are crafting an experience for our users.

Accessibility can be overwhelming, I get it. But perfection isn’t the expectation. Progress is. Over time we all get better, more efficient, and it becomes second nature in our process.

In having this progressive mindset, we make iterative improvements. As a result, we make it easier for citizens to navigate state websites — or any site for that matter — to participate in the democratic process. After all every vote counts.

This is the first in a monthly series of analysis and examples of digital accessibility in our everyday experiences.

Thank you to my Phase2 Colleagues Rachel Broardt and Brittany Alexander for their contributions to this analysis by conducting the manual testing.

Lighthouse Results by State.

If you would like the manual testing notes, please reach out to me on LinkedIn or Twitter and I’ll be happy to share!

Alabama | 83%

Lighthouse Report Viewer

Alaska | 80%

Lighthouse Report Viewer

Arkansas | 55%

Lighthouse Report Viewer

California | 91%

Lighthouse Report Viewer

Colorado | 96%

Lighthouse Report Viewer

Connecticut | 96%

Lighthouse Report Viewer

Delaware | 82%

Lighthouse Report Viewer

District of Columbia | 96%

Lighthouse Report Viewer

Florida | 96%

Lighthouse Report Viewer

Georgia | 94%

Lighthouse Report Viewer

Hawaii | 93%

Lighthouse Report Viewer

Idaho | 99%

Lighthouse Report Viewer

Illinois | 79%

Lighthouse Report Viewer

Indiana | 92%

Lighthouse Report Viewer

Iowa | 97%

Lighthouse Report Viewer

Kansas | 97%

Lighthouse Report Viewer

Kentucky | 94%

Lighthouse Report Viewer

Louisiana | 93%

Lighthouse Report Viewer

Maine | 97%

Lighthouse Report Viewer

Maryland | 86%

Lighthouse Report Viewer

Massachusetts | 96%

Lighthouse Report Viewer

Michigan | 100%

Lighthouse Report Viewer

Minnesota | 78%

Lighthouse Report Viewer

Mississippi | 98%

Lighthouse Report Viewer

Missouri | 98%

Lighthouse Report Viewer

Montana | 95%

Lighthouse Report Viewer

Nebraska | 91%

Lighthouse Report Viewer

Nevada | 93%

Lighthouse Report Viewer

New Hampshire | 97%

Lighthouse Report Viewer

New Jersey | 89%

Lighthouse Report Viewer

New Mexico | 78%

Lighthouse Report Viewer

New York | 100%

Lighthouse Report Viewer

North Carolina | 90%

Lighthouse Report Viewer

North Dakota | 74%

Lighthouse Report Viewer

Ohio | 94%

Lighthouse Report Viewer

Oklahoma | 73%

Lighthouse Report Viewer

Oregon | 91%

Lighthouse Report Viewer

Pennsylvania | 90%

Lighthouse Report Viewer

Rhode Island | 68%

Lighthouse Report Viewer

South Carolina | 98%

Lighthouse Report Viewer

South Dakota | 84%

Lighthouse Report Viewer

Tennessee | 100%

Lighthouse Report Viewer

Texas | 97%

Lighthouse Report Viewer

Utah | 85%

Lighthouse Report Viewer

Vermont | 92%

Lighthouse Report Viewer

Virginia | 92%

Lighthouse Report Viewer

Washington | 81%

Lighthouse Report Viewer

West Virginia | 96%

Lighthouse Report Viewer

Wisconsin | 96%

Lighthouse Report Viewer

Wyoming | 80%

Lighthouse Report Viewer

 

Catharine McNally

Catharine McNally

Accessibility Lead