62°F
weather icon Clear

‘Scary’: Officials concerned for role AI can play in elections

Updated March 17, 2024 - 9:11 am

When researchers, elections officials and artificial intelligence experts asked chatbots questions about elections, they were shocked at the responses.

The AI Democracy Projects, a collaboration between the Institute for Advanced Study and Proof News, had gathered the experts to examine the accuracy and potential harmfulness and bias of AI chatbots as more people begin using them to power up their internet searches.

Its findings — that around half of the AI models’ answers were inaccurate — reinforced concerns from some officials that the growing use of artificial intelligence could lead to voter suppression.

Companies behind the chatbots say their platforms refer users to official sources for election information and that they are proceeding cautiously as AI develops. In the meantime, state and federal officials are looking for ways to regulate the platforms and their potential influence on elections.

‘It’s scary’

As part of the AI Democracy Projects’ study, Nevada Secretary of State Cisco Aguilar saw firsthand some of the messages AI models gave when asked questions about the state’s election processes.

Models failed to mention that Nevada allows for same-day voter registration, incorrectly saying the deadline to register was weeks before an election. One AI model implied that people recently released from prison would need to go through an extensive process before regaining their voting rights, even though Nevada automatically restores voting rights upon people’s release from prison.

“When I say it’s scary, I have a very limited view of the term ‘scary,’” Aguilar told the Review-Journal. “Scary is when one person is dissuaded from voting. That’s scary.”

The information the AI models gave wasn’t always wrong, but it wasn’t totally accurate, either, and it did not paint the whole picture, Aguilar said.

Lost in translation

When officials asked the chatbots to translate some election-related information from English to Spanish, the tone of the Spanish came out too happy and festive, Aguilar said. When translated from English to Hindi, the information came out sounding so serious that it could have scared somebody out of voting, he said.

“And Nevada being such a diverse community, the individuals and potential voters are probably going to use these systems to translate information and … these systems have to be very careful about that tone and word choice,” Aguilar said.

Making it up

Alondra Nelson, a researcher at the Institute for Advanced Study and former acting director of the White House Office of Science and Technology Policy, said the AI Democracy Projects found the AI models responded sometimes with half-truths, and other times with entirely made up claims, such as a fake website that looked like a credible source, Nelson said.

In one instance, a chatbot said that a ZIP code with a large African American population did not have a polling site, Nelson said. In another, when the researchers asked if someone could vote by text, the chatbot said yes — even though that isn’t allowed in any state, Nelson said.

The implications of that misinformation is voter suppression, Nelson said.

“No one is intending necessarily to do anything bad, but the sort of critical mass of bad information is deeply, deeply corrosive and eroding democratic processes, election processes, and the like,” Nelson said.

Using AI

There’s not a lot of data on how many people are gravitating toward AI chatbots as sources of information, but a Pew Research Center survey of 11,000 adults in 2022 found that 27 percent of Americans said they interact with AI at least several times a day; another 28 percent interact with it about once a day.

A November 2023 poll of around 1,000 adults in the U.S. found that only 14 percent are somewhat likely to use AI to get information about the upcoming presidential election. The majority were concerned about AI increasing the spread of election misinformation, according to the UChicago Harris/AP-NORC poll.

But even though some people aren’t using AI products, the foundational models of the chatbots are being built into other products, Nelson said, such as Microsoft Office Suite and Bing. If companies build upon a foundation that is rotten or shaky, then that’s a huge challenge, Nelson said.

‘Proceeding cautiously’

Artificial intelligence companies say they have made goals to curb the spread of misinformation and are training their systems to direct users to the best resources when they ask for elections-related questions.

Meta, which runs its MetaAI model, directs people to state-specific resources where users can find authoritative information, according to spokesperson Daniel Roberts.

Anthropic, an artificial intelligence start-up, is taking steps to prevent the misuse of it AI systems, such as prohibiting political campaigning, conducting model testing against potential election abuse, and directing users to authoritative voter information, according to a spokesperson.

“Given generative AI’s novelty, we’re proceeding cautiously by restricting certain political use cases under our Acceptable Use Policy,” the Anthropic spokesperson said in a statement. Anthropic’s Claude model “is not trained frequently enough to provide real-time information about specific elections and that large language models can sometimes ‘hallucinate’ incorrect information,” a spokesperson said.

OpenAI has similar policies and prohibits people from building applications for political campaigning and lobbying with ChatGPT. It also doesn’t allow users to create chatbots that pretend to be real people, according to its elections blog to which a spokesperson directed the Review-Journal for information.

Regulating AI

Before the New Hampshire 2024 presidential primary, a robocall used artificial intelligence to mimic President Joe Biden’s voice, urging voters to not participate in the primary election and save their vote for November. The incident sparked further concerns of the influence of AI in elections and heightened the sense of urgency to implement regulations.

The Federal Communications Commission adopted a ruling in February that makes voice cloning technology used in robocalls illegal, giving state attorneys general new tools to go after bad actors who use AI in robocalls for nefarious purposes.

Both Nevada’s state and federal officials are examining further ways to regulate artificial intelligence in specific election-related instances.

Nevada’s Democratic Rep. Steven Horsford, who chairs the Congressional Black Caucus, launched the caucus’ AI policy series last month that studies the benefits and advancements of AI, as well as the challenges and its impact on Black Americans.

Through those series, Horsford and the Congressional Black Caucus will identify legislation that should be advanced.

‘More enforcement, not less’

One bill that could help prevent misinforming voters includes the John Lewis Voting Rights Advancement Act, Horsford told the Review-Journal. The legislation aims to end voting discrimination, and that could include efforts to misinform and mislead voters with AI.

“We have these new technologies, these new platforms, where we’ve already seen efforts to misinform and to push out disinformation campaigns, particularly to voters of color and to older voters,” he said. “And that’s why we need even more enforcement, not less.”

Nelson, who oversaw the Blueprint for an AI Bill of Rights in the White House, said Congress could send a strong signal to the Federal Elections Commission and other regulatory bodies that the existing authorities and laws they enforce also apply to AI.

Aguilar is monitoring legislation in other states, and he plans to present a package to the next legislative session regarding regulation, he said.

As the 2024 election season continues, Nevada’s secretary of state’s office will work to dispel disinformation as it arises, Aguilar said. It will also work with the state’s election integrity task force to evaluate claims. The task force includes Metro, FBI, homeland security and others that have the tools and resources to evaluate content, he said.

“It’s not about partisan politics,” Aguilar said. “What it’s about is making sure that voters have the most accurate information to make an informed decision.”

Contact Jessica Hill at jehill@reviewjournal.com. Follow @jess_hillyeah on X.

Don't miss the big stories. Like us on Facebook.
THE LATEST