I confess to Hutchinson that if I were a politician, I would be scared to use BattlegroundAI. Generative AI tools are known to “hallucinate,” a polite way of saying that they sometimes make things up out of whole cloth. (They bullshit, to use academic parlance.) I ask how she’s ensuring that the political content BattlegroundAI generates is accurate.

“Nothing is automated,” she replies. Hutchinson notes that BattlegroundAI’s copy is a starting-off point, and that humans from campaigns are meant to review and approve it before it goes out. “You might not have a lot of time, or a huge team, but you’re definitely reviewing it.”

Of course, there’s a rising movement opposing how AI companies train their products on art, writing, and other creative work without asking for permission. I ask Hutchinson what she’d say to people who might oppose how tools like ChatGPT are trained. “Those are incredibly valid concerns,” she says. “We need to talk to Congress. We need to talk to our elected officials.”

I ask whether BattlegroundAI is looking at offering language models that train on only public domain or licensed data. “Always open to that,” she says. “We also need to give folks, especially those who are under time constraints, in resource-constrained environments, the best tools that are available to them, too. We want to have consistent results for users and high-quality information—so the more models that are available, I think the better for everybody.”

And how would Hutchinson respond to people in the progressive movement—who generally align themselves with the labor movement—objecting to automating ad copywriting? “Obviously valid concerns,” she says. “Fears that come with the advent of any new technology—we’re afraid of the computer, of the light bulb.”

Hutchinson lays out her stance: She doesn’t see this as a replacement for human labor so much as a way to reduce grunt work. “I worked in advertising for a very long time, and there’s so many elements of it that are repetitive, that are honestly draining of creativity,” she says. “AI takes away the boring elements.” She sees BattlegroundAI as a helpmeet for overstretched and underfunded teams.

Taylor Coots, a Kentucky-based political strategist who recently began using the service, describes it as “very sophisticated,” and says it helps identify groups of target voters and ways to tailor messaging to reach them in a way that would otherwise be difficult for small campaigns. In battleground races in gerrymandered districts, where progressive candidates are major underdogs, budgets are tight. “We don’t have millions of dollars,” he says. “Any opportunities we have for efficiencies, we’re looking for those.”

Will voters care if the writing in digital political ads they see is generated with the help of AI? “I’m not sure there is anything more unethical about having AI generate content than there is having unnamed staff or interns generate content,” says Peter Loge, an associate professor and program director at George Washington University who founded a project on ethics in political communication.

“If one could mandate that all political writing done with the help of AI be disclosed, then logically you would have to mandate that all political writing”—such as emails, ads, and op-eds—“not done by the candidate be disclosed,” he adds.

Still, Loge has concerns about what AI does to public trust on a macro level, and how it might impact the way people respond to political messaging going forward. “One risk of AI is less what the technology does, and more how people feel about what it does,” he says. “People have been faking images and making stuff up for as long as we’ve had politics. The recent attention on generative AI has increased peoples’ already incredibly high levels of cynicism and distrust. If everything can be fake, then maybe nothing is true.”

Hutchinson, meanwhile, is focused on her company’s shorter-term impact. “We really want to help people now,” she says. “We’re trying to move as fast as we can.”

Share.
Exit mobile version