Deepfakes vs. democracy: Navigating the age of technological misinformation, political deception – F.D. Flam

In the age of AI-generated deception, combating deepfakes demands more than technology. F.D. Flam suggests a shift to critical thinking methods like refocusing attention on policies, questioning sources, and reconsidering our own biases. AI excels at tricking rapid judgments, making it crucial to prioritise slow thinking for effective scrutiny. Obsessing over image authenticity may be futile; instead, focus on substantive issues. Trusting AI-generated perfection can be misleading. Encouraging discernment ensures thoughtful evaluation in the face of misinformation, safeguarding the integrity of democracy.

Sign up for your early morning brew of the BizNews Insider to keep you up to speed with the content that matters. The newsletter will land in your inbox at 5:30am weekdays. Register here.

By F.D. Flam

Realistic AI-generated images and voice recordings may be the newest threat to democracy, but they’re part of a longstanding family of deceptions. The way to fight so-called deepfakes isn’t to develop some rumor-busting form of AI or to train the public to spot fake images. A better tactic would be to encourage a few well-known critical thinking methods — refocusing our attention, reconsidering our sources, and questioning ourselves.

Some of those critical thinking tools fall under the category of “system 2” or slow thinking as described in the book Thinking, Fast and Slow. AI is good at fooling the fast thinking “system 1” — the mode that often jumps to conclusions.

We can start by refocusing attention on policies and performance rather than gossip and rumors. So what if former President Donald Trump stumbled over a word and then blamed AI manipulation? So what if President Joe Biden forgot a date? Neither incident tells you anything about either man’s policy record or priorities.

Obsessing over which images are real or fake may be a waste of the time and energy. Research suggests that we’re terrible at spotting fakes.

“We are very good at picking up on the wrong things,” said computational neuroscientist Tijl Grootswagers of the University of Western Sydney. People tend to look for flaws when trying to spot fakes, but it’s the real images that are most likely to have flaws.

People may unconsciously be more trusting of deepfake images because they’re more perfect than real ones, he said. Humans tend to like and trust faces that are less quirky, and more symmetrical, so AI-generated images can often look more attractive and trustworthy than the real thing.

Asking voters to simply do more research when confronted with social media images or claims isn’t enough. Social scientists recently made the alarming finding that people were more likely to believe made-up news stories after doing some “research” using Google.

That wasn’t evidence that research is bad for people, or for democracy for that matter. The problem was that many people do a mindless form of research. They look for confirmatory evidence, which, like everything else on the internet, is abundant — however crazy the claim.

Real research involves questioning whether there’s any reason to believe a particular source. Is it a reputable news site? An expert who has earned public trust? Real research also means examining the possibility that what you want to believe might be wrong. One of the most common reasons that rumors get repeated on X, but not in the mainstream media, is lack of credible evidence.

AI has made it cheaper and easier than ever to use social media to promote a fake news site by manufacturing realistic fake people to comment on articles, said Filippo Menczer, a computer scientist and director of the Observatory on Social Media at Indiana University.

For years, he’s been studying the proliferation of fake accounts known as bots, which can have influence through the psychological principle of social proof — making it appear that many people like or agree with a person or idea. Early bots were crude, but now, he told me, they can be created to look like they’re having long, detailed and very realistic discussions.  

But this is still just a new tactic in a very old battle. “You don’t really need advanced tools to create misinformation,” said psychologist Gordon Pennycook of Cornell University. People have pulled off deceptions by using Photoshop or repurposing real images — like passing off photos of Syria as Gaza.

Pennycook and I talked about the tension between too much and too little trust. While there’s a danger that too little trust might cause people to doubt things that are real, we agreed there’s more danger from people being too trusting.

What we should really aim for is discernment — so people ask the right kinds of questions. “When people are sharing things on social media, they don’t even think about whether it’s true,” he said. They’re thinking more about how sharing it would make them look.

Considering this tendency might have spared some embarrassment for actor Mark Ruffalo, who recently apologized for sharing what is reportedly a deepfake image used to imply that Donald Trump participated in Jeffrey Epstein’s sexual assaults on underage girls.

If AI makes it impossible to trust what we see on television or on social media, that’s not altogether a bad thing, since much of it was untrustworthy and manipulative long before recent leaps in AI. Decades ago, the advent of TV notoriously made physical attractiveness a much more important factor for all candidates. There are more important criteria on which to base a vote.

Contemplating policies, questioning sources, and second-guessing ourselves requires a slower, more effortful form of human intelligence. But considering what’s at stake, it’s worth it.

Read also:

© 2024 Bloomberg L.P.