How AI Bots Spread Misinformation Online and Undermine Democratic Politics

·  The fact that humans mimic and synchronize communication styles when interacting;

·  We’re more likely to remember frequently repeated words and;

·  Our brains are more likely to connect unrelated ideas when words or phrases frequently recur together over and over again in the same context.

When we unwittingly engage with bots that repeatedly use certain phrases and terms, this subtly reinforces their association and impact.

What I Found
I conducted statistical analysis on how bot-generated content influenced Canadian Twitter users during the SNC Lavalin scandal from March 14 to April 9, 2019. My study found strong correlations between bot-generated and human tweets, suggesting people engaged closely with AI-generated content and stored it in memory for easy retrieval and recall.

My analysis shows that bot-circulated tweets shared a high degree of similarity with human-generated tweets. The similarity in the emotional salience of bot-generated and human-generated tweets was significantly pronounced.

I first used Spearman’s phi coefficient, a statistical tool, to measure how strongly bot tweets related with human tweets. I then applied linear regression to understand this relationship in more detail and to see if changes in bot tweets affected changes in human tweets.

The results show there is a strong correlation between bot and human tweets, and that the content of bot tweets significantly influences linguistic aspects of human generated tweets. In simpler terms, human tweets replicated bot tweets to a high degree.

On March 14, 2019, for instance, bot tweets shared 75 per cent similarity with human tweets, which increased to 92 per cent similarity by March 28. The emotional words used in bot tweets were reproduced in human tweets just over 97 per cent of the time on March 14. Though the reproduction of emotional words in human tweets decreased over time, the similarity remained significant. This underscores how closely bot- and human-generated content can mirror each other.

Sometimes, bots simply mirror human activity because machine learning and generative AI technology are designed to replicate our behaviors. However, my analysis aims to understand not just how similar bot tweets and human tweets are, but also which one influences the other and in what way.

My examination of prevalent words and phrases like “obstruction of justice,” “Trudeau’s scandal” and “Liberal coverup” propagated by bots were replicated at high frequency in human-generated tweets, both in unique and retweeted tweets.

Social Mimicry
This study’s findings support the idea that we are inclined to mimic language structures of tweets posted by bots, which gains traction among users engaged in a resulting ecology of tweets posted both by peers and computer programs, a phenomenon observed in human-computer interactions.

Research underscores our innate tendency to mimic the communication patterns of those around us, particularly those who share our values and beliefs — a phenomenon also evident in our digital interactions. This inclination shapes how we express ourselves online, influencing our language, intonation and even how we make arguments.

Popular words and catchphrases that appear at high frequency within online contexts help our brains organize and remember large amounts of information more easily.

But how do these phrases achieve such resonance?

We know that repeatedly hearing specific words or phrases significantly enhances our ability to recall them more effortlessly, especially when they consistently appear within the same context.

In the landscape of our social brains, single words and phrases don’t carry much weight by themselves. What makes them powerful tools for spreading misinformation is their knack for evoking mental connections to familiar ideas, shaping how we understand current political developments.

Consider this: when several disparate concepts are cleverly woven into a memorable catchy phrase, they merge those underlying concepts in our minds, making them feel logical and familiar. For example: “Make America Great Again.”

This fusion is especially potent when these ideas strike a chord with our core values and beliefs, making catchphrases highly effective at molding public opinion in profound ways.

As we navigate this era of digital discourse, awareness of blind spots in our social psychology is our best defense. Understanding how cues or triggers affect us can reduce their influence over time. The more aware we are of bots and how they work, the more able we are to protect ourselves from misleading rhetoric, ensuring our democratic processes remain robust and inclusive.

Sophia Melanson Ricciardone is Postdoctoral Fellow, Department of Psychology, Neuroscience & Behaviour, McMaster University. This article is published courtesy of The Conversation.