Disinformation dismantles public trust and disorients electoral ethics

Nick Hajli
Loughborough, UK, April 11, 2024

Social media platforms have become more than mere tools for communication. They’ve evolved into bustling arenas where truth and falsehood collide. Among these platforms, X stands out as a prominent battleground. It’s a place where disinformation campaigns thrive, perpetuated by armies of AI-powered bots programmed to sway public opinion and manipulate narratives.

AI-powered bots are automated accounts that are designed to mimic human behaviour. Bots on social media, chat platforms and conversational AI are integral to modern life. They are needed to make AI applications run effectively, for example.

But some bots are crafted with malicious intent. Shockingly, bots constitute a significant portion of X’s user base. In 2017 it was estimated that there were approximately 23 million social bots accounting for 8.5% of total users. More than two-thirds of tweets originated from these automated accounts, amplifying the reach of disinformation and muddying the waters of public discourse.

How bots work

Social influence is now a commodity that can be acquired by purchasing bots. Companies sell fake followers to artificially boost the popularity of accounts. These followers are available at remarkably low prices, with many celebrities among the purchasers.

In the course of our research, for example, colleagues and I detected a bot that had posted 100 tweets offering followers for sale.

Using AI methodologies and a theoretical approach called actor-network theory, my colleagues and I dissected how malicious social bots manipulate social media, influencing what people think and how they act with alarming efficacy. We can tell if fake news was generated by a human or a bot with an accuracy rate of 79.7%. It is crucial to comprehend how both humans and AI disseminate disinformation in order to grasp the ways in which humans leverage AI for spreading misinformation.

To take one example, we examined the activity of an account named “True Trumpers” on Twitter.

The account was established in August 2017, has no followers and no profile picture, but had, at the time of the research, posted 4423 tweets. These included a series of entirely fabricated stories. It’s worth noting that this bot originated from an eastern European country.

Research such as this influenced X to restrict the activities of social bots. In response to the threat of social media manipulation, X has implemented temporary reading limits to curb data scraping and manipulation. Verified accounts have been limited to reading 6,000 posts a day, while unverified accounts can read 600 a day. This is a new update, so we don’t yet know if it has been effective.

Can we protect ourselves?

However, the onus ultimately falls on users to exercise caution and discern truth from falsehood, particularly during election periods. By critically evaluating information and checking sources, users can play a part in protecting the integrity of democratic processes from the onslaught of bots and disinformation campaigns on X. Every user is, in fact, a frontline defender of truth and democracy. Vigilance, critical thinking, and a healthy dose of scepticism are essential armour.

With social media, it’s important for users to understand the strategies employed by malicious accounts.

Malicious actors often use networks of bots to amplify false narratives, manipulate trends and swiftly disseminate misinformation. Users should exercise caution when encountering accounts exhibiting suspicious behaviour, such as excessive posting or repetitive messaging.

Disinformation is also frequently propagated through dedicated fake news websites. These are designed to imitate credible news sources. Users are advised to verify the authenticity of news sources by cross-referencing information with reputable sources and consulting fact-checking organisations.

Self awareness is another form of protection, especially from social engineering tactics. Psychological manipulation is often deployed to deceive users into believing falsehoods or engaging in certain actions. Users should maintain vigilance and critically assess the content they encounter, particularly during periods of heightened sensitivity such as elections.

By staying informed, engaging in civil discourse and advocating for transparency and accountability, we can collectively shape a digital ecosystem that fosters trust, transparency and informed decision-making.

Nick Hajli is AI Strategist and Professor of Digital Strategy, Loughborough University,  Leicestershire, UK. The above article published in The Conversation, has been reproduced here under Creative Commons Licence.

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.

Share this story

Related Stories

Indian Newslink

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide

Advertisement

Previous slide
Next slide