The Disruption of a Russian AI-Powered Bot Farm: A New Wave of Propaganda
In a major collaborative effort, intelligence agencies across the globe have successfully identified and shut down a highly sophisticated Russian bot farm that employed artificial intelligence (AI) technology to spread disinformation and pro-Russian sentiments on social media platforms.
The scheme, carried out by a digital media department within RT (formerly known as Russia Today), a state-controlled media outlet in Russia, utilized a powerful software called “Meliorator.” This software, reportedly developed by RT’s deputy editor-in-chief in 2022 with the approval and funding of an officer at Russia’s Federal Security Service, enabled the creation of “authentic appearing social media personas en masse.” These personas were used to generate text messages, images, and mirror disinformation from other bot personas.
The FBI, in collaboration with intelligence officers from the Netherlands and cybersecurity authorities from Canada, issued a cybersecurity advisory, highlighting the specific tool called “Meliorator” responsible for creating these AI-powered bot accounts. This tool not only allowed the bots to create social media profiles that appeared genuine, but it also facilitated the mass dissemination of false narratives and manipulated content.
To carry out their operations, the bot farm relied on X, the platform formerly known as Twitter. The Russian actors involved in this scheme utilized two seized domains to create email addresses necessary for signing up to X and creating the bot accounts. X, although equipped with various safeguards, struggled to detect and block these bots due to their ability to copy-paste One-Time Passwords (OTPs) from their email accounts to log in.
The US Justice Department, taking swift action, has already identified and suspended all of the 968 accounts used by the Russian actors to spread false information on X. However, officials are still in the process of locating any additional accounts involved in the disinformation campaign. The use of US-based domain names by the operation was a violation of the International Emergency Economic Powers Act, while the payment process involved in creating these accounts violated federal money laundering laws in the United States.
One noteworthy aspect of this operation was the creation of bot profiles that impersonated Americans. These accounts used American-sounding names and set their locations on X to various places within the US, lending them an air of authenticity. However, upon close analysis, it became evident that the profile photos of these accounts, which were predominantly headshots against gray backgrounds, were most likely generated using AI technology.
The content disseminated by these bots aimed to promote pro-Russian sentiments and disinformation. For instance, an account named Ricardo Abbott, claiming to be from Minneapolis, posted a video of Russian President Vladimir Putin justifying Russia’s actions in Ukraine. Another account named Sue Williamson shared a video in which Putin expressed that the war in Ukraine is not about territorial conflict but rather reflects the “principles on which the New World Order will be based.” These posts were liked and reposted by other bots within the network, amplifying their reach and impact.
While this particular bot farm was limited to X, it is crucial to recognize that the actors behind it had intentions to expand their operations to other platforms. Analyzing the Meliorator software revealed that foreign actors, aiming to spread political disinformation, have been increasingly utilizing AI technology to manipulate social media platforms and influence political outcomes. OpenAI reported dismantling five covert influence operations originating from Russia, China, Israel, and Iran back in May, revealing the widespread adoption of AI in these disinformation campaigns.
Christopher Wray, the Director of the FBI, emphasized the significance of this achievement, stating, “Russia intended to use this bot farm to disseminate AI-generated foreign disinformation, scaling their work with the assistance of AI to undermine our partners in Ukraine and influence geopolitical narratives favorable to the Russian government.” He further expressed the FBI’s commitment to collaborating with international partners and utilizing advanced technology to strategically disrupt similar nefarious activities in the future.
Responding to the allegations, RT downplayed the severity of the situation, stating that farming (referring to the creation of such bot farms) is a beloved pastime for millions of Russians. However, the implications of this bot farm’s activities extend far beyond mere pastime hobbies. The successful disruption of this Russian AI-powered bot farm underscores the urgent need for increased vigilance and countermeasures to identify, counter, and dismantle such operations that aim to manipulate public opinions and destabilize democratic systems.
Source link