China & North Korea: AI-Driven Election Interference Targets US, South Korea & India in 2024
China and North Korea are preparing to influence upcoming elections in the United States, South Korea, and India through AI-generated content, following previous attempts in Taiwan. According to a Microsoft report, Chinese state-backed cyber groups aim to manipulate these elections on a global scale.
The report, released by Microsoft on Friday, suggests that China intends to manipulate the information environment in these democracies, particularly during the election period. Beijing is expected to create and distribute AI-generated memes, videos, and other content to sway voters and promote its political stance.
AI-Generated Content: Tools of Political Disruption
Microsoft noted that while the current impact of AI-generated content is minor, it could increase as China refines its tactics. During the Taiwanese election, Chinese actors attempted to disrupt the political process using AI-generated content, but their reach was limited. However, as these tools evolve, their effectiveness may improve significantly.
The Beijing-backed group Storm 1376, also known as Spamouflage or Dragonbridge, was active during Taiwan's elections, using AI-generated audio to influence voters. The group posted fake recordings, such as one involving the candidate Terry Gou endorsing another contender despite dropping out of the race. YouTube promptly removed this content, minimizing its reach.
In addition to fake audio, AI-generated memes aimed at discrediting William Lai, the successful pro-sovereignty candidate opposed by Beijing, circulated on various social media platforms. The group falsely accused Lai of embezzling state funds and spread defamatory content about his private life, including allegations of fathering illegitimate children. Microsoft noted that the content often featured AI-generated avatars resembling news anchors, generated using ByteDance's CapCut tool, similar to tactics employed by Iran, where AI-generated avatars have also been used to disseminate state propaganda and spread disinformation.
Upcoming Elections in Focus: US, South Korea, and India
China's influence activities are not limited to Taiwan. The report indicates that similar campaigns are expected to target the upcoming elections in India, South Korea, and the United States in 2024. Chinese cyber and influence groups, alongside North Korean operatives, are predicted to create and disseminate AI-generated content aimed at manipulating political perceptions.
In the United States, Beijing-backed groups are employing social media accounts to pose divisive questions, ostensibly to gather intelligence and better understand voter dynamics ahead of the 2024 US Presidential election. One example cited involved a post highlighting a $118 billion bipartisan bill that included funds for the US-Mexico border and Ukraine-Israel aid. The intention behind such posts appears to be stirring debate and gaining insights into public sentiment.
Microsoft's Warnings on Emerging AI Disinformation Tactics
Microsoft warns that China's attempts at AI-enabled disinformation are likely just the beginning of a larger-scale campaign that could influence global electoral processes. While the current influence of AI-generated content remains limited, its potential could increase as China and other actors further develop these tools and tactics.
The company also pointed out that these activities coincide with recent concerns about China's cyber capabilities. A recent official review board appointed by the White House blamed Microsoft's security shortcomings for allowing state-backed Chinese cyber operators to infiltrate email accounts of senior US officials. Furthermore, both the United States and the United Kingdom recently accused Chinese-backed hackers of running a years-long campaign targeting politicians, journalists, and businesses.
Growing Concerns Over Digital Influence and Cybersecurity
This latest development underlines the growing concerns surrounding the use of artificial intelligence for geopolitical influence. The report from Microsoft serves as a warning to nations worldwide about the potential misuse of AI by state actors to manipulate democratic processes and undermine free elections. The expansion of these campaigns into high-profile elections raises questions about the ability of governments and tech firms to detect and mitigate the impact of AI-driven disinformation effectively.
As the political landscape in several countries braces for elections in 2024, it is becoming increasingly crucial for social media companies, government agencies, and private firms to work in coordination to tackle disinformation campaigns that rely on AI-generated content. The evolving nature of influence tactics, especially involving sophisticated technologies, signifies a new era in the battle for narrative control and the protection of democratic integrity. Governments and technology companies must prioritize coordinated efforts to safeguard democratic processes.
Source: The Guardian