×
Deepfakes Threaten Democracy: Google Study Reveals AI’s Role in Swaying Public Opinion
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

A new study from Google’s DeepMind reveals that the most common misuse of AI is creating political deepfakes to sway public opinion, raising concerns about the impact on elections and the spread of misinformation.

Key findings: The research, conducted in collaboration with Google’s Jigsaw unit, analyzed around 200 incidents of AI misuse and found that:

  • Creating realistic fake images, videos, and audio of politicians and celebrities was the most prevalent misuse, nearly twice as common as the next highest category.
  • Shaping public opinion was the primary goal, accounting for 27% of misuse cases, followed by financial gain through services like generating deepfakes or fake news articles.
  • Most incidents involved easily accessible tools requiring minimal technical expertise, enabling a wider range of bad actors to misuse AI.

Implications for elections and democracy: The prevalence of political deepfakes is particularly concerning given their potential to influence voters and distort the collective understanding of sociopolitical reality:

  • Deepfakes of global leaders, including UK Prime Minister Rishi Sunak, have appeared on various social media platforms in recent months, coinciding with upcoming elections.
  • Despite efforts by platforms to label or remove such content, there are fears that audiences may not recognize the fakes, and their dissemination could sway voters.
  • Ardi Janjeva from The Alan Turing Institute emphasized the long-term risks to democracies posed by the distortion of publicly accessible information through AI-generated content.

Industry response and future steps: As major tech companies rush to release generative AI products to the public, they are beginning to monitor the flood of misinformation and harmful content created by their tools:

  • OpenAI recently revealed that operations linked to Russia, China, Iran, and Israel had been using its tools to create and spread disinformation.
  • Google DeepMind’s research will influence how it improves its own model safety evaluations and aims to shape how competitors and stakeholders view the manifestation of AI-related harms.
  • The findings highlight the need for continued monitoring, research, and development of strategies to mitigate the misuse of AI, particularly in the context of political manipulation.

Analyzing deeper: While the study sheds light on the current landscape of AI misuse, it also underscores the ongoing challenge of staying ahead of malicious actors as the technology becomes more accessible and sophisticated. As generative AI tools become more integrated into various aspects of society, it is crucial for policymakers, tech companies, and the public to remain vigilant and proactive in addressing the potential risks to democracy and the integrity of information. Collaboration between stakeholders and continued research will be essential in developing effective countermeasures and fostering a more resilient information ecosystem in the face of AI-powered manipulation.

Political deepfakes are the most popular way to misuse AI

Recent News

Nvidia’s new AI agents can search and summarize huge quantities of visual data

NVIDIA's new AI Blueprint combines computer vision and generative AI to enable efficient analysis of video and image content, with potential applications across industries and smart city initiatives.

How Boulder schools balance AI innovation with student data protection

Colorado school districts embrace AI in classrooms, focusing on ethical use and data privacy while preparing students for a tech-driven future.

Microsoft Copilot Vision nears launch — here’s what we know right now

Microsoft's new AI feature can analyze on-screen content, offering contextual assistance without the need for additional searches or explanations.