×
North Korean hackers are using AI to infiltrate workplaces
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI-powered impersonation threatens workforce security: Recent developments in artificial intelligence have enabled sophisticated impersonation techniques, posing significant risks to companies’ hiring processes and overall security.

North Korean threat actors lead the charge: State-sponsored hackers from North Korea are at the forefront of this emerging threat, using a combination of deepfake technology and stolen American identities to infiltrate organizations.

  • The FBI warned in May 2022 about North Korean IT workers posing as non-North Korean nationals to gain employment and fund weapons development.
  • By October 2023, the FBI issued additional guidance on identifying deepfake job candidates, citing red flags such as reluctance to appear on camera and mismatched social media profiles.
  • In May 2024, the Department of Justice announced arrests related to a scheme targeting Fortune 500 companies using stolen American identities.

The AI interview advantage: Generative AI tools are being leveraged by both legitimate job seekers and fraudsters to gain an edge in the hiring process.

  • A 2024 Capterra survey found that 58% of job seekers use AI in their job search, with 83% admitting to exaggerating or lying about their skills using AI.
  • Software like Interview Copilot and Sensei AI can generate tailored answers to interview questions in real-time, often undetectable by interviewers.
  • These tools, combined with deepfake technology for creating fake ID documents and profile photos, allow threat actors to bypass traditional hiring and background check procedures.

Existing security measures fall short: Current hiring practices and identity verification methods are proving insufficient to combat this sophisticated threat.

  • Visual confirmation of identity and asking candidates to show photo ID on video calls can be easily circumvented by high-quality deepfakes and counterfeit IDs.
  • The case of security firm KnowBe4 unknowingly hiring a North Korean spy demonstrates the effectiveness of these impersonation techniques in bypassing multiple layers of security.

Enhanced prevention strategies: Organizations need to implement more robust measures to protect against AI-powered impersonation attempts.

  • Be alert for signs of candidates using AI interview aids, such as delayed responses or suspiciously vague answers.
  • Implement strong identity verification (IDV) systems at new user account provisioning, using factors that cannot be phished or fooled by deepfakes.
  • Consider reverifying existing employees using scalable, automated, and trustworthy methods that go beyond simple email or phone passcodes.

Broader implications: The rise of AI-powered impersonation in the workforce highlights the need for a fundamental shift in how organizations approach hiring and security.

  • As AI technology continues to advance, the line between legitimate job seekers and threat actors will become increasingly blurred.
  • Companies must balance the benefits of AI-assisted hiring processes with the potential security risks they introduce.
  • This trend may lead to a reevaluation of remote work policies and a return to more in-person interactions during the hiring process.

By implementing stronger identity verification measures and staying vigilant against AI-powered impersonation attempts, organizations can better protect themselves from sophisticated threats while maintaining efficient hiring practices in an increasingly digital world.

Is your coworker a North Korean hacker? How AI impersonation is compromising the workforce

Recent News

7 ways to optimize your business for ChatGPT recommendations

Companies must adapt their digital strategy with specific expertise, consistent information across platforms, and authoritative content to appear in AI-powered recommendation results.

Robin Williams’ daughter Zelda slams OpenAI’s Ghibli-style images amid artistic and ethical concerns

Robin Williams' daughter condemns OpenAI's AI-generated Ghibli-style images, highlighting both environmental costs and the contradiction with Miyazaki's well-documented opposition to artificial intelligence in creative work.

AI search tools provide wrong answers up to 60% of the time despite growing adoption

Independent testing reveals AI search tools frequently provide incorrect information, with error rates ranging from 37% to 94% across major platforms despite their growing popularity as Google alternatives.