×
North Korean hackers are using AI to infiltrate workplaces
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI-powered impersonation threatens workforce security: Recent developments in artificial intelligence have enabled sophisticated impersonation techniques, posing significant risks to companies’ hiring processes and overall security.

North Korean threat actors lead the charge: State-sponsored hackers from North Korea are at the forefront of this emerging threat, using a combination of deepfake technology and stolen American identities to infiltrate organizations.

  • The FBI warned in May 2022 about North Korean IT workers posing as non-North Korean nationals to gain employment and fund weapons development.
  • By October 2023, the FBI issued additional guidance on identifying deepfake job candidates, citing red flags such as reluctance to appear on camera and mismatched social media profiles.
  • In May 2024, the Department of Justice announced arrests related to a scheme targeting Fortune 500 companies using stolen American identities.

The AI interview advantage: Generative AI tools are being leveraged by both legitimate job seekers and fraudsters to gain an edge in the hiring process.

  • A 2024 Capterra survey found that 58% of job seekers use AI in their job search, with 83% admitting to exaggerating or lying about their skills using AI.
  • Software like Interview Copilot and Sensei AI can generate tailored answers to interview questions in real-time, often undetectable by interviewers.
  • These tools, combined with deepfake technology for creating fake ID documents and profile photos, allow threat actors to bypass traditional hiring and background check procedures.

Existing security measures fall short: Current hiring practices and identity verification methods are proving insufficient to combat this sophisticated threat.

  • Visual confirmation of identity and asking candidates to show photo ID on video calls can be easily circumvented by high-quality deepfakes and counterfeit IDs.
  • The case of security firm KnowBe4 unknowingly hiring a North Korean spy demonstrates the effectiveness of these impersonation techniques in bypassing multiple layers of security.

Enhanced prevention strategies: Organizations need to implement more robust measures to protect against AI-powered impersonation attempts.

  • Be alert for signs of candidates using AI interview aids, such as delayed responses or suspiciously vague answers.
  • Implement strong identity verification (IDV) systems at new user account provisioning, using factors that cannot be phished or fooled by deepfakes.
  • Consider reverifying existing employees using scalable, automated, and trustworthy methods that go beyond simple email or phone passcodes.

Broader implications: The rise of AI-powered impersonation in the workforce highlights the need for a fundamental shift in how organizations approach hiring and security.

  • As AI technology continues to advance, the line between legitimate job seekers and threat actors will become increasingly blurred.
  • Companies must balance the benefits of AI-assisted hiring processes with the potential security risks they introduce.
  • This trend may lead to a reevaluation of remote work policies and a return to more in-person interactions during the hiring process.

By implementing stronger identity verification measures and staying vigilant against AI-powered impersonation attempts, organizations can better protect themselves from sophisticated threats while maintaining efficient hiring practices in an increasingly digital world.

Is your coworker a North Korean hacker? How AI impersonation is compromising the workforce

Recent News

Social network Bluesky says it won’t train AI on user posts

As social media platforms debate AI training practices, Bluesky stakes out a pro-creator stance by pledging not to use user content for generative AI.

New research explores how cutting-edge AI may advance quantum computing

AI is being leveraged to address key challenges in quantum computing, from hardware design to error correction.

Navigating the ethical minefield of AI-powered customer segmentation

AI-driven customer segmentation provides deeper insights into consumer behavior, but raises concerns about privacy and potential bias.