×
North Korean hackers are using AI to infiltrate workplaces
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI-powered impersonation threatens workforce security: Recent developments in artificial intelligence have enabled sophisticated impersonation techniques, posing significant risks to companies’ hiring processes and overall security.

North Korean threat actors lead the charge: State-sponsored hackers from North Korea are at the forefront of this emerging threat, using a combination of deepfake technology and stolen American identities to infiltrate organizations.

  • The FBI warned in May 2022 about North Korean IT workers posing as non-North Korean nationals to gain employment and fund weapons development.
  • By October 2023, the FBI issued additional guidance on identifying deepfake job candidates, citing red flags such as reluctance to appear on camera and mismatched social media profiles.
  • In May 2024, the Department of Justice announced arrests related to a scheme targeting Fortune 500 companies using stolen American identities.

The AI interview advantage: Generative AI tools are being leveraged by both legitimate job seekers and fraudsters to gain an edge in the hiring process.

  • A 2024 Capterra survey found that 58% of job seekers use AI in their job search, with 83% admitting to exaggerating or lying about their skills using AI.
  • Software like Interview Copilot and Sensei AI can generate tailored answers to interview questions in real-time, often undetectable by interviewers.
  • These tools, combined with deepfake technology for creating fake ID documents and profile photos, allow threat actors to bypass traditional hiring and background check procedures.

Existing security measures fall short: Current hiring practices and identity verification methods are proving insufficient to combat this sophisticated threat.

  • Visual confirmation of identity and asking candidates to show photo ID on video calls can be easily circumvented by high-quality deepfakes and counterfeit IDs.
  • The case of security firm KnowBe4 unknowingly hiring a North Korean spy demonstrates the effectiveness of these impersonation techniques in bypassing multiple layers of security.

Enhanced prevention strategies: Organizations need to implement more robust measures to protect against AI-powered impersonation attempts.

  • Be alert for signs of candidates using AI interview aids, such as delayed responses or suspiciously vague answers.
  • Implement strong identity verification (IDV) systems at new user account provisioning, using factors that cannot be phished or fooled by deepfakes.
  • Consider reverifying existing employees using scalable, automated, and trustworthy methods that go beyond simple email or phone passcodes.

Broader implications: The rise of AI-powered impersonation in the workforce highlights the need for a fundamental shift in how organizations approach hiring and security.

  • As AI technology continues to advance, the line between legitimate job seekers and threat actors will become increasingly blurred.
  • Companies must balance the benefits of AI-assisted hiring processes with the potential security risks they introduce.
  • This trend may lead to a reevaluation of remote work policies and a return to more in-person interactions during the hiring process.

By implementing stronger identity verification measures and staying vigilant against AI-powered impersonation attempts, organizations can better protect themselves from sophisticated threats while maintaining efficient hiring practices in an increasingly digital world.

Is your coworker a North Korean hacker? How AI impersonation is compromising the workforce

Recent News

GE HealthCare is using AI to transform every aspect of patient care

GE HealthCare aims to triple its AI-enabled medical devices while using previously untapped hospital data to improve patient care and reduce physician burnout.

What experts say about AI chatbots for mental health support

Mental health experts warn that while AI chatbots offer accessible emotional support, they lack crucial safety protocols and clinical expertise found in traditional therapy.

Fury vs Usyk heavyweight boxing championship to be the first ever judged by AI

Historic title fight between Fury and Usyk will feature an AI judge alongside human officials, though its scores won't affect the official result.