×
UK targets AI-related national security threats with new LASR initiative
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The United Kingdom has launched a new research laboratory focused on addressing AI-related national security threats, marking an intensified approach to defending against emerging technological risks.

Key initiative details: The Laboratory for AI Security Research (LASR) was announced by Pat McFadden, chancellor of the Duchy of Lancaster, during a NATO Cyber Defence Conference in London.

  • The UK government has committed £8.2 million in initial funding for the laboratory
  • Multiple government departments are involved, including FCDO, DSIT, GCHQ, NCSC, and the MOD’s Defence Science and Technology Laboratory
  • Private sector partners include the Alan Turing Institute, University of Oxford, Queen’s University Belfast, and Plexal

Strategic focus and threats: The laboratory aims to assess and counter AI-based security challenges in both cyber and physical domains.

  • AI and machine learning are increasingly being used to automate cyber attacks and evade detection systems
  • The initiative specifically addresses concerns about state-sponsored hacker groups utilizing AI capabilities
  • McFadden explicitly named Russia as a primary threat, stating that the UK is actively monitoring and countering their attacks

Collaborative approach: LASR represents a multi-stakeholder effort to combine expertise from various sectors.

  • The laboratory brings together experts from industry, academia, and government
  • The AI Safety Institute will contribute its expertise, though there appears to be some overlap in mission
  • Private sector organizations are being invited to provide additional funding and support

International context: The launch comes amid growing concerns about the effectiveness of global AI governance agreements.

  • The initiative follows the Bletchley Declaration, a multilateral pledge by 28 countries to ensure responsible AI development
  • The creation of LASR suggests skepticism about the effectiveness of international commitments to responsible AI development
  • The UK acknowledges an ongoing AI arms race with potential adversaries

Looking ahead: The security paradox: While LASR represents a significant step in defending against AI threats, it also highlights the growing tension between AI’s defensive capabilities and its potential for weaponization.

  • AI technology offers enhanced cyber defense tools and intelligence gathering capabilities
  • However, these same advances can be turned against their creators, creating a complex security challenge
  • The race to stay ahead of adversaries while maintaining responsible development practices will likely remain a critical balance
UK points LASR at AI-based threats to national security

Recent News

Register now: Dec. ‘Ideathon’ to discuss assignment of legal responsibility in AI era 

Experts and developers collaborate on new frameworks to assign legal responsibility as AI systems become more autonomous and complex.

Newcomer ‘SmolVLM’ is a small but mighty Vision Language Model

A lightweight vision-language model running on modest hardware shows that advanced AI capabilities don't always require massive computing resources.

Enterprises embrace AI and IoT to drive secure growth

Large enterprises are prioritizing tech investments to drive growth and customer experience through 2030, with AI and 5G receiving significant portions of budgets that will reach 11% of revenue.