×
US Government Partners with OpenAI and Anthropic for AI Safety Testing
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI safety collaboration takes center stage: OpenAI and Anthropic have entered into groundbreaking agreements with the US government, granting early access to their latest AI models for safety testing before public release.

  • The National Institute of Standards and Technology (NIST) announced formal collaborations with both companies and the US Artificial Intelligence Safety Institute to conduct AI safety research, testing, and evaluation.
  • This partnership aims to ensure that public safety assessments are not solely dependent on the companies’ internal evaluations but also include collaborative research with the US government.
  • The US AI Safety Institute will work in conjunction with its UK counterpart to examine models and identify potential safety risks.

Broader context of AI regulation: The agreements come amid ongoing debates about AI regulation and safety measures at both state and federal levels.

  • California is on the verge of passing one of the country’s first AI safety bills, which includes controversial provisions such as requiring AI companies to implement a “kill switch” for models that could pose novel threats to public safety.
  • Critics argue that the California bill may overlook existing AI risks while potentially stifling innovation, urging Governor Gavin Newsom to veto the legislation.
  • Anthropic has cautiously supported the California bill after recent amendments, while OpenAI has joined critics in opposing it.

Industry perspectives on AI safety: AI companies express varying views on the balance between safety measures and innovation in the rapidly evolving field.

  • Anthropic’s co-founder, Jack Clark, emphasized that safe and trustworthy AI is crucial for the technology’s positive impact, supporting collaboration with the US AI Safety Institute.
  • OpenAI’s chief strategy officer, Jason Kwon, advocated for federal leadership in regulating frontier AI models, citing implications for national security and competitiveness.
  • Both companies acknowledge the importance of safety in driving technological innovation, albeit with different approaches to regulation.

Government’s role in AI safety: The US government is taking an active stance in AI safety research and evaluation through these collaborations.

  • Elizabeth Kelly, director of the US AI Safety Institute, described the agreements as an important milestone in responsibly stewarding the future of AI.
  • The institute plans to conduct its own research to advance the science of AI safety, leveraging the government’s expertise to rigorously test models before widespread deployment.
  • This collaboration aims to provide feedback to OpenAI and Anthropic on potential safety improvements for their models.

Implications for AI development and deployment: The partnerships between AI companies and government agencies signal a shift towards more collaborative approaches to AI safety.

  • These agreements build upon voluntary AI safety commitments previously made by AI companies to the Biden administration.
  • The collaboration may serve as a framework for global AI safety efforts, potentially influencing international standards and practices.
  • By involving government agencies in pre-release testing, the initiative aims to address public concerns about AI safety while supporting continued innovation in the field.

Balancing innovation and regulation: The differing stances of OpenAI and Anthropic on state-level regulation highlight the ongoing challenge of finding the right balance between fostering innovation and ensuring public safety.

  • While Anthropic supports California’s AI safety bill with some reservations, OpenAI argues for federal-level regulation to address national security and competitiveness concerns.
  • These contrasting positions reflect the broader debate within the tech industry about the most effective approach to AI governance and safety measures.
  • The collaboration with the US AI Safety Institute may represent a middle ground, allowing for government oversight while maintaining the pace of technological advancement.

Looking ahead: Potential impacts and challenges: As these collaborations unfold, several key questions and considerations emerge for the future of AI development and regulation.

  • The effectiveness of pre-release testing in identifying and mitigating potential risks associated with advanced AI models remains to be seen.
  • The balance between transparency and protecting proprietary information may pose challenges as government agencies gain early access to cutting-edge AI technologies.
  • The outcomes of these partnerships could significantly influence future AI policies and regulations at both national and international levels, potentially setting precedents for government-industry collaborations in emerging technologies.
Feds to get early access to OpenAI, Anthropic AI to test for doomsday scenarios

Recent News

‘Heretic’ film directors include anti-AI disclaimer in film credits

Hollywood directors' anti-AI stance reflects growing concerns about automation in creative industries and its potential impact on jobs.

AI at the edge: Key architecture decisions for future success

Edge intelligence brings AI processing closer to data sources, enabling faster and more reliable decision-making across industries.

Why new AI data centers may spike Americans’ electricity bills

The growing energy demands of AI data centers are causing electricity costs to rise for consumers in some parts of the U.S., highlighting the unintended consequences of rapid technological expansion.