×
US Government Partners with OpenAI and Anthropic for AI Safety Testing
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

AI safety collaboration takes center stage: OpenAI and Anthropic have entered into groundbreaking agreements with the US government, granting early access to their latest AI models for safety testing before public release.

  • The National Institute of Standards and Technology (NIST) announced formal collaborations with both companies and the US Artificial Intelligence Safety Institute to conduct AI safety research, testing, and evaluation.
  • This partnership aims to ensure that public safety assessments are not solely dependent on the companies’ internal evaluations but also include collaborative research with the US government.
  • The US AI Safety Institute will work in conjunction with its UK counterpart to examine models and identify potential safety risks.

Broader context of AI regulation: The agreements come amid ongoing debates about AI regulation and safety measures at both state and federal levels.

  • California is on the verge of passing one of the country’s first AI safety bills, which includes controversial provisions such as requiring AI companies to implement a “kill switch” for models that could pose novel threats to public safety.
  • Critics argue that the California bill may overlook existing AI risks while potentially stifling innovation, urging Governor Gavin Newsom to veto the legislation.
  • Anthropic has cautiously supported the California bill after recent amendments, while OpenAI has joined critics in opposing it.

Industry perspectives on AI safety: AI companies express varying views on the balance between safety measures and innovation in the rapidly evolving field.

  • Anthropic’s co-founder, Jack Clark, emphasized that safe and trustworthy AI is crucial for the technology’s positive impact, supporting collaboration with the US AI Safety Institute.
  • OpenAI’s chief strategy officer, Jason Kwon, advocated for federal leadership in regulating frontier AI models, citing implications for national security and competitiveness.
  • Both companies acknowledge the importance of safety in driving technological innovation, albeit with different approaches to regulation.

Government’s role in AI safety: The US government is taking an active stance in AI safety research and evaluation through these collaborations.

  • Elizabeth Kelly, director of the US AI Safety Institute, described the agreements as an important milestone in responsibly stewarding the future of AI.
  • The institute plans to conduct its own research to advance the science of AI safety, leveraging the government’s expertise to rigorously test models before widespread deployment.
  • This collaboration aims to provide feedback to OpenAI and Anthropic on potential safety improvements for their models.

Implications for AI development and deployment: The partnerships between AI companies and government agencies signal a shift towards more collaborative approaches to AI safety.

  • These agreements build upon voluntary AI safety commitments previously made by AI companies to the Biden administration.
  • The collaboration may serve as a framework for global AI safety efforts, potentially influencing international standards and practices.
  • By involving government agencies in pre-release testing, the initiative aims to address public concerns about AI safety while supporting continued innovation in the field.

Balancing innovation and regulation: The differing stances of OpenAI and Anthropic on state-level regulation highlight the ongoing challenge of finding the right balance between fostering innovation and ensuring public safety.

  • While Anthropic supports California’s AI safety bill with some reservations, OpenAI argues for federal-level regulation to address national security and competitiveness concerns.
  • These contrasting positions reflect the broader debate within the tech industry about the most effective approach to AI governance and safety measures.
  • The collaboration with the US AI Safety Institute may represent a middle ground, allowing for government oversight while maintaining the pace of technological advancement.

Looking ahead: Potential impacts and challenges: As these collaborations unfold, several key questions and considerations emerge for the future of AI development and regulation.

  • The effectiveness of pre-release testing in identifying and mitigating potential risks associated with advanced AI models remains to be seen.
  • The balance between transparency and protecting proprietary information may pose challenges as government agencies gain early access to cutting-edge AI technologies.
  • The outcomes of these partnerships could significantly influence future AI policies and regulations at both national and international levels, potentially setting precedents for government-industry collaborations in emerging technologies.
Feds to get early access to OpenAI, Anthropic AI to test for doomsday scenarios

Recent News

7 ways to optimize your business for ChatGPT recommendations

Companies must adapt their digital strategy with specific expertise, consistent information across platforms, and authoritative content to appear in AI-powered recommendation results.

Robin Williams’ daughter Zelda slams OpenAI’s Ghibli-style images amid artistic and ethical concerns

Robin Williams' daughter condemns OpenAI's AI-generated Ghibli-style images, highlighting both environmental costs and the contradiction with Miyazaki's well-documented opposition to artificial intelligence in creative work.

AI search tools provide wrong answers up to 60% of the time despite growing adoption

Independent testing reveals AI search tools frequently provide incorrect information, with error rates ranging from 37% to 94% across major platforms despite their growing popularity as Google alternatives.