×
OpenAI accused of profiting from model inspection in NYT lawsuit
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The struggle between technology companies and media organizations over AI model transparency and copyright protection has reached a critical juncture in the legal battle between OpenAI and The New York Times.

Core dispute: OpenAI’s proposed model inspection protocol has sparked controversy over access costs and limitations placed on the examination process.

  • OpenAI suggested allowing NYT to hire an expert to review confidential materials in a controlled environment
  • The company proposed capping queries at $15,000 worth of retail credits, with additional queries charged at half-retail prices
  • NYT estimates needing $800,000 worth of credits for a thorough inspection, claiming OpenAI’s pricing far exceeds actual costs

Legal implications: The case highlights significant challenges in balancing intellectual property protection with the need for transparency in AI development.

  • OpenAI defends its query cap as necessary to prevent unrestricted investigation and limit operational burden
  • Technical difficulties have hampered NYT’s attempts to inspect OpenAI’s training data
  • The outcome could set precedents for future cases involving AI model inspection rights

Regulatory context: Current AI safety testing frameworks in the United States reveal gaps in oversight and accountability.

  • The AI Safety Institute (AISI) is designed to test AI models for potential harms before deployment
  • Participation in AISI testing remains voluntary for AI companies
  • Concerns exist about AISI’s funding adequacy to fulfill its mandate effectively

Technical considerations: Model inspection methods vary in effectiveness and accessibility.

  • Public models are generally easier to examine for potential issues
  • API access to original models provides more comprehensive evidence gathering capabilities
  • Without robust government testing protocols, the public largely depends on AI companies’ internal safety measures

Future implications: This legal dispute illuminates growing tensions between AI innovation and accountability, raising questions about how to balance commercial interests with public oversight in the rapidly evolving AI landscape.

  • The case may establish important precedents for future AI model inspections
  • Cost structures for model access could influence future litigation strategies
  • The outcome might shape how AI companies approach transparency and external auditing
OpenAI accused of trying to profit off AI model inspection in court

Recent News

7 ways to optimize your business for ChatGPT recommendations

Companies must adapt their digital strategy with specific expertise, consistent information across platforms, and authoritative content to appear in AI-powered recommendation results.

Robin Williams’ daughter Zelda slams OpenAI’s Ghibli-style images amid artistic and ethical concerns

Robin Williams' daughter condemns OpenAI's AI-generated Ghibli-style images, highlighting both environmental costs and the contradiction with Miyazaki's well-documented opposition to artificial intelligence in creative work.

AI search tools provide wrong answers up to 60% of the time despite growing adoption

Independent testing reveals AI search tools frequently provide incorrect information, with error rates ranging from 37% to 94% across major platforms despite their growing popularity as Google alternatives.