×
OpenAI accused of profiting from model inspection in NYT lawsuit
Written by
Published on
Join our daily newsletter for breaking news, product launches and deals, research breakdowns, and other industry-leading AI coverage
Join Now

The struggle between technology companies and media organizations over AI model transparency and copyright protection has reached a critical juncture in the legal battle between OpenAI and The New York Times.

Core dispute: OpenAI’s proposed model inspection protocol has sparked controversy over access costs and limitations placed on the examination process.

  • OpenAI suggested allowing NYT to hire an expert to review confidential materials in a controlled environment
  • The company proposed capping queries at $15,000 worth of retail credits, with additional queries charged at half-retail prices
  • NYT estimates needing $800,000 worth of credits for a thorough inspection, claiming OpenAI’s pricing far exceeds actual costs

Legal implications: The case highlights significant challenges in balancing intellectual property protection with the need for transparency in AI development.

  • OpenAI defends its query cap as necessary to prevent unrestricted investigation and limit operational burden
  • Technical difficulties have hampered NYT’s attempts to inspect OpenAI’s training data
  • The outcome could set precedents for future cases involving AI model inspection rights

Regulatory context: Current AI safety testing frameworks in the United States reveal gaps in oversight and accountability.

  • The AI Safety Institute (AISI) is designed to test AI models for potential harms before deployment
  • Participation in AISI testing remains voluntary for AI companies
  • Concerns exist about AISI’s funding adequacy to fulfill its mandate effectively

Technical considerations: Model inspection methods vary in effectiveness and accessibility.

  • Public models are generally easier to examine for potential issues
  • API access to original models provides more comprehensive evidence gathering capabilities
  • Without robust government testing protocols, the public largely depends on AI companies’ internal safety measures

Future implications: This legal dispute illuminates growing tensions between AI innovation and accountability, raising questions about how to balance commercial interests with public oversight in the rapidly evolving AI landscape.

  • The case may establish important precedents for future AI model inspections
  • Cost structures for model access could influence future litigation strategies
  • The outcome might shape how AI companies approach transparency and external auditing
OpenAI accused of trying to profit off AI model inspection in court

Recent News

67% of EU businesses struggle to understand AI Act compliance

Critical guidance remains unpublished just weeks before key deadlines take effect.

Google AI Pro now offers annual billing at $199.99, saving users 16%

The plan bundles 2TB storage with Gemini access and video generation tools.

Everyday AI Value: Five Below’s 4-step AI blueprint drives 19.5% sales growth

Strategic focus on business constraints beats the typical "scaling meetings" trap that derails most AI initiatives.