Signal/Noise
Signal/Noise
2025-12-12
Disney’s $1B OpenAI deal isn’t about Mickey Mouse videos—it’s the moment media companies pivoted from resisting AI to weaponizing it. While everyone debates Trump’s AI executive order blocking state regulations, the real story is how content owners are racing to monetize their moats before AI makes ownership meaningless. We’re witnessing the great IP arbitrage: those with content libraries are cashing out while they still can.
The Great IP Liquidation Sale Has Begun
Disney’s $1B OpenAI partnership represents the most significant shift in media strategy since Netflix went streaming-first. But strip away the Mickey Mouse headlines and you see something more profound: content owners are racing to monetize their IP before AI makes ownership irrelevant.
The same day Disney announced its OpenAI deal, it sent cease-and-desist letters to Google, accusing them of training on Disney content without permission. This isn’t contradiction—it’s strategy. Disney is simultaneously licensing its characters to OpenAI while using legal warfare to prevent competitors from getting the same content for free. They’re creating artificial scarcity in an AI world that threatens to make all content infinitely replicable.
The timing is everything. Disney CEO Bob Iger has been telegraphing this move for months, talking about AI coming to Disney+. But the urgency became clear when Chinese firm DeepSeek spooked markets with a breakthrough model on Trump’s inauguration day. The message was unmistakable: the AI race had entered a new phase, and content companies needed to pick sides fast.
What Disney figured out that other media companies haven’t: their real asset isn’t Mickey Mouse cartoons, it’s the legal right to Mickey Mouse. In an AI world where anyone can generate infinite Mickey-adjacent content, the only value is in legitimate licensing. Disney isn’t selling content—they’re selling permission. And in a world of legal gray areas around AI training, permission is the new scarcity.
This explains why Disney is getting $1B while other content creators get nothing. They have the lawyers, the copyright registrations, and the enforcement apparatus to make their permission stick. Smaller creators don’t. The IP arbitrage isn’t just about having good content—it’s about having defensible content ownership at scale.
Trump’s AI Order Reveals The Real Regulatory Capture
While everyone focuses on Trump’s executive order blocking state AI regulations, the more interesting story is who’s actually writing AI policy. David Sacks, Trump’s AI czar, is a venture capitalist with extensive AI investments who stood at Trump’s shoulder during the signing. This isn’t regulatory capture—it’s regulatory abdication.
The conventional narrative frames this as federal vs. state power. The reality is simpler: AI companies convinced Trump that a patchwork of state regulations would slow American competitiveness against China. But look at the specifics and you see something more nuanced. The order targets laws requiring algorithmic bias testing and transparency—exactly the kinds of consumer protections that create compliance costs for AI companies.
California leads with 70+ AI laws since 2016, including requirements for safety testing and bias audits. Colorado mandated algorithmic discrimination assessments. These aren’t anti-innovation—they’re basic consumer protection. But in an AI arms race framed as existential competition with China, consumer protection becomes a luxury America can’t afford.
The real tell is in the exemptions. The order specifically carves out protections for children and data center infrastructure—areas where tech companies actually want federal standards. They’re not against all regulation; they’re against regulations that slow deployment or increase costs. The pattern is clear: federalize the rules that help big tech, preempt the rules that constrain them.
What makes this different from typical regulatory capture is the speed and scope. Usually, industries spend years lobbying for favorable rules. Here, AI companies convinced a president to preempt unfavorable state laws before they could take effect. It’s proactive regulatory arbitrage—preventing bad rules rather than overturning them.
The Productivity Paradox Goes Agentic
OpenAI’s GPT-5.2 release, coming weeks after an internal ‘code red’ memo, reveals something crucial about the current AI moment: we’re optimizing for benchmarks that don’t translate to business value. The model beats industry professionals 71% of the time on ‘well-specified professional tasks’—but those aren’t the tasks that matter.
The productivity paradox that economist Robert Solow identified in the 1980s—’you can see the computer age everywhere except in the productivity statistics’—is playing out again with AI. Companies are deploying ChatGPT Enterprise to 120,000+ employees, building elaborate agent frameworks, and achieving impressive demo results. But actual productivity gains remain elusive.
The problem isn’t capability—it’s context. AI excels at isolated tasks with clear parameters but struggles with the messy, interdependent work that defines most jobs. A travel booking agent can handle simple requests but breaks down when flights get canceled and hotels overbooked. An HR chatbot answers policy questions but can’t navigate the political dynamics of actually implementing those policies.
This creates what we’re calling the ‘last mile problem’ for enterprise AI. Companies can automate 80% of a workflow, but that final 20%—the exceptions, edge cases, and human judgment calls—still requires full human oversight. Instead of 5x productivity gains, they get 1.2x improvements wrapped in 10x complexity.
The race to deploy AI agents ignores this fundamental limitation. Every AI implementation creates new categories of work: monitoring AI outputs, handling AI failures, and managing human-AI handoffs. The total work doesn’t decrease—it just shifts to different types of cognitive overhead that are harder to measure and manage.
Questions
- If Disney’s licensing deal with OpenAI is really about legal permission rather than content, what happens when the first major fair use case establishes precedent for AI training?
- Trump’s AI order preempts state consumer protection laws in favor of federal standards that don’t exist yet—is this regulatory vacuum intentional or incompetent?
- As AI agents become more capable at isolated tasks, are we accidentally creating a new class of ‘AI middle management’ jobs to coordinate between human work and automated processes?
Past Briefings
Bill Gurley Says the AI Bubble Is About to Burst. Travis Kalanick’s Timing Says He’s Right.
THE NUMBER: $300 billion — HSBC's estimate of cumulative cash burn by foundational AI model companies through 2030. Bill Gurley sat on Uber's board while it burned $2 billion a year and says it gave him "high anxiety." OpenAI and Anthropic make Uber's bonfire look like a birthday candle. "God bless them," Gurley told CNBC. "It's a scary way to run a company." Travis Kalanick showed up on the All-In podcast this week with a new robotics venture called Atoms and opinions about who's winning the autonomy race. That's the headline most people caught. But the deeper signal is the...
Mar 17, 2026Anthropic Is Winning the Product War. The $575 Billion Question Is Whether Anyone Can Afford to Keep Fighting
THE NUMBER: 12x — For every dollar the hyperscalers earn from AI today, they're spending twelve dollars building more capacity. That's $575 billion in capex this year. Alphabet just issued a century bond — the first by a tech company since Motorola in 1997 — to fund it. The debt matures in 2126. The chips it buys will be obsolete by 2029. Anthropic now wins 70% of new enterprise deals in direct matchups with OpenAI, according to Ramp's March 2026 AI Index. Claude Code generates $2.5 billion in annualized revenue. OpenAI's Codex manages $1 billion. OpenAI's enterprise share dropped from...
Mar 16, 2026Chamath Says Your Portfolio Is Worth 75% Less Than You Think. Karpathy’s Data Suggests He’s Right.
THE NUMBER: 60-80% — the share of a typical equity valuation derived from terminal value. That's the portion of every stock price that assumes competitive advantages persist for a decade or more. Chamath Palihapitiya just argued that AI makes that assumption unpriceable. If he's even half right, the math doesn't bend. It breaks. Chamath Palihapitiya posted a note this weekend titled "The Collapse of Terminal Value" that should be required reading for anyone who allocates capital — including the capital of their own career. His thesis: AI accelerates disruption so fast that no company can credibly project cash flows beyond five...