Signal/Noise
Signal/Noise
2025-12-05
While AI companies race to build bigger models and grab headlines with trillion-dollar valuations, the real action is happening in the unglamorous business of making AI actually work reliably at scale. The gap between AI demos and production reality is creating a hidden infrastructure play that will determine which companies survive the inevitable consolidation.
The Great AI Reality Check: When Silicon Valley Dreams Meet Production Nightmares
Beneath the venture capital euphoria and billion-dollar AI startups lies an uncomfortable truth: most AI systems are brittle, unreliable, and nowhere near production-ready. Anthropic’s internal research reveals that even their own engineers can only “fully delegate” 0-20% of their work to Claude, despite claiming massive productivity gains. Meanwhile, coding agents—supposedly the poster child for AI automation—are failing spectacularly when faced with real-world complexity. They break when context windows overflow, fumble basic refactoring, and lack the operational awareness to handle production environments. This isn’t a temporary growing pain; it’s a fundamental architecture problem. The AI industry has optimized for demo-ability over deployability, creating systems that wow in controlled settings but crumble under real-world pressure. The companies that recognize this gap and build boring, reliable infrastructure will capture disproportionate value as the market matures. Look for businesses focused on data quality, model reliability, and operational monitoring—the plumbing that makes AI actually work.
The Data Gold Rush: How Training Data Became the New Oil (And Why It’s Getting Dirty)
The AI training data market has exploded from virtually nothing to a multi-billion dollar industry, with companies like Micro1 crossing $100M ARR in eight months by connecting domain experts with AI labs hungry for high-quality human feedback. But this gold rush is creating its own problems. Academic researchers are warning of a “slop problem”—low-quality, AI-generated content polluting training datasets and degrading model performance. Meanwhile, the race for specialized human trainers has created a new gig economy where Harvard professors earn $100/hour grading AI outputs. This isn’t sustainable. As models become more capable, the bar for useful human feedback rises exponentially. Companies are already struggling to find experts who can meaningfully improve frontier models. The winning strategy isn’t just accumulating more data—it’s building systems that can identify and filter high-quality training signals while maintaining data integrity at scale. The firms that solve this curation problem will control the chokepoint between raw human expertise and AI capability.
The Platform Wars Are Over Before They Started
While OpenAI panics about ChatGPT’s “code red” competitive situation and races to build AI agents, the real platform battle is being won by the infrastructure layer. Nvidia’s position remains unassailable not because of GPU performance, but because they control the entire stack from silicon to software. Their CUDA ecosystem creates switching costs that make even trillion-dollar competitors think twice about alternatives. Meanwhile, Google’s Gemini 3 launch signals a different strategy: embedding AI so deeply into existing workflows that users never have to choose a “primary” AI assistant. This isn’t about building the best chatbot; it’s about becoming invisible infrastructure. Meta’s poaching of Apple’s top designers reveals another angle—the winners will be companies that make AI feel like a natural extension of existing tools rather than a separate application. The consumer AI platform war was decided before it began: the platforms that already own distribution (Google, Apple, Microsoft) will win by making AI a feature, not a product.
Questions
- If AI coding agents can’t handle production complexity, what does this mean for the $7 trillion infrastructure buildout everyone is betting on?
- When training data quality becomes the limiting factor, do we end up with a few AI monopolies controlling the best datasets?
- Is the current AI bubble actually two bubbles—one for capabilities that will deflate, and another for infrastructure that will grow?
Past Briefings
OpenAI Deleted ‘Safely.’ NVIDIA Reports. Karpathy Is Still Learning
THE NUMBER: 6 — times OpenAI changed its mission in 9 years. The most recent edit deleted one word: safely. TL;DR Andrej Karpathy — the engineer who wrote the curriculum that trained a generation of developers, ran AI at Tesla, and helped found OpenAI — posted in December that he's never felt so behind as a programmer. Fourteen million people saw it. Tonight, NVIDIA reports Q4 fiscal 2026 earnings after market close: analysts expect $65.7 billion in revenue, up 67% year over year. The numbers will almost certainly land. What matters is what Jensen Huang says about the next two quarters to...
Feb 23, 2026Altman lied about a handshake on camera. CrowdStrike fell 8%. Google just killed the $3,000 photo shoot.
Sam Altman told reporters he was "confused" when Narendra Modi grabbed his hand at the India AI Impact Summit. He said he "wasn't sure what was happening." The video, which has been watched by tens of millions of people, shows Altman looking directly at Dario Amodei before raising his fist. He knew exactly what was happening. He chose not to do it, and then he lied about it. On camera. In multiple interviews. With the footage playing on every screen behind him. That would be a minor character note in any other industry. In this one, it isn't. Because on...
Feb 20, 2026We’re Building the Agentic Web Faster Than We’re Protecting It
Google's WebMCP gives agents structured access to every website. Anthropic's data shows autonomy doubling with oversight thinning. OpenAI's agent already drains crypto vaults. Google shipped working code Thursday that hands AI agents a structured key to every website on the internet. WebMCP, running in Chrome 146 Canary, lets sites expose machine-readable "Tool Contracts" so agents can book a flight, file a support ticket, or complete a checkout without parsing screenshots or scraping HTML. Early benchmarks show 67% less compute overhead than visual approaches. Microsoft co-authored the spec. The W3C is incubating it. This isn't a proposal. It's production software already...