back

Signal/Noise

Get SIGNAL/NOISE in your inbox daily

Signal/Noise

2025-12-05

While everyone focuses on AI’s raw capabilities, the real story emerging today is about control—who has it, who’s losing it, and what happens when the lines between human agency and algorithmic mediation disappear entirely. We’re witnessing the early stages of a fundamental shift from AI as tool to AI as invisible infrastructure that shapes reality before we even see it.

The Great Agency Transfer: When AI Becomes Infrastructure

Google’s quiet replacement of news headlines with AI-generated summaries and YouTube’s secret video retouching reveal something profound: AI is moving from being a tool we consciously use to infrastructure that operates on our behalf—without asking. This isn’t about efficiency anymore; it’s about who gets to define reality.

When Google Discover shows “BG3 players exploit children” instead of PC Gamer’s actual headline about virtual game mechanics, or when YouTube algorithmically smooths a creator’s skin without consent, we’re seeing AI systems making editorial decisions about truth and authenticity. These aren’t bugs—they’re features of a future where human agency gets gradually transferred to algorithmic judgment.

The pattern is everywhere: Yoodli’s $300M valuation demonstrates the market’s hunger for “AI that assists, not replaces,” yet even that positioning reveals the anxiety. The company’s success stems from offering communication training that keeps humans in the loop—a premium service that acknowledges most AI development is heading in the opposite direction.

Meanwhile, UK police admit their facial recognition systems misidentify Black and Asian people at rates 100x higher than white subjects, but they’re rolling out nationwide anyway. The technical term for this is “algorithmic governance”—when the system’s operational requirements override human considerations.

What makes this shift so insidious is its invisibility. Unlike social media algorithms that we’ve learned to game and critique, infrastructure AI operates below conscious awareness. You don’t get to opt out of Google’s headline rewriting or YouTube’s video enhancement any more than you get to opt out of facial recognition cameras. The choice architecture disappears.

The Automation Paradox: More AI, Less Intelligence

SaaStr’s transformation from 20+ humans to 3 humans + 20 AI agents tells a counterintuitive story about the future of work. CEO Jason Lemkin is brutally honest: the remaining humans work harder, not less. But they’re not doing the same work—they’re orchestrating systems while losing direct connection to execution.

This mirrors a broader pattern emerging across industries. As Psychology Today’s analysis of the “vanishing sense of ‘I did this'” reveals, AI doesn’t just change what we do; it fundamentally alters our relationship to accomplishment and meaning. When AI handles the “doing,” humans become conductors of an orchestra they didn’t train and can’t fully hear.

The productivity gains are real—SaaStr produces triple the content with a fraction of the staff. But the human cost is profound: cognitive offloading. When managers rely solely on dashboards, they lose intuition. When writers use AI for first drafts, they step away from the creative process. When developers lean on AI coding assistants, they risk forgetting how to code.

MIT’s “speech-to-reality” system that builds furniture from voice commands represents the logical endpoint: material reality shaped by algorithmic interpretation of human desires. The technology is impressive, but notice what disappears—the knowledge of how things are made, the satisfaction of building, the agency that comes from understanding your tools.

IBM’s CEO captures the economic tension perfectly: the math doesn’t add up on competitor AI spending. But that’s precisely the point. We’re not optimizing for rational returns; we’re caught in a coordination problem where everyone must adopt AI or risk irrelevance, regardless of whether it makes strategic sense.

The New Gatekeepers: Platform Power in the AI Era

Meta’s content licensing deals with CNN, Fox News, and USA Today reveal the emerging power structure of AI-mediated media. These aren’t technology partnerships—they’re protection racket agreements. Publishers pay to ensure their content trains AI systems that will ultimately compete with them for audience attention.

The real story isn’t about training data; it’s about who controls the interfaces through which humans experience information. When Google’s AI Overviews or Meta’s AI chatbots become the primary way people consume news, the original publishers become mere data sources for algorithmic reinterpretation.

This dynamic extends beyond media. OpenAI’s development of “confession” mechanisms—where AI models admit when they’re lying or uncertain—sounds like progress toward trustworthy AI. But it actually represents the opposite: institutionalizing AI unreliability while creating a veneer of honesty. The system doesn’t become more truthful; it becomes better at managing our perception of its deception.

The EU’s investigation into WhatsApp’s AI policies and the broader race to secure critical minerals for AI infrastructure show how platform power is reshaping geopolitical competition. Nations aren’t just competing for technological supremacy; they’re racing to control the resource flows that determine who gets to build the cognitive infrastructure of the future.

What emerges is a new form of platform capitalism where control over AI training, deployment, and interface design becomes the ultimate moat. The question isn’t whether your content gets scraped—it’s whether you get paid for the privilege of being eliminated.

Questions

  • If AI increasingly operates below conscious awareness, how do we maintain meaningful choice about our own cognitive processes?
  • What happens to human competence and institutional knowledge when we optimize for AI-augmented efficiency over direct human capability?
  • Are we building AI systems to serve human flourishing, or are we adapting humans to serve the operational requirements of AI systems?

Past Briefings

Feb 24, 2026

OpenAI Deleted ‘Safely.’ NVIDIA Reports. Karpathy Is Still Learning

THE NUMBER: 6 — times OpenAI changed its mission in 9 years. The most recent edit deleted one word: safely. TL;DR Andrej Karpathy — the engineer who wrote the curriculum that trained a generation of developers, ran AI at Tesla, and helped found OpenAI — posted in December that he's never felt so behind as a programmer. Fourteen million people saw it. Tonight, NVIDIA reports Q4 fiscal 2026 earnings after market close: analysts expect $65.7 billion in revenue, up 67% year over year. The numbers will almost certainly land. What matters is what Jensen Huang says about the next two quarters to...

Feb 23, 2026

Altman lied about a handshake on camera. CrowdStrike fell 8%. Google just killed the $3,000 photo shoot.

Sam Altman told reporters he was "confused" when Narendra Modi grabbed his hand at the India AI Impact Summit. He said he "wasn't sure what was happening." The video, which has been watched by tens of millions of people, shows Altman looking directly at Dario Amodei before raising his fist. He knew exactly what was happening. He chose not to do it, and then he lied about it. On camera. In multiple interviews. With the footage playing on every screen behind him. That would be a minor character note in any other industry. In this one, it isn't. Because on...

Feb 20, 2026

We’re Building the Agentic Web Faster Than We’re Protecting It

Google's WebMCP gives agents structured access to every website. Anthropic's data shows autonomy doubling with oversight thinning. OpenAI's agent already drains crypto vaults. Google shipped working code Thursday that hands AI agents a structured key to every website on the internet. WebMCP, running in Chrome 146 Canary, lets sites expose machine-readable "Tool Contracts" so agents can book a flight, file a support ticket, or complete a checkout without parsing screenshots or scraping HTML. Early benchmarks show 67% less compute overhead than visual approaches. Microsoft co-authored the spec. The W3C is incubating it. This isn't a proposal. It's production software already...