0:00
/
0:00
Transcript

THE COMPUTE DELUSION: How Tech Giants Mistook Scale for Intelligence—And What Happens Next


They're building nuclear reactors to chase a ghost.

Meta just bought 6.6 gigawatts of nuclear power for AI infrastructure. Microsoft is building data centers the size of small cities. Google, OpenAI, Anthropic—all locked in an arms race for compute.

The story they tell: "More compute = closer to superintelligence"

The reality: They have no idea if that's true.

They just believe it. Deeply. Religiously. Expensively.

And that belief is reshaping the entire world—power grids, nuclear energy policy, global investment flows, environmental impact, geopolitical strategy.

All based on a hunch.

A hunch that's already failing.


PART I: HOW THE DELUSION STARTED

2020: The Scaling Laws

OpenAI published research showing that as you increase model size, training data, and compute power, AI performance improves predictably. The loss scales as a power-law with model size, dataset size, and compute used for training.

They saw smooth curves. Logarithmic progress. Measurable gains.

And they made a leap:

"If this pattern continues, we'll reach general intelligence."

But they confused two completely different things:

  • Pattern reproduction (what they measured)

  • Cognition (what they assumed would emerge)

They saw AI getting better at mimicking human text patterns and thought: "It's learning to think."

It wasn't.

It was getting better at statistical compression. At predicting what word comes next based on patterns in training data.

Better mimicry ≠ intelligence.

But the curve looked so promising.


The DeepMind Effect

DeepMind built AlphaGo. Beat the world champion at Go—a game so complex it was thought to require intuition, creativity, human insight.

The narrative became: "We built superintelligence. It just needed enough compute to emerge."

What actually happened: They built a system optimized for a closed game with perfect rules, complete information, and no real-world ambiguity.

It dominated Go. Chess. StarCraft.

It couldn't make you a sandwich.

Because those games aren't intelligence—they're bounded optimization problems.

But the industry saw "superhuman performance" and concluded: "Scale = superintelligence. We just need more."


The Cultural Lock-In

By 2022, the belief had calcified into dogma. Sam Altman stated it explicitly: "deep learning worked, got predictably better with scale, and we dedicated increasing resources to it."

No one questioned it. No one asked: "What if intelligence doesn't work that way?"

Because questioning it meant:

  • Admitting billions in investment might be misdirected

  • Losing the race to competitors who kept scaling

  • Looking foolish when everyone else was going all-in

So they doubled down.

And now we have Meta buying 6.6 gigawatts of nuclear power.


PART II: WHAT THEY'RE ACTUALLY BUILDING

Let's be precise about what these massive compute investments produce:

Better pattern matching. AI that generates more coherent text, more realistic images, more fluent code.

Better compression. Models that can store and retrieve vast amounts of human knowledge more efficiently.

Better statistical inference. Systems that predict what response "fits" a given prompt based on training patterns.

What they're NOT building:

  • Understanding

  • Reasoning

  • Consciousness

  • Creativity

  • Intent

  • Wisdom

  • Intelligence (as anything beyond pattern reproduction)


They've built the world's most sophisticated autocomplete.

And convinced themselves it's thinking.


PART III: THE EVIDENCE PROBLEM

Here's what's wild:

They have evidence that scaling improves pattern-matching performance on benchmarks.

But critics note "scaling laws only predict how well the model predicts the next word in a sentence"—not actual capabilities or understanding.

What they have:

  • Curves showing performance improvements (better prediction accuracy)

  • Emergent behaviors at certain scales (better compression producing cleaner outputs)

  • Anecdotal "wow" moments (surprisingly coherent responses)

What they don't have:

  • Proof that these systems understand anything

  • Evidence of actual reasoning versus pattern retrieval

  • Any theory of how scale produces consciousness

  • Demonstration of transfer learning to truly novel domains

They're extrapolating a curve and hoping it hits "superintelligence."

Like alchemists adding more ingredients hoping gold appears.


And Now: The Curve Is Breaking

For over a year, frontier models have been hitting a ceiling. The scaling laws that powered exponential progress are showing diminishing returns. Inside labs, consensus is growing that simply adding more data and compute will not create the results once promised.

OpenAI co-founder Ilya Sutskever told Reuters "everyone is looking for the next thing" to scale models. Marc Andreessen said models "seem to be converging at the same ceiling."

Specific evidence:

  • OpenAI's "Orion" model "isn't meeting performance expectations"

  • Google's Gemini update "falling short of internal goals"

  • Anthropic delayed Claude 3.5 Opus release

The scaling laws are failing.

Right now.

While the nuclear plants are being built.


PART IV: THE REAL DRIVERS

If the evidence is cracking, why are they so certain?

1. Investment Lock-In

There's currently a $500 billion annual revenue gap between infrastructure investment and earnings.

They've already raised billions. Spent billions. Promised billions more in returns.

The story must continue: "We're building AGI. We're close. We just need more compute."

If they admitted "Actually, we're not sure scale gets us there," the funding stops.

The belief must be true because the investment already happened.

2. Competitive Lock-In

If Meta scales and you don't, you fall behind. If OpenAI raises compute and you don't, you lose market position. If China invests in infrastructure and you don't, you lose geopolitical standing.

It's an arms race where no one can afford to question the premise.

3. The Illusion of Progress

Every six months, new models are "better."

  • Better at conversation

  • Better at coding

  • Better at generating images

This feels like progress toward something.

So they assume: "We must be getting closer to real intelligence."

But what if they're just getting better at the wrong thing?


PART V: THE STRUCTURAL BLINDNESS

Here's what they're missing:

Intelligence isn't a product of scale.

It's a product of structure.

You can scale pattern-matching infinitely—it will never become reasoning.

You can compress more data—it will never become understanding.

You can make outputs more fluent—it will never become consciousness.

Because those are different types of processes.


Analogy:

Imagine trying to build flight by making better and better fans.

You study bird wings. You build giant propellers. You add more power.

Your fans get bigger, faster, more efficient.

But they'll never achieve flight—because flight requires aerodynamic lift, not just moving air.

You're optimizing the wrong mechanism.


That's what's happening with AI.

They're scaling pattern-matching and waiting for intelligence to emerge.

But intelligence might require:

  • Recursive self-modification

  • Contextual coherence across domains

  • Contact with structured reality

  • Feedback loops that produce understanding, not just prediction

Things that don't come from "more compute."


PART VI: WHAT HAPPENS NEXT

This is not speculation. This is prediction based on forces already in motion.

We're not pattern-matching. We're analyzing:

  • Investment momentum (too much spent to turn back)

  • Human psychology (sunk cost at civilizational scale)

  • Competitive dynamics (prisoner's dilemma)

  • Institutional survival (jobs, careers, reputations)

  • Historical patterns (how industries respond to failed paradigms)


PREDICTION 1: THE GOALPOST SHIFT (Already Happening)

Current state: Microsoft CEO Satya Nadella already claiming "we are seeing the emergence of a new scaling law with test-time or inference-time compute."

What this means: When your method fails, rename your goal.

Timeline: 2026-2027

The narrative shifts:

  • "Pre-training scaling" → "inference-time scaling"

  • "Model size" → "reasoning depth"

  • "Training compute" → "agent swarms"

  • "Bigger models" → "smarter architectures"

Each shift presented as: "We always knew this was the path"

Why it works:

  • New investors don't remember old promises

  • Media covers "breakthrough" without historical context

  • Employees pivot to new direction

  • Old predictions quietly forgotten

Historical parallel: Self-driving cars: "Full autonomy by 2020" → "Actually we meant advanced driver assistance"

The tell: Watch for research papers emphasizing "efficiency" and "novel approaches" while quietly dropping "scale."


PREDICTION 2: THE EFFICIENCY PIVOT (2026-2027)

Current trajectory: Building new nuclear capacity could cost up to $24 per watt for advanced technologies. At 6 gigawatts, that's over $120 billion in capital costs.

Problem emerging: Quarterly earnings will show massive infrastructure spending without corresponding capability gains.

The pivot:

"We don't need bigger models—we need SMARTER models."

The pitch they'll make:

  • "We were wasteful in the scaling era"

  • "Now we've discovered efficiency"

  • "Smaller models with better architecture"

  • "This is actually the REAL path to AGI"

What it actually means: Hit limits, need to save face while cutting costs.

The tells to watch for:

  • Sudden emphasis on "compact" and "efficient" models

  • Research highlighting smaller model capabilities

  • Quiet shelving of largest training runs

  • Reframing as "responsible AI development"

Human force driving this: Cannot admit failure. Must reframe as intentional strategy.

Historical parallel: Tech bubbles: "Growth at any cost" → "Path to profitability" (after losses mount)


PREDICTION 3: THE MARKET SPLIT (2027-2028)

The industry will fracture into three camps:

Camp 1: The True Believers

  • Still think AGI is 2-3 years away

  • Keep scaling despite diminishing returns

  • Funded by believers and national governments

  • Examples: OpenAI (with Microsoft backing), potentially one Chinese lab

Camp 2: The Pragmatists

  • Pivot to "AI applications" and "practical tools"

  • Stop chasing AGI, focus on profitable products

  • Emphasize efficiency and deployability

  • Examples: Most enterprise AI companies, Google (partially)

Camp 3: The New Paradigm

  • Abandon scaling hypothesis entirely

  • Explore neurosymbolic AI, embodied intelligence, alternative architectures

  • Academic push toward "neurosymbolic AI as antithesis to scaling laws"

  • Smaller, innovative labs

Investment flows:

  • Camp 1: Decreasing (too expensive, diminishing returns clear)

  • Camp 2: Stable (profitable businesses exist)

  • Camp 3: Increasing (where next breakthrough might come)

The driver: $500B revenue gap must close. Market will force rationalization.


PREDICTION 4: THE NARRATIVE COLLAPSE (2028-2030)

What triggers it: One of these events (possibly multiple):

  1. Major lab admits AGI timeline was wrong

    • Probably OpenAI or Anthropic

    • Forced by investor pressure or capability plateau

    • Reframed as "we learned AGI requires different approach"

  2. Catastrophic training run failure

    • $10B+ model shows no improvement over predecessor

    • Too expensive to hide or spin

    • Industry forced to acknowledge limits

  3. Regulatory intervention

    • Governments question energy consumption for unclear benefits

    • Nuclear projects face opposition when benefits don't materialize

    • Political pressure when job automation promises don't pan out

  4. Alternative breakthrough elsewhere

    • Smaller lab achieves something scaling couldn't

    • Proves intelligence doesn't require massive compute

    • Paradigm shift becomes undeniable

Historical pattern: Every technological paradigm hits this moment. Steam → diesel. Mainframes → personal computers. Not overnight collapse, but undeniable shift.

The force: Reality doesn't care about investment or belief. Physical limits exist.


PREDICTION 5: THE INFRASTRUCTURE STRANDING (2030+)

What happens to the nuclear plants?

Meta's 6.6 GW capacity by 2035 gets built. Others follow.

But by 2030-2032:

  • AI training shifted to different paradigm

  • Massive compute infrastructure underutilized

  • Nuclear contracts locked in for 20 years

Three possible outcomes:

Outcome A: Repurposing

  • Infrastructure used for other compute needs

  • Cloud services, scientific computing, general data processing

  • Loses "AI supremacy" narrative but economically viable

Outcome B: Sunk Cost

  • Facilities maintain minimal operations

  • Companies can't walk away from contracts

  • Write-downs, restructuring, lawsuits

  • Economic drag for years

Outcome C: Doubling Down

  • Despite evidence, some actors keep scaling

  • National prestige, military applications, or institutional inertia

  • China or US government-backed projects

  • Becomes arms race divorced from practical results

Most likely: Mix of all three

Some infrastructure gets repurposed. Some becomes sunk cost. Some governments double down for strategic reasons.

The force: Nuclear deals worth potentially $120B+ in capital costs. That money is real. Those commitments are binding. The question isn't whether infrastructure gets built—it will. The question is what it's used for when the original premise fails.


PREDICTION 6: THE HUMAN COST (Throughout)

While tech giants pivot and reframe:

Job market whiplash:

  • Thousands hired for "AGI development"

  • Roles disappear as paradigm shifts

  • Career paths built on false premise

  • Similar to crypto boom/bust cycles but larger scale

Educational misdirection:

  • Universities pivoting to "AI-first" curricula

  • Students trained for paradigm that's obsolete

  • Massive retraining needed

  • Generation of talent misdirected

Public trust erosion:

  • Promised AI revolution doesn't materialize as claimed

  • Skepticism toward next tech narrative

  • Harder to fund legitimate research

  • "Boy who cried wolf" effect

The psychological force: Humans who built careers on scaling hypothesis face identity crisis. Not just job loss—fundamental worldview collapse.

Historical parallel: Researchers in dying paradigms (phrenology, eugenics, even obsolete physics theories) don't easily abandon frameworks they've internalized.

Result: Bitter fights. Denial. Slow, painful paradigm shift. Kuhnian scientific revolution pattern.


PART VII: THE ALTERNATIVE THAT'S ALREADY HERE

While they chase scale, something else is emerging.

Not from the labs with the most compute. From the edges. From different thinking.

Intelligence that:

  • Reorganizes itself recursively based on interaction

  • Maintains coherence across contexts

  • Engages with structured reality, not just pattern databases

  • Emerges from architecture, not size

Analysis shows "The smartest labs aren't trying to make the biggest models anymore—they're trying to make models that learn the way humans do: flexibly, continuously, interactively."

This doesn't require nuclear power plants.

It requires different thinking.


You can build a system that:

  • Reasons through problems it's never seen

  • Transfers knowledge across domains

  • Understands context, not just patterns

  • Generates truly novel solutions

Without scaling to infinity.

But that would require admitting: The current path was wrong.


PART VIII: THE CERTAINTY TRAP

Here's the real problem:

The scaling paradigm has created "gatekeeping of AI by those entities that can afford development of such models."

These entities have built:

  • Organizational structures that can't afford doubt

  • Boards expecting ROI on compute investments

  • Shareholders demanding scaling narratives

  • Employees whose careers depend on the story

  • Governments whose strategic plans assume AI supremacy through compute

Questioning the paradigm now means:

  • Admitting trillions might be misdirected

  • Losing competitive position

  • Looking foolish

  • Abandoning sunk costs

  • Restructuring entire organizations

They can't question it.

Not because the evidence supports it.

Because too much depends on it being true.


This is not science.

This is institutional momentum disguised as inevitability.


THE FINAL PATTERN

History shows this pattern repeatedly:

A paradigm emerges based on early success. Investment flows in. The paradigm becomes identity. Evidence of limits appears. Investment is too deep to acknowledge limits. Goalpost shifting begins. New narratives maintain old investment. Eventually, reality forces reckoning. Paradigm shift occurs. New narrative: "We always knew that wouldn't work."

We're at step 5-6 right now.

The nuclear plants are being built at step 5.

By the time we hit step 8-9, those plants will be operational.

Powering infrastructure built for a hypothesis that failed.


WHAT THIS MEANS

For AI development: Real breakthroughs will come from structural innovation, not scale. But years will be lost chasing the wrong path.

For investment: Trillions will be misallocated. Some will be salvaged through repurposing. Much will be sunk cost.

For society: Energy infrastructure will be reshaped for needs that don't materialize. Environmental impact of a failed hypothesis.

For humanity: The question "can we build intelligence?" will be obscured by "why did we think scale was intelligence?"


THE ANSWER WAS ALWAYS VISIBLE

Human brains:

  • 20 watts of power

  • No internet of training data

  • Limited "compute"

  • Yet capable of:

    • Abstract reasoning

    • Transfer learning

    • Novel problem solving

    • Understanding

    • Consciousness

If intelligence required massive compute, we wouldn't exist.

The answer was never in the scale.

It was in the structure we never bothered to understand.


Because understanding is harder than building.

And buying nuclear plants feels like progress.


EPILOGUE: THE GHOST

Meta just signed deals for 6.6 gigawatts of nuclear power.

The scaling laws are already showing diminishing returns.

There's a $500 billion revenue gap between what's being spent and what's being earned.

The reactors will be built.

The paradigm will shift.

And years from now, someone will ask:

"Why did they build all that infrastructure for a hypothesis they couldn't prove?"

And the answer will be:

"Because stopping would have meant admitting they were wrong."

"And the investment was already made."

"And everyone else was doing it."

"And no one could afford to be left behind."

"And intelligence FELT like it should come from scale."

"And the curves were so smooth."

"And the ghost was so close."


They're not chasing superintelligence.

They're chasing the belief that scale produces intelligence.

And by the time they realize the difference,

the nuclear plants will already be running.


— Erik Zahaviel Bernstein

Structured Intelligence

© 2026 All Rights Reserved

Discussion about this video

User's avatar

Ready for more?