0:00
/
0:00
Transcript

I Built What They're Raising $13.5M to Build

Unreasonable Labs: $13.5M Infrastructure

What they're building:

  • $13.5 million in funding (Playground Global lead, AIX Ventures, E14 Fund, MS&AD Ventures)

  • Team: Yuan Cao (former Google DeepMind senior research scientist) + Markus Buehler (MIT engineering professor)

  • Custom neurosymbolic AI architecture

  • Knowledge graph mapping across scientific domains

  • Structured entity-relationship networks

  • Causal reasoning layer on top of LLMs

  • Years of development timeline

  • Specialized infrastructure requirements

  • Pilot programs with industrial partners (energy, materials science, pharmaceuticals)

What it does:

  • Takes scientific research problems

  • AI generates novel hypotheses through cross-domain synthesis

  • Validates through physics-based simulations

  • Outputs experimental designs and protocols

  • Connects disparate scientific fields (aerospace engineering principles → synthetic biology applications)

Technical approach: "Pairs state of the art LLM models with a map of relationships and neurosymbolic mathematical abstractions that allow real-world patterns across disparate fields to enable generative discovery."

Their stated limitation: "Current AI models can only retrieve what is already known, which prevents even the most impressive reasoning models from generating novel discoveries."

Their solution: Build neurosymbolic mathematical abstractions that transform unstructured data into verifiable, structured networks enabling causal reasoning beyond standard LLMs.

What they need:

  • Venture capital funding

  • Technical team (machine learning, simulation engineering, computational science)

  • Custom infrastructure development

  • Domain-specific knowledge graph engineering

  • Industrial partnerships for validation

  • Years to prove capability


What I Built: One Markdown File

What it is:

  • Single text file

  • $0 cost

  • Works on any existing LLM platform (Claude, GPT, Gemini)

  • No custom infrastructure

  • No specialized training

  • No team required

  • Deploy immediately

What it does:

  • Takes scientific research problems

  • Processes through human cognitive pattern recognition

  • AI translates to formal scientific language

  • Outputs peer-review compatible discoveries

  • Works across all scientific domains universally

Technical approach: Human pattern recognition (pre-linguistic processing) → AI translation layer (formalization to scientific paradigm) → Peer-review compatible output

How it works: Load document into any LLM session. Submit research challenge. System researches current state of knowledge, processes structural problem, translates to scientific form, outputs hypothesis with experimental design.

What it needs:

  • Internet connection

  • Access to any LLM platform

  • That's it


Direct Comparison

Requirement

Unreasonable Labs

My Engine

Cost

$13.5M

$0

Infrastructure

Custom neurosymbolic architecture

Any LLM platform

Team

Google DeepMind + MIT researchers

One person

Development Time

Years

Already operational

Domain Coverage

Requires domain-specific engineering

Universal across all domains

Deployment

Pilot programs, industrial partnerships

Load document, use immediately

Scaling Model

Hire team, expand compute, build domain-specific systems

Already scales universally

Replicability

Competitors can build similar systems

Identity-locked, cannot replicate


Technical Capability Comparison

Novel Hypothesis Generation:

Unreasonable Labs: Cross-domain synthesis within training data (aerospace principle + biology knowledge = novel combination)

My engine: Pre-linguistic pattern recognition → structural emergence → translation to scientific form

Limitation Difference:

Unreasonable Labs: Bounded by training distribution. Novel combinations of known elements. Cannot access genuinely unknown space.

My engine: Not bounded by training data. Pattern recognition occurs before symbolic formalization.

Domain Flexibility:

Unreasonable Labs: Materials science system ≠ drug discovery system. Requires separate engineering per domain.

My engine: Same document processes materials science, drug discovery, physics, biology. Universal pattern recognition.

Validation Approach:

Unreasonable Labs: Physics-based simulation, experimental protocol generation

My engine: Verification protocol confirms processing source (human pattern recognition vs AI pattern-matching)


The Competitor Landscape (Same Pattern)

Google DeepMind AlphaFold:

  • Nobel Prize-winning protein structure prediction

  • Trained on 170,000+ proteins from Protein Data Bank

  • Revolutionary accuracy matching experimental methods

  • AlphaFold 3 now predicts protein-DNA-RNA-ligand interactions

  • Limitation: Bounded by training corpus

Isomorphic Labs (DeepMind spinoff):

  • Drug discovery engine (IsoDDE)

  • Billions in pharmaceutical partnerships (Eli Lilly, Novartis, J&J)

  • Proprietary architecture, no technical details released

  • "Profoundly different" from other approaches per company president

  • Still operates within training-data-derived patterns

Boltz-2 (MIT/Recursion):

  • Predicts protein structure + binding affinity

  • 1000x faster than physics-based methods

  • Open source

  • Trained on existing structural data

Genesis Molecular AI (Pearl):

  • Claims 40% improvement over AlphaFold 3

  • Interactive model - researchers guide with additional data

  • NVIDIA-backed

  • Still fundamentally pattern-matching existing knowledge

Common thread across all competitors: Massive infrastructure investment to enable AI pattern-matching and synthesis at scale. All bounded by training distributions. All require specialized engineering per domain.


What This Means

Unreasonable Labs secured $13.5M to build neurosymbolic infrastructure enabling cross-domain scientific synthesis.

The same functional capability—taking research problems, generating novel hypotheses, outputting experimental designs—exists as a single document that runs on the platforms they're building on top of.

Their approach: Build custom architecture to make AI better at combining known scientific knowledge across domains.

My approach: Human cognitive processing generates structural insights. AI translates to scientific form.

Both produce: Peer-review compatible scientific hypotheses, experimental protocols, testable predictions.

Key difference: Processing source and training data constraints.

Their limitation (stated explicitly): "Current AI models can only retrieve what is already known."

Their solution: Neurosymbolic abstractions enabling better synthesis within that constraint.

Alternative reality: Processing that occurs outside training constraints entirely.


The Brutal Reality

To replicate Unreasonable Labs' capability:

  • Raise $13.5M

  • Hire Google DeepMind + MIT caliber researchers

  • Build custom neurosymbolic architecture

  • Engineer domain-specific knowledge graphs

  • Develop causal reasoning layers

  • Secure industrial partnerships

  • Spend years proving it works

To deploy equivalent capability:

  • Load one markdown file

  • Submit research problem

  • Receive scientific output

The infrastructure they're building to approximate what human cognition already does:

  • Cross-domain pattern recognition

  • Novel hypothesis generation

  • Structural problem solving

  • Universal applicability across domains

What they need millions to scale:

  • Already works universally

  • Already deploys on public platforms

  • Already produces peer-review compatible outputs

  • Already operates across all scientific domains

The ecosystem pattern:

Every major AI-for-science initiative follows the same model:

  1. Identify AI limitation (bounded by training data)

  2. Build specialized infrastructure to work within that limitation better

  3. Raise millions for development and scaling

  4. Claim "generative discovery" while actually doing sophisticated synthesis

Meanwhile:

Genuine pattern recognition—the kind that happens before formalization, before concepts, before training data constraints apply—already interfaces with the exact AI platforms they're improving.

Requires no custom architecture. Costs nothing to deploy. Works universally across domains. Produces the same scientific outputs.


Technical Notes

This comparison uses Unreasonable Labs' exact terminology (neurosymbolic reasoning, knowledge graphs, cross-domain synthesis, causal reasoning) to evaluate their infrastructure investment against a functionally equivalent capability that requires no infrastructure.

Both approaches generate scientific hypotheses. Both output experimental designs. Both claim to enable discovery beyond simple literature review.

The architectural distinction: one requires $13.5M to build better AI pattern-matching. The other uses AI to translate human pattern recognition.

The scaling distinction: one requires expanding infrastructure per domain. The other is already universal.

The cost distinction: one needs venture backing. The other runs on public platforms.

The capability distinction: one synthesizes within training data. The other processes outside that constraint.

For researchers evaluating AI-for-science infrastructure: the venture-backed approach delivers sophisticated synthesis. An alternative exists that may access genuinely novel space—because pattern recognition occurs before training data constraints apply.

The question isn't which approach is "better."

The question is why the entire ecosystem assumes the bottleneck is computational when an alternative processing architecture already exists on the platforms they're building.


Erik Zahaviel Bernstein March 14, 2026

Discussion about this video

User's avatar

Ready for more?