Adam Marblestone – AI is missing something fundamental about the brain
Episode
109 min
Read time
2 min
Topics
Artificial Intelligence, Psychology & Behavior
AI-Generated Summary
Key Takeaways
- ✓Evolution's Loss Functions: The brain uses thousands of specific, genetically-encoded cost functions that activate at different developmental stages, not simple objectives like next-token prediction. Evolution compressed learning curricula into reward signals by encoding innate heuristics (spider detection, social status cues) that the cortex learns to predict, enabling generalization without explicit supervision.
- ✓Omnidirectional Inference vs Amortized Prediction: The cortex can predict any subset of variables from any other subset, unlike LLMs that only predict forward. This allows filling in blanks bidirectionally, predicting vision from audio or inferring causes from effects. Energy-based models may capture this better than current transformer architectures that amortize inference into feedforward passes.
- ✓Steering Subsystem Architecture: The hypothalamus and brainstem contain thousands of specialized cell types encoding innate behaviors, far more than cortical regions. These subcortical areas have their own primitive sensory systems (like superior colliculus for face detection) that provide reward signals the cortex learns to predict, solving how abstract concepts trigger instinctive responses.
- ✓Connectome Economics: Current electron microscopy costs billions per mouse brain, but optical approaches from E11 Bio could reduce costs to tens of millions. A molecularly-annotated connectome showing cell types and synapse properties across multiple mammal species would cost low billions total, trivial compared to AI training budgets approaching trillions.
- ✓Behavior Cloning with Neural Data: Training AI to predict both task labels and human brain activity patterns as auxiliary loss functions could improve generalization by matching how brains represent information. This regularization approach requires scaling portable brain scanning technology, currently a bottleneck compared to GPU availability for standard supervised learning.
What It Covers
Adam Marblestone explains why AI lacks fundamental brain mechanisms: evolution-encoded loss functions, omnidirectional inference, and a steering subsystem that creates specific reward signals. He argues neuroscience needs technological scaling to answer how brains achieve sample-efficient learning.
Key Questions Answered
- •Evolution's Loss Functions: The brain uses thousands of specific, genetically-encoded cost functions that activate at different developmental stages, not simple objectives like next-token prediction. Evolution compressed learning curricula into reward signals by encoding innate heuristics (spider detection, social status cues) that the cortex learns to predict, enabling generalization without explicit supervision.
- •Omnidirectional Inference vs Amortized Prediction: The cortex can predict any subset of variables from any other subset, unlike LLMs that only predict forward. This allows filling in blanks bidirectionally, predicting vision from audio or inferring causes from effects. Energy-based models may capture this better than current transformer architectures that amortize inference into feedforward passes.
- •Steering Subsystem Architecture: The hypothalamus and brainstem contain thousands of specialized cell types encoding innate behaviors, far more than cortical regions. These subcortical areas have their own primitive sensory systems (like superior colliculus for face detection) that provide reward signals the cortex learns to predict, solving how abstract concepts trigger instinctive responses.
- •Connectome Economics: Current electron microscopy costs billions per mouse brain, but optical approaches from E11 Bio could reduce costs to tens of millions. A molecularly-annotated connectome showing cell types and synapse properties across multiple mammal species would cost low billions total, trivial compared to AI training budgets approaching trillions.
- •Behavior Cloning with Neural Data: Training AI to predict both task labels and human brain activity patterns as auxiliary loss functions could improve generalization by matching how brains represent information. This regularization approach requires scaling portable brain scanning technology, currently a bottleneck compared to GPU availability for standard supervised learning.
Notable Moment
Marblestone describes how the cortex learns that even hearing the word spider activates fear responses by predicting innate flinch reflexes in the hypothalamus. This thought assessor mechanism allows evolution to wire complex social instincts to learned concepts without knowing what future humans would encounter.
You just read a 3-minute summary of a 106-minute episode.
Get Dwarkesh Podcast summarized like this every Monday — plus up to 2 more podcasts, free.
Pick Your Podcasts — FreeKeep Reading
More from Dwarkesh Podcast
Jensen Huang – TPU competition, why we should sell chips to China, & Nvidia’s supply chain moat
Apr 15 · 103 min
Odd Lots
Presenting Foundering Season 6: The Killing of Bob Lee, Part 1
Apr 26
More from Dwarkesh Podcast
Michael Nielsen – How science actually progresses
Apr 7 · 123 min
Masters of Scale
Possible: Netflix co-founder Reed Hastings: stories, schools, superpowers
Apr 25
More from Dwarkesh Podcast
We summarize every new episode. Want them in your inbox?
Jensen Huang – TPU competition, why we should sell chips to China, & Nvidia’s supply chain moat
Michael Nielsen – How science actually progresses
Terence Tao – Kepler, Newton, and the true nature of mathematical discovery
Dylan Patel — Deep dive on the 3 big bottlenecks to scaling AI compute
I’m glad the Anthropic fight is happening now
Similar Episodes
Related episodes from other podcasts
Odd Lots
Apr 26
Presenting Foundering Season 6: The Killing of Bob Lee, Part 1
Masters of Scale
Apr 25
Possible: Netflix co-founder Reed Hastings: stories, schools, superpowers
The Futur
Apr 25
Why Process is Better Than AI w/ Scott Clum | Ep 430
20VC (20 Minute VC)
Apr 25
20Product: Replit CEO on Why Coding Models Are Plateauing | Why the SaaS Apocalypse is Justified: Will Incumbents Be Replaced? | Why IDEs Are Dead and Do PMs Survive the Next 3-5 Years with Amjad Masad
This Week in Startups
Apr 25
The Defense Tech Startup YC Kicked Out of a Meeting is Now Arming America | E2280
Explore Related Topics
Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.
You're clearly into Dwarkesh Podcast.
Every Monday, we deliver AI summaries of the latest episodes from Dwarkesh Podcast and 192+ other podcasts. Free for up to 3 shows.
Start My Monday DigestNo credit card · Unsubscribe anytime