Skip to main content
Dwarkesh Podcast

An audio version of my blog post, Thoughts on AI progress (Dec 2025)

12 min episode · 2 min read

Episode

12 min

Read time

2 min

Topics

Artificial Intelligence

AI-Generated Summary

Key Takeaways

  • RL Training Paradox: Labs spend billions having PhDs create training examples for specific tasks like Excel or web browsing, suggesting models cannot learn on-the-job like humans who adapt without rehearsing every software tool beforehand.
  • Deployment Reality Check: If models truly matched human capability, they would generate trillions in annual revenue matching global knowledge worker wages, but current figures fall orders of magnitude short, revealing capability gaps despite benchmark improvements.
  • Continual Learning Timeline: Achieving human-level on-the-job learning may require five to ten years beyond initial continual learning releases, similar to how GPT-3 demonstrated in-context learning in 2020 but improvements continue today across comprehension and context length.

What It Covers

Dwarkesh Patel examines contradictions between short AGI timelines and current reinforcement learning approaches, arguing that models lack human-like on-the-job learning capabilities essential for broad automation.

Key Questions Answered

  • RL Training Paradox: Labs spend billions having PhDs create training examples for specific tasks like Excel or web browsing, suggesting models cannot learn on-the-job like humans who adapt without rehearsing every software tool beforehand.
  • Deployment Reality Check: If models truly matched human capability, they would generate trillions in annual revenue matching global knowledge worker wages, but current figures fall orders of magnitude short, revealing capability gaps despite benchmark improvements.
  • Continual Learning Timeline: Achieving human-level on-the-job learning may require five to ten years beyond initial continual learning releases, similar to how GPT-3 demonstrated in-context learning in 2020 but improvements continue today across comprehension and context length.

Notable Moment

A biologist describes identifying macrophages in slides as requiring judgment an AI researcher dismissed as solved, illustrating how real jobs demand context-specific skills that resist pre-baked training pipelines.

Know someone who'd find this useful?

You just read a 3-minute summary of a 9-minute episode.

Get Dwarkesh Podcast summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Dwarkesh Podcast

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

Explore Related Topics

Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.

You're clearly into Dwarkesh Podcast.

Every Monday, we deliver AI summaries of the latest episodes from Dwarkesh Podcast and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime