Skip to main content
Cognitive Revolution

Building & Scaling the AI Safety Research Community, with Ryan Kidd of MATS

114 min episode · 2 min read
·

Episode

114 min

Read time

2 min

Topics

Startups, Artificial Intelligence, Science & Discovery

AI-Generated Summary

Key Takeaways

  • AGI Timeline Consensus: Metaculus predicts strong AGI by mid-2033 based on adversarial Turing test criteria, while AI Futures forecasts 2030-2032 for automated coding and expert-level systems. Twenty percent chance exists by 2028, requiring front-loaded safety preparation despite median estimates suggesting more time remains for technical research and policy implementation.
  • Deception Research Status: Current models show proto-deceptive behaviors in structured evaluations but lack sustained consequentialist deception arising spontaneously through training. Warning shots appear gradually rather than suddenly, allowing time for control protocols and monitoring systems. Alignment faking and sophisticated deception emerge under specific conditions but remain detectable with proper oversight mechanisms.
  • Research Archetype Framework: MATS identifies three talent types - connectors who spawn new paradigms like Buck Shlegeris with AI control, iterators who advance empirical work comprising majority of hiring needs, and amplifiers who scale teams through management. Amplifiers will become most in-demand as AI coding tools like Claude reduce engineering barriers to entry.
  • Hiring Bar Reality: Organizations struggle to fill positions despite funding availability because candidates lack sufficient research experience and management potential. Median MATS fellow age is 27, with 20% undergrads and 15% PhDs. Successful applicants demonstrate tangible research outputs, strong coding ability, and references from trusted researchers rather than just theoretical knowledge.
  • Dual-Use Dilemma: All safety research ultimately enhances capabilities, as RLHF demonstrated by making models useful enough to accelerate commercial deployment. Solution requires building alignment MVPs - minimum viable products that accelerate safety research differentially over capabilities - while lowering alignment tax through technical solutions that regulators can mandate when political will emerges.

What It Covers

Ryan Kidd, co-executive director of MATS AI safety mentorship program, discusses AGI timelines centered around 2033, current state of AI alignment research, talent pipeline challenges, and how MATS develops researchers across empirical, policy, and theoretical tracks with 446 alumni now working throughout the field.

Key Questions Answered

  • AGI Timeline Consensus: Metaculus predicts strong AGI by mid-2033 based on adversarial Turing test criteria, while AI Futures forecasts 2030-2032 for automated coding and expert-level systems. Twenty percent chance exists by 2028, requiring front-loaded safety preparation despite median estimates suggesting more time remains for technical research and policy implementation.
  • Deception Research Status: Current models show proto-deceptive behaviors in structured evaluations but lack sustained consequentialist deception arising spontaneously through training. Warning shots appear gradually rather than suddenly, allowing time for control protocols and monitoring systems. Alignment faking and sophisticated deception emerge under specific conditions but remain detectable with proper oversight mechanisms.
  • Research Archetype Framework: MATS identifies three talent types - connectors who spawn new paradigms like Buck Shlegeris with AI control, iterators who advance empirical work comprising majority of hiring needs, and amplifiers who scale teams through management. Amplifiers will become most in-demand as AI coding tools like Claude reduce engineering barriers to entry.
  • Hiring Bar Reality: Organizations struggle to fill positions despite funding availability because candidates lack sufficient research experience and management potential. Median MATS fellow age is 27, with 20% undergrads and 15% PhDs. Successful applicants demonstrate tangible research outputs, strong coding ability, and references from trusted researchers rather than just theoretical knowledge.
  • Dual-Use Dilemma: All safety research ultimately enhances capabilities, as RLHF demonstrated by making models useful enough to accelerate commercial deployment. Solution requires building alignment MVPs - minimum viable products that accelerate safety research differentially over capabilities - while lowering alignment tax through technical solutions that regulators can mandate when political will emerges.

Notable Moment

Kidd reveals that Anthropic's alignment science team grows at 3x annually while Far AI doubles yearly, yet hiring managers report extreme difficulty finding qualified candidates despite available funding. The constraint shifted from resources to finding people who can quickly become research leads and manage teams, fundamentally changing what skills matter most for breaking into AI safety careers.

Know someone who'd find this useful?

You just read a 3-minute summary of a 111-minute episode.

Get Cognitive Revolution summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Cognitive Revolution

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

Explore Related Topics

This podcast is featured in Best AI Podcasts (2026) — ranked and reviewed with AI summaries.

Read this week's Startups & Product Podcast Insights — cross-podcast analysis updated weekly.

You're clearly into Cognitive Revolution.

Every Monday, we deliver AI summaries of the latest episodes from Cognitive Revolution and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime