→ WHAT IT COVERS Swyx (Latent Space) and Jacob Efron (Redpoint/Unsupervised Learning) conduct their annual crossover episode covering the 2026 AI coding wars, agent infrastructure stability, foundation model competition, open-source model adoption shifts, and the emerging "dark factory" paradigm of zero-human-review software development. → KEY INSIGHTS - **AI Coding Market Scale:** Anthropic generates roughly $2.
Recent Episode Summaries
20 AI-powered summaries available
→ WHAT IT COVERS Shopify CTO Mikhail Parakhin details the company's AI adoption explosion in 2026, covering internal tooling including Tangle (ML experiment orchestration), Tangent (auto-research loops), and SimGym (customer behavior simulation), alongside infrastructure decisions around token budgets, PR review bottlenecks, and Liquid AI model deployment for sub-30ms search latency.
→ WHAT IT COVERS Noetik co-founders Ron Alfa and Daniel Bear explain how 90-95% of cancer drug trial failures stem from poor patient selection rather than bad pharmacology. They describe building multimodal foundation models trained on spatially-resolved human tumor data — combining H&E pathology, multiplex protein imaging, and 20,000-gene spatial transcriptomics — to match drugs to the right patient subpopulations.
→ WHAT IT COVERS Simon Last and Sarah Sachs from Notion detail five rebuilds of their AI agent system since 2022, covering the technical evolution from custom XML tool-calling to 100+ progressive disclosure tools, their MCP versus CLI tradeoffs, software factory vision, model behavior engineering as a distinct career path, and usage-based credit pricing for enterprise agentic workflows.
→ WHAT IT COVERS Ryan Lopopolo from OpenAI's Frontier team describes building a 1M+ line Electron application over five months with zero human-written code, deploying 1B tokens daily through a fully autonomous multi-agent pipeline. The episode covers harness engineering principles, the Symphony orchestration system built in Elixir, and how small teams can eliminate human bottlenecks from the software development lifecycle.
→ WHAT IT COVERS Marc Andreessen joins Latent Space to argue that current AI represents an 80-year overnight success, built on foundational research dating to 1943. He covers why this cycle differs from previous AI winters, the architectural significance of Pi and OpenClaw for agents, the death of the browser, crypto-AI convergence, and proof-of-human identity systems.
→ WHAT IT COVERS Moon Lake founders Fan-yun Sun and Chris Manning explain why causal world models require symbolic abstraction rather than pure pixel-level video generation. They contrast their multimodal reasoning approach against diffusion-based video models like Sora, arguing that action-conditioned interactivity and structured semantic representations are prerequisites for spatial intelligence and embodied AI applications.
→ WHAT IT COVERS Mistral releases Voxtral TTS, a 3B-parameter text-to-speech model supporting nine languages, built on a novel autoregressive flow matching architecture with an in-house neural audio codec. Guillaume Lample and Pavan Kumar Reddy also cover Mistral Small, the Forge deployment platform, and the LeanStral formal math reasoning project. → KEY INSIGHTS - **Autoregressive Flow Matching for TTS:** Voxtral TTS uses a flow matching head attached to a 3B Ministral backbone, processing...
→ WHAT IT COVERS MIT chemical engineering professor Heather Kulik explains why materials science lacks an AlphaFold equivalent, covering active learning for multi-objective optimization, LLM limitations in molecular design, the gap between ML potentials and experimental ground truth, and how academic researchers can differentiate from well-resourced industry labs.
→ WHAT IT COVERS David Singleton, former Stripe CTO and Dreamer co-founder, presents Dreamer—a consumer-facing platform where non-technical users build, discover, and deploy AI agents through a personal sidekick. The platform functions as an agent OS, combining a curated tool marketplace, hosted infrastructure, and a TypeScript SDK, with a 17-person team building the entire stack.
→ WHAT IT COVERS Felix Rieseberg, engineer at Anthropic, explains how Claude Cowork evolved from Claude Code into a VM-based knowledge work tool for non-terminal users. The conversation covers architecture decisions around local versus cloud compute, skills portability, sandbox security tradeoffs, and how Anthropic evaluates agent products against knowledge work tasks rather than coding benchmarks.
→ WHAT IT COVERS Simon Eski, founder of Turbopuffer, explains how his search database achieves cost reductions of up to 95% by building entirely on object storage and NVMe SSDs — an architecture made possible only after S3 gained strong consistency in December 2020 and compare-and-swap support in late 2024 — while serving customers like Cursor and Notion at billion-vector scale.
→ WHAT IT COVERS Nader Khalil (Brev/NVIDIA) and Kyle Kranen (Dynamo/NVIDIA) cover the acquisition of Brev by NVIDIA, the architecture of Dynamo's data center-scale inference engine, prefill-decode disaggregation, KV cache optimization, agent security constraints, and the trajectory of long-running autonomous agents in production environments. → KEY INSIGHTS - **Agent Security Constraint:** Agents capable of three actions — file access, internet access, and code execution — should only ever be...
→ WHAT IT COVERS Cursor's Cloud Agents launch gives AI a full persistent Linux VM with computer-use capabilities, enabling agents to install dependencies, run dev servers, reproduce bugs, record demo videos, and test changes end-to-end before returning a PR — shifting developer workflow from line-by-line editing toward high-level task delegation across parallel agent threads.
→ WHAT IT COVERS Box CEO Aaron Levie joins Latent Space with Chroma CEO Jeff Huber to examine why enterprise AI agent deployment lags behind coding agents, covering data governance, agent identity management, access control architecture, context engineering challenges, and why Fortune 500 companies face a multi-year transformation timeline before realizing compounding productivity returns from autonomous agents.
→ WHAT IT COVERS METR researcher Joel Becker explains how the organization evaluates AI capabilities using time horizon benchmarks, discusses the developer productivity RCT findings, examines why current models like GPT-5 are not yet catastrophically dangerous, and explores what conditions would signal a genuine AI capabilities explosion requiring serious concern.
→ WHAT IT COVERS Nathan Lambert, Sebastian Raschka, and Swyx analyze two converging AI stories: Anthropic's public accusation that Chinese labs — primarily MiniMax and DeepSeek — used distributed API accounts to extract training data, and OpenAI's formal deprecation of SWE-Bench Verified after discovering 59 unsolvable tasks and model memorization of benchmark solutions.
→ WHAT IT COVERS Prof. Max Welling, co-founder of CuspAI, explains how his company uses AI-driven platforms to search the entire space of possible materials — not just known ones — to accelerate solutions for climate change, carbon capture, and the energy transition, combining generative models, digital twins, and high-throughput experimentation. → KEY INSIGHTS - **Materials as the foundation layer:** Every technology stack ultimately depends on physical materials — GPUs require novel etching...
→ WHAT IT COVERS SemiAnalysis co-founder Doug O'Laughlin joins Latent Space to detail how Claude Code transformed his firm's research workflow, tracking AI-generated GitHub commits to quantify adoption, analyzing historical semiconductor memory cycles, and arguing that Claude Code 4.5's ability to one-shot complex multi-step tasks marks a genuine capability threshold for white-collar knowledge work automation.
→ WHAT IT COVERS OpenAI's Mia Glaese and Olivia Watkins explain why SWE-Bench Verified, the dominant coding benchmark since mid-2024, is now saturated and contaminated, why the field should migrate to SWE-Bench Pro, and what next-generation agentic coding evaluations need to measure. → KEY INSIGHTS - **Benchmark Saturation Signal:** When frontier models cluster around 80%+ on SWE-Bench Verified and gains shrink to 0.
Monday morning, inbox, done.
Pick your shows, and start the week knowing what happened in your world.
Pick the Podcasts You Care About
Choose from 200+ curated shows or add any public RSS feed.
AI Reads Every New Episode
Key arguments, surprising data points, and frameworks worth stealing — pulled automatically.
One Email, Every Monday
A curated brief for each episode, with links to listen if something grabs you.
Similar Podcasts You'll Love
Explore More
Get a free sample digest
See what your Monday email looks like — real AI summaries, no account needed.
One free sample — no spam, no commitment.




