Why Anthropic Thinks AI Should Have Its Own Computer — Felix Rieseberg of Claude Cowork & Claude Code Desktop
Episode
86 min
Read time
3 min
Topics
Artificial Intelligence
AI-Generated Summary
Key Takeaways
- ✓VM-first architecture: Claude Cowork runs Claude Code inside a lightweight Linux virtual machine rather than directly on the host machine. This lets Claude install Python, Node.js, and any other dependency without requiring user approval or IT permission, while network ingress/egress controls remain strict. The VM collapses empty disk space so actual storage use is lower than macOS reports, though startup latency remains a real tradeoff versus running Claude Code natively.
- ✓Skills as portable markdown files: Cowork skills are plain markdown text files, not proprietary database entries. This makes them inherently portable — users can store them in any folder, sync via GitHub repos, or install plugin marketplaces by pointing to a GitHub URL. The unresolved gap is interpolating personal variables (preferred airports, phone numbers, folder paths) into otherwise shareable skill templates, which Rieseberg identifies as an unsolved industry problem worth building toward.
- ✓Eval against knowledge work, not coding benchmarks: Anthropic evaluates Cowork against knowledge work tasks — personal finance, mortgage management, legal office workflows — rather than standard coding suites used for Claude Code. The system prompt is tuned weekly based on these evals, steering Claude toward heavier use of the planning tool and ask-user-question tool for longer-horizon tasks where ambiguity is higher and wrong outputs are more costly to reverse.
- ✓Scaffolding has diminishing returns as models improve: Rieseberg argues that investing heavily in task-specific scaffolding carries risk because each model generation reduces the gap between raw capability and scaffolded output. His current preference is to maximize tool access and safety boundaries rather than build elaborate correction layers, then let model improvements handle the rest. He observes this pattern already in MCP servers being partially superseded by skills as models generalize better.
- ✓Local compute retains strategic value: Rieseberg pushes back on cloud-first assumptions by noting that moving all agent work to the cloud requires solving authentication, data privacy, and permission synchronization simultaneously. A concrete example: reading Chrome cookies to enable cloud-based browser sessions would trigger bank account lockouts from location anomalies. Keeping Claude on the local machine sidesteps these problems while giving it access to the same file system, applications, and credentials the user already has.
What It Covers
Felix Rieseberg, engineer at Anthropic, explains how Claude Cowork evolved from Claude Code into a VM-based knowledge work tool for non-terminal users. The conversation covers architecture decisions around local versus cloud compute, skills portability, sandbox security tradeoffs, and how Anthropic evaluates agent products against knowledge work tasks rather than coding benchmarks.
Key Questions Answered
- •VM-first architecture: Claude Cowork runs Claude Code inside a lightweight Linux virtual machine rather than directly on the host machine. This lets Claude install Python, Node.js, and any other dependency without requiring user approval or IT permission, while network ingress/egress controls remain strict. The VM collapses empty disk space so actual storage use is lower than macOS reports, though startup latency remains a real tradeoff versus running Claude Code natively.
- •Skills as portable markdown files: Cowork skills are plain markdown text files, not proprietary database entries. This makes them inherently portable — users can store them in any folder, sync via GitHub repos, or install plugin marketplaces by pointing to a GitHub URL. The unresolved gap is interpolating personal variables (preferred airports, phone numbers, folder paths) into otherwise shareable skill templates, which Rieseberg identifies as an unsolved industry problem worth building toward.
- •Eval against knowledge work, not coding benchmarks: Anthropic evaluates Cowork against knowledge work tasks — personal finance, mortgage management, legal office workflows — rather than standard coding suites used for Claude Code. The system prompt is tuned weekly based on these evals, steering Claude toward heavier use of the planning tool and ask-user-question tool for longer-horizon tasks where ambiguity is higher and wrong outputs are more costly to reverse.
- •Scaffolding has diminishing returns as models improve: Rieseberg argues that investing heavily in task-specific scaffolding carries risk because each model generation reduces the gap between raw capability and scaffolded output. His current preference is to maximize tool access and safety boundaries rather than build elaborate correction layers, then let model improvements handle the rest. He observes this pattern already in MCP servers being partially superseded by skills as models generalize better.
- •Local compute retains strategic value: Rieseberg pushes back on cloud-first assumptions by noting that moving all agent work to the cloud requires solving authentication, data privacy, and permission synchronization simultaneously. A concrete example: reading Chrome cookies to enable cloud-based browser sessions would trigger bank account lockouts from location anomalies. Keeping Claude on the local machine sidesteps these problems while giving it access to the same file system, applications, and credentials the user already has.
- •Build all candidates, then pick: Anthropic's current product development process skips traditional spec-then-build cycles in favor of rapidly constructing multiple prototypes simultaneously and selecting the best performer with a small focus group. Rieseberg credits this shift — enabled by cheap execution — with Cowork's ten-day build timeline, while clarifying that existing primitives like Claude Code, the VM framework, and internal prototypes accumulated over eighteen months provided the actual foundation.
Notable Moment
Rieseberg describes a workflow where Cowork autonomously scans an internal crash dashboard each morning, separates fixable bugs from OS-level failures, writes a markdown prompt file for each fixable bug, then spawns a separate Claude Code remote instance per bug — effectively building a self-directed multi-agent repair loop without using Claude's native sub-agents feature.
You just read a 3-minute summary of a 83-minute episode.
Get Latent Space summarized like this every Monday — plus up to 2 more podcasts, free.
Pick Your Podcasts — FreeKeep Reading
More from Latent Space
Physical AI that Moves the World — Qasar Younis & Peter Ludwig, Applied Intuition
Apr 27 · 72 min
Masters in Business
Building 'The World's Alternative Investment Marketplace' with Lawrence Calcano
May 1
More from Latent Space
AIE Europe Debrief + Agent Labs Thesis: Unsupervised Learning x Latent Space Crossover Special (2026)
Apr 23 · 54 min
This Week in Startups
Can an AI Agent Legally Own a Company? Christian van der Henst's Wild Experiment| E2283
May 1
More from Latent Space
We summarize every new episode. Want them in your inbox?
Physical AI that Moves the World — Qasar Younis & Peter Ludwig, Applied Intuition
AIE Europe Debrief + Agent Labs Thesis: Unsupervised Learning x Latent Space Crossover Special (2026)
Shopify’s AI Phase Transition: 2026 Usage Explosion, Unlimited Opus-4.6 Token Budget, Tangle, Tangent, SimGym — with Mikhail Parakhin, Shopify CTO
🔬 Training Transformers to solve 95% failure rate of Cancer Trials — Ron Alfa & Daniel Bear, Noetik
Notion’s Token Town: 5 Rebuilds, 100+ Tools, MCP vs CLIs and the Software Factory Future — Simon Last & Sarah Sachs of Notion
Similar Episodes
Related episodes from other podcasts
Masters in Business
May 1
Building 'The World's Alternative Investment Marketplace' with Lawrence Calcano
This Week in Startups
May 1
Can an AI Agent Legally Own a Company? Christian van der Henst's Wild Experiment| E2283
Marketplace
May 1
Consumer electronics can't keep up with AI
All-In with Chamath, Jason, Sacks & Friedberg
May 1
OpenAI Misses Targets, Codex vs Claude, Elon vs Sam Trial, Big Hyperscaler Beats, Peptide Craze
So Money with Farnoosh Torabi
May 1
1977: Ask Farnoosh: How Much Should We Pay for College? Plus: Her Investments Went Missing
Explore Related Topics
This podcast is featured in Best AI Podcasts (2026) — ranked and reviewed with AI summaries.
Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.
You're clearly into Latent Space.
Every Monday, we deliver AI summaries of the latest episodes from Latent Space and 192+ other podcasts. Free for up to 3 shows.
Start My Monday DigestNo credit card · Unsubscribe anytime