Skip to main content
Latent Space

Why Anthropic Thinks AI Should Have Its Own Computer — Felix Rieseberg of Claude Cowork & Claude Code Desktop

86 min episode · 3 min read
·

Episode

86 min

Read time

3 min

Topics

Artificial Intelligence

AI-Generated Summary

Key Takeaways

  • VM-first architecture: Claude Cowork runs Claude Code inside a lightweight Linux virtual machine rather than directly on the host machine. This lets Claude install Python, Node.js, and any other dependency without requiring user approval or IT permission, while network ingress/egress controls remain strict. The VM collapses empty disk space so actual storage use is lower than macOS reports, though startup latency remains a real tradeoff versus running Claude Code natively.
  • Skills as portable markdown files: Cowork skills are plain markdown text files, not proprietary database entries. This makes them inherently portable — users can store them in any folder, sync via GitHub repos, or install plugin marketplaces by pointing to a GitHub URL. The unresolved gap is interpolating personal variables (preferred airports, phone numbers, folder paths) into otherwise shareable skill templates, which Rieseberg identifies as an unsolved industry problem worth building toward.
  • Eval against knowledge work, not coding benchmarks: Anthropic evaluates Cowork against knowledge work tasks — personal finance, mortgage management, legal office workflows — rather than standard coding suites used for Claude Code. The system prompt is tuned weekly based on these evals, steering Claude toward heavier use of the planning tool and ask-user-question tool for longer-horizon tasks where ambiguity is higher and wrong outputs are more costly to reverse.
  • Scaffolding has diminishing returns as models improve: Rieseberg argues that investing heavily in task-specific scaffolding carries risk because each model generation reduces the gap between raw capability and scaffolded output. His current preference is to maximize tool access and safety boundaries rather than build elaborate correction layers, then let model improvements handle the rest. He observes this pattern already in MCP servers being partially superseded by skills as models generalize better.
  • Local compute retains strategic value: Rieseberg pushes back on cloud-first assumptions by noting that moving all agent work to the cloud requires solving authentication, data privacy, and permission synchronization simultaneously. A concrete example: reading Chrome cookies to enable cloud-based browser sessions would trigger bank account lockouts from location anomalies. Keeping Claude on the local machine sidesteps these problems while giving it access to the same file system, applications, and credentials the user already has.

What It Covers

Felix Rieseberg, engineer at Anthropic, explains how Claude Cowork evolved from Claude Code into a VM-based knowledge work tool for non-terminal users. The conversation covers architecture decisions around local versus cloud compute, skills portability, sandbox security tradeoffs, and how Anthropic evaluates agent products against knowledge work tasks rather than coding benchmarks.

Key Questions Answered

  • VM-first architecture: Claude Cowork runs Claude Code inside a lightweight Linux virtual machine rather than directly on the host machine. This lets Claude install Python, Node.js, and any other dependency without requiring user approval or IT permission, while network ingress/egress controls remain strict. The VM collapses empty disk space so actual storage use is lower than macOS reports, though startup latency remains a real tradeoff versus running Claude Code natively.
  • Skills as portable markdown files: Cowork skills are plain markdown text files, not proprietary database entries. This makes them inherently portable — users can store them in any folder, sync via GitHub repos, or install plugin marketplaces by pointing to a GitHub URL. The unresolved gap is interpolating personal variables (preferred airports, phone numbers, folder paths) into otherwise shareable skill templates, which Rieseberg identifies as an unsolved industry problem worth building toward.
  • Eval against knowledge work, not coding benchmarks: Anthropic evaluates Cowork against knowledge work tasks — personal finance, mortgage management, legal office workflows — rather than standard coding suites used for Claude Code. The system prompt is tuned weekly based on these evals, steering Claude toward heavier use of the planning tool and ask-user-question tool for longer-horizon tasks where ambiguity is higher and wrong outputs are more costly to reverse.
  • Scaffolding has diminishing returns as models improve: Rieseberg argues that investing heavily in task-specific scaffolding carries risk because each model generation reduces the gap between raw capability and scaffolded output. His current preference is to maximize tool access and safety boundaries rather than build elaborate correction layers, then let model improvements handle the rest. He observes this pattern already in MCP servers being partially superseded by skills as models generalize better.
  • Local compute retains strategic value: Rieseberg pushes back on cloud-first assumptions by noting that moving all agent work to the cloud requires solving authentication, data privacy, and permission synchronization simultaneously. A concrete example: reading Chrome cookies to enable cloud-based browser sessions would trigger bank account lockouts from location anomalies. Keeping Claude on the local machine sidesteps these problems while giving it access to the same file system, applications, and credentials the user already has.
  • Build all candidates, then pick: Anthropic's current product development process skips traditional spec-then-build cycles in favor of rapidly constructing multiple prototypes simultaneously and selecting the best performer with a small focus group. Rieseberg credits this shift — enabled by cheap execution — with Cowork's ten-day build timeline, while clarifying that existing primitives like Claude Code, the VM framework, and internal prototypes accumulated over eighteen months provided the actual foundation.

Notable Moment

Rieseberg describes a workflow where Cowork autonomously scans an internal crash dashboard each morning, separates fixable bugs from OS-level failures, writes a markdown prompt file for each fixable bug, then spawns a separate Claude Code remote instance per bug — effectively building a self-directed multi-agent repair loop without using Claude's native sub-agents feature.

Know someone who'd find this useful?

You just read a 3-minute summary of a 83-minute episode.

Get Latent Space summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Latent Space

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

Explore Related Topics

This podcast is featured in Best AI Podcasts (2026) — ranked and reviewed with AI summaries.

Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.

You're clearly into Latent Space.

Every Monday, we deliver AI summaries of the latest episodes from Latent Space and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime