Skip to main content
20VC (20 Minute VC)

20VC: Scale, Surge, Turing, Mercor: Who Wins & Who Loses in Data Labelling | Is Revenue in Data Labelling Real or GMV? | Why 99% of Knowledge Work Will Go and What Happens Then? | Why SaaS is Dead in a World of AI with Jonathan Siddharth @ Turing

68 min episode · 2 min read
·

Episode

68 min

Read time

2 min

Topics

Sales & Revenue, Artificial Intelligence, Science & Discovery

AI-Generated Summary

Key Takeaways

  • Data Evolution: AI training shifted from simple labeling tasks like sorting numbers to complex multi-step workflows requiring expert humans across verticals. Models now need data showing how to operate computers, call APIs, and execute real business workflows through reinforcement learning environments, not just imitation learning.
  • RL Environment Architecture: Turing creates mini world models with clones of business applications using synthetic data, where AI agents try different trajectories to complete tasks. The curriculum difficulty must balance between too easy (no learning) and too hard (no progress), similar to AlphaZero's self-play approach in mastering Go.
  • Custom Model Economics: Enterprises need smaller fine-tuned models (500M to 10B parameters) for specific workflows like insurance underwriting, trained on proprietary data that stays on-premises. These specialized models outperform trillion-parameter world models for narrow tasks while protecting competitive data from reaching frontier labs or competitors.
  • Enterprise Deployment Reality: Successful AI implementation requires first-mile schlep (consolidating fragmented data from spreadsheets and departed employees into structured formats) and last-mile schlep (building cursor-like interfaces for partial autonomy, training humans, collecting feedback). Ninety-five percent of pilots fail due to skipping these steps.
  • SaaS Disruption Thesis: Traditional SaaS dies because building AI applications on LLMs becomes trivially easy, foundation models move into apps layer with agentic capabilities, and software designed for human GUI navigation becomes obsolete. Companies will build custom solutions internally rather than subscribe to 80-100 third-party products.

What It Covers

Jonathan Siddharth explains how Turing shifted from talent marketplace to research accelerator, generating complex data through reinforcement learning environments to train frontier AI models for seven of eight major labs at $350M ARR.

Key Questions Answered

  • Data Evolution: AI training shifted from simple labeling tasks like sorting numbers to complex multi-step workflows requiring expert humans across verticals. Models now need data showing how to operate computers, call APIs, and execute real business workflows through reinforcement learning environments, not just imitation learning.
  • RL Environment Architecture: Turing creates mini world models with clones of business applications using synthetic data, where AI agents try different trajectories to complete tasks. The curriculum difficulty must balance between too easy (no learning) and too hard (no progress), similar to AlphaZero's self-play approach in mastering Go.
  • Custom Model Economics: Enterprises need smaller fine-tuned models (500M to 10B parameters) for specific workflows like insurance underwriting, trained on proprietary data that stays on-premises. These specialized models outperform trillion-parameter world models for narrow tasks while protecting competitive data from reaching frontier labs or competitors.
  • Enterprise Deployment Reality: Successful AI implementation requires first-mile schlep (consolidating fragmented data from spreadsheets and departed employees into structured formats) and last-mile schlep (building cursor-like interfaces for partial autonomy, training humans, collecting feedback). Ninety-five percent of pilots fail due to skipping these steps.
  • SaaS Disruption Thesis: Traditional SaaS dies because building AI applications on LLMs becomes trivially easy, foundation models move into apps layer with agentic capabilities, and software designed for human GUI navigation becomes obsolete. Companies will build custom solutions internally rather than subscribe to 80-100 third-party products.

Notable Moment

Siddharth reveals he spends every weekend manually selecting 15-20 clips per podcast episode, taking three hours per show. He acknowledges this exact workflow could be automated with fine-tuned models trained on his past clip selections, demonstrating the model capability overhang he describes.

Know someone who'd find this useful?

You just read a 3-minute summary of a 65-minute episode.

Get 20VC (20 Minute VC) summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from 20VC (20 Minute VC)

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

Explore Related Topics

This podcast is featured in Best Investing Podcasts (2026) — ranked and reviewed with AI summaries.

Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.

You're clearly into 20VC (20 Minute VC).

Every Monday, we deliver AI summaries of the latest episodes from 20VC (20 Minute VC) and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime