20VC: Cohere Founder on How Cohere Compete with OpenAI and Anthropic $BNs | Why Counties Should Fund Their Own Models & the Need for Model Sovereignty | How Sam Altman Has Done a Disservice to AI with Nick Frosst
Episode
67 min
Read time
2 min
Topics
Startups, Artificial Intelligence
AI-Generated Summary
Key Takeaways
- ✓Enterprise Model Differentiation: Cohere trains models specifically for enterprise tool use and business data integration rather than consumer engagement metrics, using synthetic data from fake companies, emails, and APIs to optimize workplace augmentation over conversational ability or entertainment value.
- ✓Scaling Law Limitations: Throwing more compute at models does not guarantee exponential progress, as evidenced by GPT-5's worse user experience with auto-selection delays. The industry still uses 2017 transformer architecture with minimal algorithmic changes, making data quality and product work more critical than raw compute power.
- ✓Efficient Model Training: Cohere trains models to fit on two GPUs, spending orders of magnitude less than competitors on foundational models. This efficiency addresses enterprise deployment bottlenecks where companies lack GPU access, making the sweet spot between performance, cost, and available infrastructure crucial for production deployment.
- ✓Benchmark Gaming Reality: Industry benchmarks like HellaSWAG and ARC AGI challenge do not reflect actual enterprise utility. Models can be trained specifically to perform well on benchmarks without improving real workplace value. Customer success depends on practical task completion, not leaderboard rankings or mathematical reasoning tests.
- ✓Forward-Deployed Engineering Value: Enterprise AI deployment requires forward-deployed engineers to customize models for specific business contexts, internal tools, and documentation. This approach is not poor technology but necessary infrastructure work, similar to how industrial revolution required labor policy alongside technological advancement to create sustainable productivity gains.
What It Covers
Nick Frosst, Cohere cofounder, discusses competing against OpenAI and Anthropic with enterprise-focused models, challenges Sam Altman's AGI predictions, explains why scaling laws have limits, and advocates for sovereign AI models and forward-deployed engineering approaches.
Key Questions Answered
- •Enterprise Model Differentiation: Cohere trains models specifically for enterprise tool use and business data integration rather than consumer engagement metrics, using synthetic data from fake companies, emails, and APIs to optimize workplace augmentation over conversational ability or entertainment value.
- •Scaling Law Limitations: Throwing more compute at models does not guarantee exponential progress, as evidenced by GPT-5's worse user experience with auto-selection delays. The industry still uses 2017 transformer architecture with minimal algorithmic changes, making data quality and product work more critical than raw compute power.
- •Efficient Model Training: Cohere trains models to fit on two GPUs, spending orders of magnitude less than competitors on foundational models. This efficiency addresses enterprise deployment bottlenecks where companies lack GPU access, making the sweet spot between performance, cost, and available infrastructure crucial for production deployment.
- •Benchmark Gaming Reality: Industry benchmarks like HellaSWAG and ARC AGI challenge do not reflect actual enterprise utility. Models can be trained specifically to perform well on benchmarks without improving real workplace value. Customer success depends on practical task completion, not leaderboard rankings or mathematical reasoning tests.
- •Forward-Deployed Engineering Value: Enterprise AI deployment requires forward-deployed engineers to customize models for specific business contexts, internal tools, and documentation. This approach is not poor technology but necessary infrastructure work, similar to how industrial revolution required labor policy alongside technological advancement to create sustainable productivity gains.
Notable Moment
Frosst directly criticizes Sam Altman for making obviously wrong predictions about AGI timelines and existential threats, arguing this world tour warning global leaders was academically disingenuous and damaged productive discourse about real AI risks like income inequality and workforce disruption rather than imagined digital gods.
You just read a 3-minute summary of a 64-minute episode.
Get 20VC (20 Minute VC) summarized like this every Monday — plus up to 2 more podcasts, free.
Pick Your Podcasts — FreeKeep Reading
More from 20VC (20 Minute VC)
20VC: Anthropic Raises $45BN but Falls Short on Compute | OpenAI Crushes with GPT5.5 and Codex: Back in the Game? | China Blocks Manus $2BN Deal to Meta | Thoma Bravo Hand Back Medallia Keys to Creditors | Why Google is a Bigger Buy Than Ever Before
Apr 30 · 85 min
The TWIML AI Podcast
How to Engineer AI Inference Systems with Philip Kiely - #766
Apr 30
More from 20VC (20 Minute VC)
20VC: Applovin: $160BN Market Cap, $5.48BN Revenue, $10M EBITDA Per Head | Why the Best Do Not Need Mentorship | Why Founders Should Not Angel Invest | Why Kindness in Business Will Slow You Down with Adam Foroughi
Apr 27 · 80 min
Eye on AI
#341 Celia Merzbacher: Beyond the Buzzword: The Real State of Quantum Computing, Sensing, and AI in 2025
Apr 30
More from 20VC (20 Minute VC)
We summarize every new episode. Want them in your inbox?
20VC: Anthropic Raises $45BN but Falls Short on Compute | OpenAI Crushes with GPT5.5 and Codex: Back in the Game? | China Blocks Manus $2BN Deal to Meta | Thoma Bravo Hand Back Medallia Keys to Creditors | Why Google is a Bigger Buy Than Ever Before
20VC: Applovin: $160BN Market Cap, $5.48BN Revenue, $10M EBITDA Per Head | Why the Best Do Not Need Mentorship | Why Founders Should Not Angel Invest | Why Kindness in Business Will Slow You Down with Adam Foroughi
20Product: Replit CEO on Why Coding Models Are Plateauing | Why the SaaS Apocalypse is Justified: Will Incumbents Be Replaced? | Why IDEs Are Dead and Do PMs Survive the Next 3-5 Years with Amjad Masad
20VC: Cursor Acquired for $60BN by xAI | Anthropic Hits $1TRN in Secondary Markets | Did Anthropic Just Kill Figma, Adobe and Canva | Rippling Hits $1BN in ARR | Salesforce Goes Headless: Smart or Stupid | Cerebras IPO 2.0
20VC: Everyone is Wrong; We Will Have More Developers in Five Years | Why Frontier Labs Will Be Way More Valuable Than They Are Today | Are SaaS Companies Cooked: Which Thrive & Which Die with Aaron Levie, Founder at Box
Similar Episodes
Related episodes from other podcasts
The TWIML AI Podcast
Apr 30
How to Engineer AI Inference Systems with Philip Kiely - #766
Eye on AI
Apr 30
#341 Celia Merzbacher: Beyond the Buzzword: The Real State of Quantum Computing, Sensing, and AI in 2025
The Readout Loud
Apr 30
399: Hair-raising trial results, and Servier’s M&A wishlist
This Week in Startups
Apr 30
Mastering AI Video Marketing w/ Magnific CEO Joaquín Cuenca Abela | AI Basics
Moonshots with Peter Diamandis
Apr 30
Google Invests $40B Into Anthropic, GPT 5.5 Drops, and Google Cloud Dominates | EP #252
Explore Related Topics
This podcast is featured in Best Investing Podcasts (2026) — ranked and reviewed with AI summaries.
Read this week's Startups & Product Podcast Insights — cross-podcast analysis updated weekly.
You're clearly into 20VC (20 Minute VC).
Every Monday, we deliver AI summaries of the latest episodes from 20VC (20 Minute VC) and 192+ other podcasts. Free for up to 3 shows.
Start My Monday DigestNo credit card · Unsubscribe anytime