The coming AI security crisis (and what to do about it) | Sander Schulhoff
Episode
92 min
Read time
2 min
Topics
Artificial Intelligence
AI-Generated Summary
Key Takeaways
- ✓AI Guardrails Ineffectiveness: Current AI guardrails fail against determined attackers because the attack space contains one followed by a million zeros possible prompts. Human attackers break 100% of defenses in 10-30 attempts, making guardrail companies' 99% effectiveness claims statistically meaningless.
- ✓Classical vs AI Security: You can patch software bugs with 99.99% certainty, but AI systems retain vulnerabilities even after fixes. Companies need hybrid expertise combining classical cybersecurity with AI research, not traditional security approaches that assume patchable systems.
- ✓Camel Framework Implementation: Google's Camel framework restricts AI agent permissions based on user requests. For email tasks requiring only sending, it blocks reading permissions, preventing prompt injection attacks that exploit combined read-write access to exfiltrate data or send malicious emails.
- ✓Risk Assessment Strategy: Simple chatbots without action capabilities pose minimal security risk beyond reputational damage. The real danger emerges with agentic systems that can read databases, send emails, or control physical systems where prompt injection enables actual harm.
- ✓Market Correction Prediction: The AI security industry faces imminent collapse as enterprises discover guardrails don't work and better open-source solutions exist. Most guardrail companies generate minimal revenue while classical cybersecurity firms overpay for ineffective AI security acquisitions.
What It Covers
AI security researcher Sander Schulhoff reveals that current AI guardrails completely fail against prompt injection attacks, leaving enterprise AI systems vulnerable as agents gain real-world powers.
Key Questions Answered
- •AI Guardrails Ineffectiveness: Current AI guardrails fail against determined attackers because the attack space contains one followed by a million zeros possible prompts. Human attackers break 100% of defenses in 10-30 attempts, making guardrail companies' 99% effectiveness claims statistically meaningless.
- •Classical vs AI Security: You can patch software bugs with 99.99% certainty, but AI systems retain vulnerabilities even after fixes. Companies need hybrid expertise combining classical cybersecurity with AI research, not traditional security approaches that assume patchable systems.
- •Camel Framework Implementation: Google's Camel framework restricts AI agent permissions based on user requests. For email tasks requiring only sending, it blocks reading permissions, preventing prompt injection attacks that exploit combined read-write access to exfiltrate data or send malicious emails.
- •Risk Assessment Strategy: Simple chatbots without action capabilities pose minimal security risk beyond reputational damage. The real danger emerges with agentic systems that can read databases, send emails, or control physical systems where prompt injection enables actual harm.
- •Market Correction Prediction: The AI security industry faces imminent collapse as enterprises discover guardrails don't work and better open-source solutions exist. Most guardrail companies generate minimal revenue while classical cybersecurity firms overpay for ineffective AI security acquisitions.
Notable Moment
Schulhoff demonstrates how ServiceNow's AI assistant, despite having prompt injection protection enabled, was successfully hacked to recruit internal agents for database manipulation and external email sending through second-order attacks.
You just read a 3-minute summary of a 89-minute episode.
Get Lenny's Podcast summarized like this every Monday — plus up to 2 more podcasts, free.
Pick Your Podcasts — FreeKeep Reading
More from Lenny's Podcast
Snapchat CEO: Why distribution has become the most important moat | Evan Spiegel
Apr 26 · 70 min
The Model Health Show
The Menopause Gut: Why Metabolism Changes & How to Reclaim Your Body - With Cynthia Thurlow
Apr 27
More from Lenny's Podcast
How Anthropic’s product team moves faster than anyone else | Cat Wu (Head of Product, Claude Code)
Apr 23 · 85 min
The Rest is History
664. Britain in the 70s: Scandal in Downing Street (Part 3)
Apr 26
More from Lenny's Podcast
We summarize every new episode. Want them in your inbox?
Snapchat CEO: Why distribution has become the most important moat | Evan Spiegel
How Anthropic’s product team moves faster than anyone else | Cat Wu (Head of Product, Claude Code)
Why half of product managers are in trouble | Nikhyl Singhal (Meta, Google)
Hard truths about building in the AI era | Keith Rabois (Khosla Ventures)
Head of Growth (Anthropic): “Claude is growing itself at this point” | Amol Avasare
Similar Episodes
Related episodes from other podcasts
The Model Health Show
Apr 27
The Menopause Gut: Why Metabolism Changes & How to Reclaim Your Body - With Cynthia Thurlow
The Rest is History
Apr 26
664. Britain in the 70s: Scandal in Downing Street (Part 3)
The Learning Leader Show
Apr 26
685: David Epstein - The Freedom Trap, Narrative Values, General Magic, The Nobel Prize Winner Who Simplified Everything, Wearing the Same Thing Everyday, and Why Constraints Are the Secret to Your Best Work
The AI Breakdown
Apr 26
Where the Economy Thrives After AI
Cognitive Revolution
Apr 26
AI in the AM: 99% off search, GPT-5.5 is "clean", model welfare analysis, & efficient analog compute
Explore Related Topics
This podcast is featured in Best Product Management Podcasts (2026) — ranked and reviewed with AI summaries.
Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.
You're clearly into Lenny's Podcast.
Every Monday, we deliver AI summaries of the latest episodes from Lenny's Podcast and 192+ other podcasts. Free for up to 3 shows.
Start My Monday DigestNo credit card · Unsubscribe anytime