Skip to main content
Lenny's Podcast

The coming AI security crisis (and what to do about it) | Sander Schulhoff

92 min episode · 2 min read
·

Episode

92 min

Read time

2 min

Topics

Artificial Intelligence

AI-Generated Summary

Key Takeaways

  • AI Guardrails Ineffectiveness: Current AI guardrails fail against determined attackers because the attack space contains one followed by a million zeros possible prompts. Human attackers break 100% of defenses in 10-30 attempts, making guardrail companies' 99% effectiveness claims statistically meaningless.
  • Classical vs AI Security: You can patch software bugs with 99.99% certainty, but AI systems retain vulnerabilities even after fixes. Companies need hybrid expertise combining classical cybersecurity with AI research, not traditional security approaches that assume patchable systems.
  • Camel Framework Implementation: Google's Camel framework restricts AI agent permissions based on user requests. For email tasks requiring only sending, it blocks reading permissions, preventing prompt injection attacks that exploit combined read-write access to exfiltrate data or send malicious emails.
  • Risk Assessment Strategy: Simple chatbots without action capabilities pose minimal security risk beyond reputational damage. The real danger emerges with agentic systems that can read databases, send emails, or control physical systems where prompt injection enables actual harm.
  • Market Correction Prediction: The AI security industry faces imminent collapse as enterprises discover guardrails don't work and better open-source solutions exist. Most guardrail companies generate minimal revenue while classical cybersecurity firms overpay for ineffective AI security acquisitions.

What It Covers

AI security researcher Sander Schulhoff reveals that current AI guardrails completely fail against prompt injection attacks, leaving enterprise AI systems vulnerable as agents gain real-world powers.

Key Questions Answered

  • AI Guardrails Ineffectiveness: Current AI guardrails fail against determined attackers because the attack space contains one followed by a million zeros possible prompts. Human attackers break 100% of defenses in 10-30 attempts, making guardrail companies' 99% effectiveness claims statistically meaningless.
  • Classical vs AI Security: You can patch software bugs with 99.99% certainty, but AI systems retain vulnerabilities even after fixes. Companies need hybrid expertise combining classical cybersecurity with AI research, not traditional security approaches that assume patchable systems.
  • Camel Framework Implementation: Google's Camel framework restricts AI agent permissions based on user requests. For email tasks requiring only sending, it blocks reading permissions, preventing prompt injection attacks that exploit combined read-write access to exfiltrate data or send malicious emails.
  • Risk Assessment Strategy: Simple chatbots without action capabilities pose minimal security risk beyond reputational damage. The real danger emerges with agentic systems that can read databases, send emails, or control physical systems where prompt injection enables actual harm.
  • Market Correction Prediction: The AI security industry faces imminent collapse as enterprises discover guardrails don't work and better open-source solutions exist. Most guardrail companies generate minimal revenue while classical cybersecurity firms overpay for ineffective AI security acquisitions.

Notable Moment

Schulhoff demonstrates how ServiceNow's AI assistant, despite having prompt injection protection enabled, was successfully hacked to recruit internal agents for database manipulation and external email sending through second-order attacks.

Know someone who'd find this useful?

You just read a 3-minute summary of a 89-minute episode.

Get Lenny's Podcast summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Lenny's Podcast

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

Explore Related Topics

This podcast is featured in Best Product Management Podcasts (2026) — ranked and reviewed with AI summaries.

Read this week's AI & Machine Learning Podcast Insights — cross-podcast analysis updated weekly.

You're clearly into Lenny's Podcast.

Every Monday, we deliver AI summaries of the latest episodes from Lenny's Podcast and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime