Software Engineering Radio - the podcast for professional software developers

SE Radio 692: Sourabh Satish on Prompt Injection

14 snips
Oct 28, 2025
Sourabh Satish, CTO and co-founder of Pangea and seasoned security entrepreneur, breaks down prompt injection risks for LLMs. He covers the OWASP Top 10 for generative AI, the $10K prompt-injection challenge and attacker techniques like obfuscation and style injection. Learn about ingress/egress filters, system-prompt guardrails, detection methods, and priorities for securing LLM deployments.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Layered Defenses Dramatically Cut Success Rates

  • Simple system-prompt guardrails are trivial to bypass: about 1 in 5 players succeeded.
  • Stronger layered defenses (input/output inspection plus injection detection) reduced success to nearly zero.
INSIGHT

Creative Token Tricks Evade Filters

  • Attackers use distractor instructions, cognitive hacking, and style injection to evade detectors.
  • Tiny token tricks (spaces, newlines, hyphens) and encoding change token interpretation and bypass filters.
ADVICE

Use Ingress And Egress Filters

  • Apply ingress filters to inspect user input and contextual data before sending to the LLM.
  • Apply egress filters to redact or block sensitive outputs like social security or full credit card numbers.
Get the Snipd Podcast app to discover more snips from this episode
Get the app