Practical AI

Threat modeling LLM apps

28 snips
Aug 22, 2024
Donato Capitella, Principal Security Consultant at WithSecure, specializes in threat modeling for AI applications. He discusses the critical need for threat modeling in the context of large language models (LLMs) and shares insights on vulnerabilities, such as prompt injection risks. Donato emphasizes the importance of validating outputs to maintain trustworthiness and explores innovative strategies for secure integration in AI systems. The conversation also touches on the exciting future of LLM technology and the role of ethical hackers in enhancing cybersecurity.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ADVICE

Untrusted LLM Output

  • Treat LLM output as untrusted, similar to handling emails from unknown senders.
  • Apply security controls to mitigate risks from this untrusted data, verifying and validating information before acting on it.
INSIGHT

LLM Security Canvas

  • Donato Capitella's LLM Security Canvas focuses on input and output controls for secure LLM application deployment.
  • It prioritizes output validation, checking for harmful content and format, especially in rendered outputs like Markdown or HTML.
ADVICE

Input Validation

  • Implement input validation using semantic routing, topical guardrails, and prompt injection detection models.
  • Check input length, format, and character set, especially considering potential exploits with low-resource languages.
Get the Snipd Podcast app to discover more snips from this episode
Get the app