LLM01
Prompt Injection
Untrusted content manipulates model instructions and control flow.
Injected prompts can bypass guardrails and trigger unauthorized model behavior.
Prevention Checklist
- Separate trusted system instructions from untrusted user/context data.
- Apply policy checks before and after model calls.
- Minimize tool permissions and require explicit user confirmation for high-risk actions.