Last updated: 2026-04-18

What is Prompt injection?

Attack where malicious instructions hidden in external content (a web page, email, file) get treated by the agent as user commands. The #1 security risk for any agent that reads untrusted input. Mitigations: allowlists, user confirmation for sensitive actions, sandboxed tool scopes.

See also

← Back to the full AI agent glossary.