Last updated: 2026-04-18
What is Prompt injection?
Attack where malicious instructions hidden in external content (a web page, email, file) get treated by the agent as user commands. The #1 security risk for any agent that reads untrusted input. Mitigations: allowlists, user confirmation for sensitive actions, sandboxed tool scopes.
See also
← Back to the full AI agent glossary.