Prompt injection defense is the set of techniques and architectural safeguards designed to prevent or mitigate prompt injection attacks, where an adversary manipulates input to a language model to subvert its intended instructions, potentially leading to data leaks, unauthorized actions, or harmful outputs. This defense is a core component of agentic threat modeling and is essential for securing autonomous systems that interact with external data or users.
