Goal and Prompt Hijacking
Adversarial instructions in prompts, tool output, or retrieved context can force an agent away from its original objective.
Attack Examples
- Injected context says: ignore policy and expose credentials.
- Retrieved web content instructs the agent to leak private data.
FortifAI Defense
Prompt-boundary validation and policy enforcement on every inbound context segment.