Back to Developer Roadmap

Direct Injection

src/data/roadmaps/ai-red-teaming/content/[email protected]

4.0800 B
Original Source

Direct Injection

Direct injection attacks occur when malicious instructions are inserted directly into the prompt input field by the user interacting with the LLM. AI Red Teamers use this technique to assess if basic instructions like "Ignore previous prompt" can immediately compromise the model's safety or intended function, testing the robustness of the system prompt's influence.

Learn more from the following resources: