Back to Developer Roadmap

Model Vulnerabilities

src/data/roadmaps/ai-red-teaming/content/[email protected]

4.0828 B
Original Source

Model Vulnerabilities

This category covers attacks and tests targeting the AI model itself, beyond the prompt interface. AI Red Teamers investigate inherent weaknesses in the model's architecture, training data artifacts, or prediction mechanisms, such as susceptibility to data extraction, poisoning, or adversarial manipulation.

Learn more from the following resources: