AI Red Teaming for complete first-timers
Intro to AI red teaming​
Is this your first foray into AI red teaming? And probably red teaming in general? Great. This is for you.
Red teaming is the process of simulating real-world attacks to identify vulnerabilities.
AI red teaming is the process of simulating real-world attacks to identify vulnerabilities in artificial-intelligence systems. There are two scopes people often use to refer to AI red teaming:
- Prompt injection of LLMs
- A wider scope of testing pipelines, plugins, agents, and broader system dynamics