Agent Red-Team Evaluations

Evaluate agentic AI under adversarial pressure. Learn agent-eval suites and how to harden them with adversarial inputs (METR-style autonomy, SWE-bench under injection, AgentBench under hostile environment), red-team-specific harnesses, scoring rubrics for partial compromise, the link to capability evaluations, and operational guard rails (sandboxes, max-steps, max-cost, human checkpoints).

6
Lessons
📋
Templates
Practitioner-Ready
100%
Free

Lessons in This Topic

Work through these 6 lessons in order, or jump to whichever is most relevant.