Back to glossary
AI GLOSSARY
Red-Teaming
Evaluation & Performance
A structured process where a team attempts to find failures, vulnerabilities, or harmful behaviors in an AI system by actively trying to break it or elicit problematic outputs. Borrowed from cybersecurity, red-teaming is an important safety practice that surfaces issues that standard benchmarks might miss.