Skip to content

Red Teaming

Deliberately testing an AI system by attempting to make it produce harmful, incorrect, or unintended outputs. Red teams probe for safety vulnerabilities, bias, and failure modes. It is a critical part of responsible AI development before deployment.

Related terms

AI SafetyGuardrailsAdversarial Attack
← Back to glossary