Skip to content

AI Safety

The field of research focused on ensuring AI systems behave as intended and do not cause unintended harm. AI safety covers alignment, robustness, interpretability, and governance, aiming to keep systems reliable, controllable, and beneficial.

Related terms

AlignmentResponsible AIGuardrails
← Back to glossary