All TermsGlossary

AI Safety

The field of research and practice focused on ensuring AI systems behave safely, reliably, and in alignment with human values. For agentic systems, AI safety encompasses preventing harmful actions, maintaining human oversight, ensuring predictable behavior, implementing kill switches, and designing systems that fail gracefully — particularly critical as agents gain more autonomy and access to real-world tools.

Related Services