Safety
AI Safety
Definition
“
The interdisciplinary field focused on ensuring AI systems operate reliably, ethically, and without causing unintended harm. AI safety encompasses alignment research, robustness testing, interpretability, and the development of evaluation frameworks to assess model behavior before deployment in critical applications.
”
Related Terms
No related terms linked yet.
Explore all terms →