Distributional AGI Safety
By: Nenad Tomašev, Matija Franklin, Julian Jacobs, Sébastien Krier, Simon Osindero
Published: 2025-12-19
View on arXiv →#cs.AI
Abstract
We introduce the concept of Distributional AGI Safety, a framework for analyzing and ensuring the safety of Artificial General Intelligence (AGI) systems across diverse operational contexts and potential failure modes. This approach moves beyond single-point safety assessments to consider the full distribution of possible AGI behaviors and their societal impacts. We propose methods for robust safety alignment and risk mitigation, emphasizing the need for adaptable and context-aware safety measures to address the multifaceted challenges of AGI deployment.