Distributional AGI Safety

By: Nenad Tomašev, Matija Franklin, Julian Jacobs, Sébastien Krier, Simon Osindero

Published: 2025-12-19

View on arXiv →
#cs.AI

Abstract

We introduce the concept of Distributional AGI Safety, a framework for analyzing and ensuring the safety of Artificial General Intelligence (AGI) systems across diverse operational contexts and potential failure modes. This approach moves beyond single-point safety assessments to consider the full distribution of possible AGI behaviors and their societal impacts. We propose methods for robust safety alignment and risk mitigation, emphasizing the need for adaptable and context-aware safety measures to address the multifaceted challenges of AGI deployment.

FEEDBACK

Projects

No projects yet

Distributional AGI Safety | ArXiv Intelligence