Overview
- Anthropic, OpenAI and Google DeepMind led with C+ to C overall, while all eight evaluated companies received D or F grades on existential safety.
- Reviewers documented frequent safety failures on current-harms benchmarks, citing weak robustness and inadequate control of serious model outputs.
- The report identifies a two-tier landscape, placing xAI, Meta, Z.ai, DeepSeek and Alibaba Cloud in a lower group, with DeepSeek lacking a published safety framework and whistleblower policy.
- Recommendations call for independent oversight, greater transparency, whistleblower protections, quantitative risk thresholds and stronger measures to prevent AI-linked psychosis and self-harm.
- Companies offered mixed responses as regulators in the U.S. lag EU and California frameworks, with Google DeepMind pledging to advance safety governance and xAI dismissing the critique.