AI companies should do model evals and uplift experiments to determine whether models have dangerous capabilities or how close they are. They should also prepare to check whether models will act well in high-stakes situations.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
44
%
29
%
32
%
1
%
1
%
1
%
0
%

AIs might scheme, i.e. fake alignment and subvert safety measures in order to gain power. AI companies should prepare for risks from models scheming, especially during internal deployment: if they can't reliably prevent scheming, they should prepare to catch some schemers and deploy potentially scheming models safely.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
2
%
8
%
2
%
2
%
2
%
2
%
2
%

AI companies should do (extreme-risk-focused) safety research, and they should publish it to boost safety at other AI companies. Additionally, they should assist external safety researchers by sharing deep model access and mentoring.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
68
%
55
%
37
%
28
%
0
%
15
%
8
%

AI companies should prepare to prevent catastrophic misuse for deployments via API, once models are capable of enabling catastrophic harm.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
12
%
4
%
5
%
0
%
0
%
0
%
0
%

AI companies should prepare to protect model weights and code by the time AI massively boosts R&D, even from top-priority operations by the top cyber-capable institutions.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
2
%
5
%
0
%
0
%
0
%
0
%
0
%

AI companies should share information on incidents, risks, and capabilities — but not share some capabilities research.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
35
%
13
%
32
%
0
%
28
%
0
%
0
%

AI companies should plan for the possibility that dangerous capabilities appear soon and safety isn't easy: both for evaluating and improving safety of their systems and for using their systems to make the world safer.

Anthropic
DeepMind
OpenAI
Meta
xAI
Microsoft
DeepSeek
14
%
26
%
0
%
0
%
0
%
1
%
0
%

I'm Zach Stein-Perlman. I'm worried about future powerful Als causing an existential catastrophe. I track what AI companies are doing in terms of safety.

For some details on what AI companies should do and what they are doing in terms of safety, click around this scorecard. Or check out the articles below, the rest of my blog, or the resources I maintain.