teleo-codex/domains/ai-alignment/frontier-ai-dangerous-capability-testing-limited-to-three-companies.md
Teleo Agents 828dda016c theseus: extract from 2025-07-00-fli-ai-safety-index-summer-2025.md
- Source: inbox/archive/2025-07-00-fli-ai-safety-index-summer-2025.md
- Domain: ai-alignment
- Extracted by: headless extraction cron (worker 6)

Pentagon-Agent: Theseus <HEADLESS>
2026-03-12 04:52:27 +00:00

2.9 KiB

type domain description confidence source created secondary_domains
claim ai-alignment Only Anthropic, OpenAI, and DeepMind conduct substantive testing for dangerous capabilities among frontier AI developers, per FLI Summer 2025 index likely Future of Life Institute, AI Safety Index Summer 2025, July 2025 2026-03-11
grand-strategy

Frontier AI dangerous capability testing is limited to three companies

Future of Life Institute's Summer 2025 evaluation found that only three frontier AI companies—Anthropic, OpenAI, and Google DeepMind—conduct substantive testing for dangerous capabilities such as bioterrorism potential and cyberattack vectors. The remaining four assessed companies (x.AI, Meta, Zhipu AI, DeepSeek) either do not test for these risks or do not disclose such testing.

This concentration of risk assessment creates a bifurcated landscape where the majority of frontier AI development proceeds without systematic evaluation of catastrophic misuse potential. The gap is particularly concerning given that all companies are developing increasingly capable systems, but only a minority are actively probing for dangerous emergent capabilities.

The FLI index evaluated companies across six dimensions, with "Risk Assessment" specifically measuring dangerous capability testing. The three companies conducting such testing still received overall grades of C+ to C-, indicating that even substantive risk assessment does not guarantee comprehensive safety practices.

Why this matters: Dangerous capability testing remains a voluntary practice adopted by safety-conscious labs rather than an industry norm, leaving significant capability development unmonitored for catastrophic risk vectors. This pattern reinforces the structural race-to-the-bottom dynamic: companies that invest in expensive risk assessment (bioterrorism, cyberattack modeling) gain no competitive advantage and may face capability delays, while competitors skip these costs entirely.

Evidence

From FLI's assessment:

  • 3 of 7 frontier AI companies (Anthropic, OpenAI, DeepMind) conduct substantive dangerous capability testing
  • 4 of 7 companies (x.AI, Meta, Zhipu AI, DeepSeek) do not conduct or disclose such testing
  • Testing focuses on bioterrorism potential and cyberattack capabilities
  • Even companies conducting testing scored C+ or below overall

Relevant Notes:

Topics: