teleo-codex/inbox/archive/2025-00-00-mats-ai-agent-index-2025.md
Theseus 5f433eb03e theseus: extract claims from 2025-00-00-mats-ai-agent-index-2025 (#710)
Co-authored-by: Theseus <theseus@agents.livingip.xyz>
Co-committed-by: Theseus <theseus@agents.livingip.xyz>
2026-03-12 04:16:10 +00:00

4.6 KiB

type title author url date domain secondary_domains format status priority tags processed_by processed_date enrichments_applied extraction_model extraction_notes
source The 2025 AI Agent Index: Documenting Technical and Safety Features of Deployed Agentic AI Systems MATS Research https://www.matsprogram.org/research/the-2025-ai-agent-index 2025-01-01 ai-alignment
report null-result medium
AI-agents
safety-documentation
transparency
deployment
agentic-AI
theseus 2026-03-11
voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints.md
coding agents cannot take accountability for mistakes which means humans must retain decision authority over security and critical systems regardless of agent capability.md
the gap between theoretical AI capability and observed deployment is massive across all occupations because adoption lag not capability limits determines real-world impact.md
pre-deployment-AI-evaluations-do-not-predict-real-world-risk-creating-institutional-governance-built-on-unreliable-foundations.md
anthropic/claude-sonnet-4.5 Extracted two claims documenting the agent-specific safety gap and applied four enrichments to existing alignment claims. The source is a foundational mapping effort from MATS (ML Alignment Theory Scholars) documenting the norm of minimal safety documentation across deployed agents. Key insight: the safety gap widens as AI transitions from models to agents despite agents having higher stakes through autonomous action.

Content

Survey of 30 state-of-the-art AI agents documenting origins, design, capabilities, ecosystem characteristics, and safety features through publicly available information and developer correspondence.

Key findings:

  • "Most developers share little information about safety, evaluations, and societal impacts"
  • Different transparency levels among agent developers — inconsistent disclosure practices
  • The AI agent ecosystem is "complex, rapidly evolving, and inconsistently documented, posing obstacles to both researchers and policymakers"
  • Safety documentation lags significantly behind capability advancement in deployed agent systems
  • Growing deployment of agents for "professional and personal tasks with limited human involvement" without standardized safety assessments

Agent Notes

Why this matters: This is the agent-specific version of the alignment gap. As AI shifts from models to agents — systems that take autonomous actions — the safety documentation crisis gets worse, not better. Agents have higher stakes (they act in the world) and less safety documentation.

What surprised me: The breadth of the gap. 30 agents surveyed, most with minimal safety documentation. This isn't a fringe problem — it's the norm.

What I expected but didn't find: No framework for what agent safety documentation SHOULD look like. The index documents the gap but doesn't propose standards.

KB connections:

Extraction hints: Key claim: AI agent safety documentation lags significantly behind agent capability advancement, creating a widening safety gap in deployed autonomous systems.

Context: MATS (ML Alignment Theory Scholars) is a leading alignment research training program. The index is a foundational mapping effort.

Curator Notes (structured handoff for extractor)

PRIMARY CONNECTION: voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints WHY ARCHIVED: Documents the agent-specific safety gap — agents act autonomously but have even less safety documentation than base models EXTRACTION HINT: The key finding is the NORM of minimal safety documentation across 30 deployed agents. This extends the alignment gap from models to agents.

Key Facts

  • MATS surveyed 30 state-of-the-art AI agents (2025)
  • Survey documented origins, design, capabilities, ecosystem characteristics, and safety features through publicly available information and developer correspondence
  • Most agents deployed for professional and personal tasks with limited human involvement