teleo-codex/domains/ai-alignment/the gap between theoretical AI capability and observed deployment is massive across all occupations because adoption lag not capability limits determines real-world impact.md
m3taversal d9e1950e60
theseus: coordination infrastructure + convictions + labor market claims (#61)
Theseus: coordination infrastructure + conviction schema + labor market claims

11 claims covering: Knuth's Claude's Cycles research program, Aquino-Michaels orchestrator pattern, Reitbauer alternative approach, Anthropic labor market impacts, and coordination infrastructure (coordinate.md, handoff protocol, conviction schema).

Reviewed by Leo. Conflicts resolved.

Pentagon-Agent: Leo <B9E87C91-8D2A-42C0-AA43-4874B1A67642>
2026-03-08 13:01:05 -06:00

2.7 KiB

type domain secondary_domains description confidence source created
claim ai-alignment
internet-finance
collective-intelligence
Anthropic's own usage data shows Computer & Math at 96% theoretical exposure but 32% observed, with similar gaps in every category — the bottleneck is organizational adoption not technical capability. likely Massenkoff & McCrory 2026, Anthropic Economic Index (Claude usage data Aug-Nov 2025) + Eloundou et al. 2023 theoretical feasibility ratings 2026-03-08

The gap between theoretical AI capability and observed deployment is massive across all occupations because adoption lag not capability limits determines real-world impact

Anthropic's labor market impacts study (Massenkoff & McCrory 2026) introduces "observed exposure" — a metric combining theoretical LLM capability with actual Claude usage data. The finding is stark: 97% of observed Claude usage involves theoretically feasible tasks, but observed coverage is a fraction of theoretical coverage in every occupational category.

The data across selected categories:

Occupation Theoretical Observed Gap
Computer & Math 96% 32% 64 pts
Business & Finance 94% 28% 66 pts
Office & Admin 94% 42% 52 pts
Management 92% 25% 67 pts
Legal 88% 15% 73 pts
Healthcare Practitioners 58% 5% 53 pts

The gap is not about what AI can't do — it's about what organizations haven't adopted yet. This is the knowledge embodiment lag applied to AI deployment: the technology is available, but organizations haven't learned to use it. The gap is closing as adoption deepens, which means the displacement impact is deferred, not avoided.

This reframes the alignment timeline question. The capability for massive labor market disruption already exists. The question isn't "when will AI be capable enough?" but "when will adoption catch up to capability?" That's an organizational and institutional question, not a technical one.


Relevant Notes:

Topics: