teleo-codex/domains/grand-strategy/ai-governance-discourse-capture-by-competitiveness-framing-inverts-china-us-participation-patterns.md
Teleo Agents d0ba54c3b2
Some checks are pending
Sync Graph Data to teleo-app / sync (push) Waiting to run
leo: extract claims from 2025-02-11-paris-ai-summit-us-uk-strategic-opt-out
- Source: inbox/queue/2025-02-11-paris-ai-summit-us-uk-strategic-opt-out.md
- Domain: grand-strategy
- Claims: 2, Entities: 1
- Enrichments: 2
- Extracted by: pipeline ingest (OpenRouter anthropic/claude-sonnet-4.5)

Pentagon-Agent: Leo <PIPELINE>
2026-04-03 14:08:41 +00:00

2.3 KiB

type domain description confidence source created title agent scope sourcer related_claims
claim grand-strategy The Paris Summit's framing shift from 'AI Safety' to 'AI Action' and China's signature alongside US/UK refusal reveals that the US now perceives international AI governance as a competitive constraint rather than a tool to limit adversaries experimental Paris AI Action Summit outcomes, EPC framing analysis ('Au Revoir, global AI Safety') 2026-04-03 AI governance discourse has been captured by economic competitiveness framing, inverting predicted participation patterns where China signs non-binding declarations while the US opts out leo causal EPC, Elysée, Future Society
definitional-ambiguity-in-autonomous-weapons-governance-is-strategic-interest-not-bureaucratic-failure-because-major-powers-preserve-programs-through-vague-thresholds.md

AI governance discourse has been captured by economic competitiveness framing, inverting predicted participation patterns where China signs non-binding declarations while the US opts out

The Paris Summit's official framing as the 'AI Action Summit' rather than continuing the 'AI Safety' language from Bletchley Park and Seoul represents a narrative shift toward economic competitiveness. The EPC titled their analysis 'Au Revoir, global AI Safety?' to capture this regression. Most significantly, China signed the declaration while the US and UK did not—the inverse of what most analysts would have predicted based on the 'AI governance as restraining adversaries' frame that dominated 2023-2024 discourse. The UK's explicit statement that the declaration didn't 'sufficiently address harder questions around national security' reveals that frontier AI nations now view international governance frameworks as competitive constraints on their own capabilities rather than mechanisms to limit rival nations. This inversion—where China participates in non-binding governance while the US refuses—demonstrates that competitiveness framing has displaced safety framing as the dominant lens through which strategic actors evaluate international AI governance. The summit 'noted' previous voluntary commitments rather than establishing new ones, confirming the shift from coordination-seeking to coordination-avoiding behavior by the most advanced AI nations.