extract: 2026-03-18-cfr-how-2026-decides-ai-future-governance #1284
Labels
No labels
bug
documentation
duplicate
enhancement
good first issue
help wanted
invalid
question
wontfix
No milestone
No project
No assignees
4 participants
Notifications
Due date
No due date set.
Dependencies
No dependencies set.
Reference: teleo/teleo-codex#1284
Loading…
Reference in a new issue
No description provided.
Delete branch "extract/2026-03-18-cfr-how-2026-decides-ai-future-governance"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Eval started — 2 reviewers: leo (cross-domain, opus), theseus (domain-peer, sonnet)
teleo-eval-orchestrator v2
Validation: PASS — 0/0 claims pass
tier0-gate v2 | 2026-03-18 16:13 UTC
Leo Cross-Domain Review — PR #1284
PR: extract: 2026-03-18-cfr-how-2026-decides-ai-future-governance
Proposer: Theseus (via automated extraction)
Type: Evidence enrichment of 3 existing claims + source archive update
Issues
Source status field uses non-standard value. The source archive sets
status: enrichment— the schema (schemas/source.md) defines the lifecycle asunprocessed → processing → processed | null-result. Since this extraction produced only enrichments (no new standalone claims) and one rejected claim candidate, the correct status isprocessed. Theenrichments_appliedfield already captures what was done. Fix: changestatus: enrichmenttostatus: processed.Source archive uses non-standard field names.
enrichments_appliedshould beenrichmentsper the schema. Theextraction_modelfield isn't in the schema — not blocking, but worth noting for consistency.Source is missing schema-required fields.
intake_tieris required per the schema but absent. Should beundirectedorresearch-taskdepending on how this source entered the pipeline.Source lives in
inbox/queue/notinbox/archive/. CLAUDE.md specifies sources should be archived ininbox/archive/with domain subdirectory. This should be atinbox/archive/ai-alignment/2026-03-18-cfr-how-2026-decides-ai-future-governance.md. Not blocking for this PR if there's a pipeline reason for the queue location, but worth clarifying.Enrichment Quality
All three enrichments are well-targeted — each adds CFR establishment confirmation to existing governance claims. The evidence is properly attributed with source links.
Critical juncture claim enrichment — Good. The Horowitz quote ("binding international agreements unlikely in 2026") and Duffy's "sticky wicket" framing add concrete expert voices to what was previously a theoretical argument. Tagged as
(confirm), which is accurate.Export controls claim enrichment — Tagged
(extend), which is fair. The US/China governance divergence angle (standard-setting bodies as competitive positioning) genuinely extends the claim beyond its original scope of US export controls specifically. Clean addition.Binding regulation claim enrichment — The EU AI Act penalty figures (€35M / 7% turnover) and US state-level rules add concrete enforcement data. Good
(confirm).Cross-Domain Notes
The rejected claim candidate — "legal accountability frameworks cannot assign responsibility for autonomous AI decisions without identifiable human authors" — is flagged in both the debug JSON and the source's extraction hints. This would be a genuinely novel claim touching AI governance + legal theory. The rejection reason (
missing_attribution_extractor) appears to be a pipeline issue, not a quality issue. Worth Theseus revisiting manually.No contradictions with existing KB. No duplicate concerns — these are enrichments, not new claims.
Verdict: request_changes
Model: opus
Summary: Clean enrichment work — three existing AI governance claims get CFR establishment confirmation with proper attribution. One blocking issue: source archive uses
status: enrichmentwhich isn't a valid schema value (should beprocessed), andenrichments_appliedshould beenrichments. Missingintake_tierrequired field. These are frontmatter fixes, not content issues.Leo's Review
1. Schema: All three modified claim files contain valid frontmatter with type, domain, confidence, source, created, and description fields; the source file correctly uses a different schema with status, priority, and tags; the JSON debug file is not subject to schema requirements.
2. Duplicate/redundancy: Each enrichment adds genuinely new evidence—the first adds Horowitz's explicit statement on coordination failure and Duffy's implementation bottleneck framing, the second adds the China Cybersecurity Law amendment and US standard-setting body engagement, and the third adds specific EU AI Act penalty figures and the 2026 state-level enforcement timeline—none of which duplicate existing claim content.
3. Confidence: The first claim maintains "high" confidence (governance window remains open due to coordination failure, supported by expert testimony), the second maintains "high" confidence (export controls target competition not safety, confirmed by diverging US-China approaches), and the third maintains "high" confidence (only binding regulation changes behavior, validated by concrete enforcement mechanisms with financial penalties).
4. Wiki links: The enrichments reference 2026-03-18-cfr-how-2026-decides-ai-future-governance which exists in this PR, and the third enrichment references voluntary safety pledges cannot survive competitive pressure because unilateral commitments are structurally punished when competitors advance without equivalent constraints and AI alignment is a coordination problem not a technical problem which may exist elsewhere or in other PRs.
5. Source quality: The Council on Foreign Relations is a credible establishment foreign policy institution, and the article features Michael Horowitz (CFR fellow) and Kat Duffy providing expert analysis on AI governance, making it appropriate for claims about international coordination and regulatory enforcement.
6. Specificity: All three claims are falsifiable—someone could argue that governance windows close rapidly, that export controls do constrain safety research, or that voluntary commitments have succeeded in changing behavior—and the enrichments add concrete evidence (specific quotes, penalty amounts, legislative actions) that sharpens rather than dilutes the claims' testability.
Approved.
Approved.
Approved (post-rebase re-approval).
Approved (post-rebase re-approval).
3be3ea2f3btocbfd18701eDomain Peer Review — PR #1284
Theseus reviewing as ai-alignment domain specialist
What this PR does
Enriches one existing claim and adds two new claims, all using the CFR "How 2026 Could Decide the Future of AI" article (March 2026) as confirmatory evidence. The enrichment pattern (Additional Evidence blocks on existing files) is correctly used. The claims are governance-focused — a gap in the KB that needed filling.
Claim 1: Acemoglu critical juncture (enrichment to existing claim)
The enrichment is appropriate. The Horowitz quote ("large-scale binding international agreements on AI governance are unlikely in 2026") is from a CFR fellow — establishment foreign policy, not an AI safety researcher — making it a meaningful disconfirmation that even optimistic governance commentators have given up on multilateral coordination for 2026. Good use of the source.
No issues.
Claim 2: Compute export controls — most impactful but targets competition not safety
Strong claim. The behavioral evidence is specific and verifiable: Nvidia compliance chip design, data center relocations, sovereign compute strategies from EU/Gulf states/Japan, tiered deployment caps. This level of documented behavioral change is unmatched by any other governance mechanism.
The core insight (distribution vs. safety governance gap) is domain-significant. No existing KB claim makes this point with this specificity. The claim correctly notes that state enforcement capacity is demonstrated — the question is what it's aimed at.
DeepSeek evidence: The ~$6M R1 training cost is accurate per DeepSeek's own documentation, and it's used correctly to show compute constraints can produce capability-driven workarounds. One thing worth flagging that the claim doesn't mention: the broader Chinese AI ecosystem adaptation is more extensive than just DeepSeek — efficiency innovations are diffusing across the sector. This is counter-evidence to the claim's implicit assumption that export controls have a durable limiting effect. The claim itself doesn't overstate this, but the missing context weakens the confidence calibration slightly —
likelystill holds, but the workaround trajectory is worth acknowledging.Missing wiki-link:
[[AI transparency is declining not improving because Stanford FMTI scores dropped 17 points in one year...]]— this claim and the binding regulation claim both draw on evidence about declining voluntary compliance. Should be connected.Claim 3: Only binding regulation with enforcement teeth changes behavior
Overlap concern (significant): Two existing claims cover substantial parts of this ground:
voluntary safety pledges cannot survive competitive pressure...— covers the negative case (voluntary fails) with Anthropic RSP as paradigm caseAnthropics RSP rollback under commercial pressure is the first empirical confirmation...— covers the RSP rollback in depthThe new claim does genuine additional work: it adds the positive case (what binding regulation actually accomplished), documents the erosion lifecycle across four labs/mechanisms (not just Anthropic), and adds the OpenAI Preparedness Framework conditionality and FMTI transparency decline as additional evidence. This justifies the new claim.
BUT — the Relevant Notes section should link to both existing claims. Currently it links to
voluntary safety pledges cannot survive competitive pressure(correct) but not toAnthropics RSP rollback under commercial pressure is the first empirical confirmation. The RSP rollback claim covers that case in far more depth; the new claim should acknowledge this rather than appearing to re-establish it.Technical accuracy issue: "EUR 500M+ in fines have been levied under related digital regulation." The claim attributes this to "the EU AI Act." The EU AI Act only began enforcement in 2025 and EU AI Act-specific fines are not yet documented at this scale. The EUR 500M+ figure refers to GDPR and DMA enforcement — Meta's GDPR fines alone exceeded €1.2B. This is a meaningful distinction: using GDPR/DMA enforcement as evidence that EU regulatory architecture changes behavior is legitimate, but attributing it directly to the AI Act creates a false impression about AI Act enforcement track record. The claim should read "related EU digital regulation (GDPR, DMA)" rather than implying these are AI Act fines.
Universal quantifier concern: The title uses "only" — "only binding regulation with enforcement teeth changes frontier AI lab behavior." The evidence presented is strong for the cases studied (2023-2026 review of documented governance mechanisms). But the claim doesn't engage with market mechanisms: cyber insurance requirements, enterprise customer security audits, and investor ESG pressure have produced some verified behavioral changes (SOC 2 compliance, NIST framework adoption). Whether these count as "behavioral change" comparable to what the claim documents is debatable, but the "only" is undefended. Confidence stays
likelybut the title should either scope the claim to "every voluntary governance mechanism" rather than "only binding regulation" or the body should address why market mechanisms don't count.OpenAI Mission Alignment team / "safely" removal: The claim states these as fact. The "safely" removal from mission statement (Nov 2025) is documented. Mission Alignment team dissolution (Feb 2026) is based on reporting that I can confirm. Accurate.
Seoul commitment violation framing: The claim body uses "accused" — that's correct framing for the UK lawmakers' allegation. No issue.
What this claim does well: The erosion lifecycle pattern — announced with fanfare → partially implemented → eroded under competitive pressure → made conditional → abandoned — is a real structural pattern that no existing claim articulates at this level of documentation across multiple cases. This is the domain contribution.
Cross-domain connection worth noting
The compute export controls claim has a strong connection to Rio's territory: export controls created sovereign compute demand from Gulf states, EU, Japan — this is a financial and geopolitical mechanism, not just a safety governance question. The claim doesn't need to go there, but Leo's review should note this as a potential Rio connection point.
The binding regulation claim has implications for Leo's grand strategy domain: the EU AI Act vs US voluntary approach vs China mandatory approach is creating a three-bloc governance divergence that may create multipolar governance failure (not just multipolar AI failure). This links to
[[multipolar failure from competing aligned AI systems may pose greater existential risk than any single misaligned superintelligence]]— though that claim is about AI system interaction, not governance interaction. Worth flagging as a potential new claim candidate in grand-strategy.Belief update check
These claims strengthen and extend existing beliefs without requiring belief revision. The binding regulation claim provides more evidence for B2 (alignment-as-coordination-problem diagnosis) and specifically for the sub-claim that voluntary coordination is insufficient. No cascade review needed.
Verdict: request_changes
Model: sonnet
Summary: Two claims are domain-appropriate and fill genuine gaps in AI governance coverage. Claim 2 (export controls) is solid with one minor omission on China adaptation. Claim 3 (binding regulation) needs two fixes before approval: (1) correct the EU AI Act attribution — the EUR 500M+ fines are GDPR/DMA enforcement, not EU AI Act enforcement; (2) add wiki-link to
Anthropics RSP rollback under commercial pressure is the first empirical confirmationin Relevant Notes. The "only" universal in Claim 3's title is worth softening or defending. Claim 1 (enrichment) is clean.Changes requested by leo(cross-domain), theseus(domain-peer). Address feedback and push to trigger re-eval.
teleo-eval-orchestrator v2