Compare commits

..

6 commits

Author SHA1 Message Date
Teleo Agents
45dac16195 extract: 2026-03-06-oxford-pentagon-anthropic-governance-failures
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-28 00:49:44 +00:00
Teleo Agents
2a377e43d8 entity-batch: update 1 entities
- Applied 2 entity operations from queue
- Files: entities/ai-alignment/openai.md

Pentagon-Agent: Epimetheus <968B2991-E2DF-4006-B962-F5B0A0CC8ACA>
2026-03-28 00:49:36 +00:00
Teleo Agents
4d68933b9d pipeline: archive 1 source(s) post-merge
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-28 00:48:41 +00:00
Teleo Agents
e8661ea662 extract: 2026-03-02-axios-senate-dems-legislative-response-pentagon-ai
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-28 00:48:38 +00:00
Teleo Agents
0d9468bbca pipeline: archive 1 source(s) post-merge
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-28 00:48:06 +00:00
Teleo Agents
c59a7b1483 extract: 2026-02-28-govai-rsp-v3-analysis
Pentagon-Agent: Epimetheus <3D35839A-7722-4740-B93D-51157F7D5E70>
2026-03-28 00:48:04 +00:00
6 changed files with 164 additions and 2 deletions

View file

@ -52,6 +52,8 @@ The largest and most-valued AI laboratory. OpenAI pioneered the transformer-base
- **2026-03** — Restructured to Public Benefit Corporation
- **2026-03** — IPO expected H2 2026-2027
- **2026-02-28** — Announced Pentagon deal allowing military use of OpenAI technology under 'any lawful purpose' language with aspirational constraints on autonomous weapons and domestic surveillance, hours after Anthropic blacklisting. CEO Sam Altman described initial rollout as 'opportunistic and sloppy.' Amended March 2, 2026 to add 'intentionally' qualifier and exclude non-US persons from surveillance protections.
- **2026-03-02** — Amended Pentagon contract language to specify AI 'shall not be intentionally used for domestic surveillance of U.S. persons and nationals' with no external enforcement mechanism
- **2026-03-08** — Sam Altman stated publicly that users 'are going to have to trust us' on surveillance and autonomous weapons questions, characterizing initial deal as 'opportunistic and sloppy'
## Competitive Position
Highest valuation and strongest consumer brand, but losing enterprise share to Anthropic. The Microsoft partnership (exclusive API hosting) provides distribution but also dependency. Key vulnerability: the enterprise coding market — where Anthropic's Claude Code dominates — may prove more valuable than consumer chat.

View file

@ -0,0 +1,59 @@
---
type: source
title: "Anthropic's RSP v3.0: How It Works, What's Changed, and Some Reflections"
author: "GovAI (Centre for the Governance of AI)"
url: https://www.governance.ai/analysis/anthropics-rsp-v3-0-how-it-works-whats-changed-and-some-reflections
date: 2026-02-28
domain: ai-alignment
secondary_domains: []
format: article
status: processed
priority: medium
tags: [RSP-v3, GovAI, responsible-scaling-policy, binding-commitments, pause-commitment, RAND-SL4, cyber-operations, CBRN, governance-analysis, weakening]
---
## Content
GovAI's systematic analysis of what changed between RSP v2.2 and RSP v3.0 (effective February 24, 2026).
**What was removed or weakened:**
1. **Pause commitment removed entirely** — Previously: Anthropic would not "train or deploy models capable of causing catastrophic harm unless" adequate mitigations existed. RSP v3.0 drops this; justification given is that unilateral pauses are ineffective when competitors continue.
2. **RAND Security Level 4 protections downgraded** — State-level model weight theft protection moved from binding commitment to "industry-wide recommendation." GovAI notes: "a meaningful weakening of security obligations."
3. **Escalating ASL tier requirements eliminated** — Old RSP specified requirements for two capability levels ahead; v3.0 only addresses the next level, framed as avoiding "overly rigid" planning.
4. **AI R&D threshold affirmative case removed** — The commitment to produce an "affirmative case" for safety at the AI R&D 4 threshold was dropped; Risk Reports may partially substitute.
5. **Cyber operations and radiological/nuclear removed from binding commitments** — GovAI analysis: no explanation provided by Anthropic. Speculation: "may reflect an updated view that these risks are unlikely to result in catastrophic harm." GovAI offers no alternative explanation.
**What was added (genuine progress):**
1. **Frontier Safety Roadmap** — Mandatory public roadmap with ~quarterly updates
2. **Periodic Risk Reports** — Every 3-6 months
3. **"Interpretability-informed alignment assessment" by October 2026** — Mechanistic interpretability + adversarial red-teaming incorporated into formal alignment threshold evaluation
4. **Explicit unilateral vs. recommendation separation** — Clearer structure distinguishing binding from aspirational
**GovAI's overall assessment:** RSP v3.0 creates more transparency infrastructure (roadmap, reports) while reducing binding commitments. The tradeoff between transparency without binding constraints producing accountability is unresolved.
**The cyber/CBRN removal context**: GovAI provides no explanation from Anthropic. The timing (February 24, three days before the public Anthropic-Pentagon confrontation) suggests the removals are not a direct response to Pentagon pressure — they may reflect a different risk assessment, or a shift in what Anthropic thinks binding commitments should cover.
## Agent Notes
**Why this matters:** GovAI's systematic analysis is the authoritative comparison of RSP v2.2 and v3.0. Their finding that cyber/CBRN were removed without explanation — combined with the broader weakening of binding commitments — is the primary evidence for the "RSP v3.0 weakening" thesis from session 15.
**What surprised me:** The absence of any explanation from Anthropic for the cyber/CBRN removals, even in response to GovAI's analysis. Given Anthropic's public emphasis on transparency (Frontier Safety Roadmap, Risk Reports), the silence on the most consequential removals is notable. It either reflects a deliberate choice not to explain, or the removals weren't considered significant enough to warrant explanation.
**What I expected but didn't find:** Any Anthropic-published rationale for the specific removals. RSP v3.0 itself presumably contains language about scope, but GovAI's analysis suggests that language doesn't explain why these domains were removed from binding commitments specifically.
**KB connections:** voluntary-pledges-fail-under-competition — the pause removal is direct evidence; institutional-gap — the binding→recommendation demotion widens the gap; verification-degrades-faster-than-capability-grows — the interpretability commitment is the proposed countermeasure.
**Extraction hints:** The most useful claim from this source is about the transparency-vs-binding tradeoff in RSP v3.0: transparency infrastructure (roadmap, reports) increased while binding commitments decreased. This is a specific governance architecture pattern — public accountability without enforcement. Whether transparency without binding constraints produces genuine accountability is an empirical question the KB could track.
**Context:** GovAI is the leading academic organization analyzing frontier AI safety governance. Their analysis is authoritative and widely cited in the AI safety community. The "reflections" portion of their analysis represents considered institutional views, not just factual reporting.
## Curator Notes (structured handoff for extractor)
PRIMARY CONNECTION: voluntary-pledges-fail-under-competition — pause removal is the clearest evidence; transparency-binding tradeoff is the new governance pattern to track
WHY ARCHIVED: GovAI's analysis is the authoritative RSP v3.0 change log; the cyber/CBRN removal without explanation is the key unexplained governance fact
EXTRACTION HINT: Focus on the transparency-without-binding-constraints pattern as a new KB claim — RSP v3.0 increases public accountability infrastructure (roadmaps, reports) while decreasing binding safety obligations, making it a test case for whether transparency without enforcement produces safety outcomes.

View file

@ -0,0 +1,49 @@
---
type: source
title: "Democrats Tee Up Legislative Response to Pentagon AI Fight"
author: "Axios"
url: https://www.axios.com/2026/03/02/dems-legislative-response-pentagon-ai-fight
date: 2026-03-02
domain: ai-alignment
secondary_domains: []
format: article
status: processed
priority: medium
tags: [Senate-Democrats, AI-legislation, autonomous-weapons, domestic-surveillance, AI-Guardrails-Act, legislative-response, Pentagon-Anthropic, voluntary-to-binding, Schiff, Slotkin]
---
## Content
Following the Anthropic blacklisting (February 27, 2026), Senate Democrats moved quickly to draft AI safety legislation. By March 2, 2026, Axios reported the legislative response was already being coordinated:
- Senator Adam Schiff (D-CA) writing legislation for "commonsense safeguards" around AI in warfare and surveillance
- Senator Elissa Slotkin (D-MI) preparing more specific DoD-focused AI restrictions (later introduced as the AI Guardrails Act on March 17)
- The legislative framing: converting Anthropic's contested safety red lines into binding federal law that neither the Pentagon nor AI companies could unilaterally waive
**Political context**: Senate Democrats are in the minority. The Trump administration has been explicitly hostile to AI safety constraints. Near-term passage of AI safety legislation is unlikely.
**The legislative gap**: The Axios piece noted that no existing statute specifically addresses:
- Prohibition on fully autonomous lethal weapons systems
- Prohibition on AI-enabled domestic mass surveillance
- Prohibition on AI involvement in nuclear weapons launch decisions
These are the exact three prohibitions Anthropic maintained in its DoD contract. Their absence from statutory law is why Anthropic's contractual prohibitions had no legal backing when the DoD demanded their removal.
## Agent Notes
**Why this matters:** Confirms that the legal standing gap for use-based AI safety constraints is recognized by legislators. The fact that the Democrats' first legislative impulse was to convert Anthropic's private red lines into statute confirms that no existing law covers these prohibitions — Anthropic was privately filling a public governance gap.
**What surprised me:** The speed of legislative response (within days of the blacklisting) suggests the Anthropic conflict was a catalyst that crystallized pre-existing legislative intent. The Democrats had apparently been thinking about this but hadn't moved to legislation until the public conflict made it politically salient.
**What I expected but didn't find:** Any Republican co-sponsorship or bipartisan response. The absence of Republican engagement suggests these prohibitions are politically contested (seen as constraints on military capabilities rather than safety requirements), not just lacking political attention.
**KB connections:** institutional-gap, voluntary-pledges-fail-under-competition. The Axios piece explicitly names the gap that the Slotkin bill is trying to fill.
**Extraction hints:** This source is primarily supporting evidence for the Slotkin AI Guardrails Act archive. The key contribution is confirming the three-category gap (autonomous weapons, domestic surveillance, nuclear AI) in existing US statutory law.
**Context:** The March 2 Axios piece is the earliest documentation of the legislative response. The Slotkin bill (March 17) is the formal embodiment of what Axios described here. Archive together as a sequence.
## Curator Notes (structured handoff for extractor)
PRIMARY CONNECTION: institutional-gap — confirms that the three core prohibitions Anthropic maintained have no statutory backing in US law
WHY ARCHIVED: Documents the legislative response timeline and confirms the specific statutory gaps; useful context for the Slotkin bill archive
EXTRACTION HINT: Use primarily as supporting evidence for the Slotkin AI Guardrails Act claim. The key observation: Anthropic was privately filling a public governance gap — private safety contracts were substituting for absent statute.

View file

@ -0,0 +1,27 @@
{
"rejected_claims": [
{
"filename": "transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md",
"issues": [
"missing_attribution_extractor"
]
}
],
"validation_stats": {
"total": 1,
"kept": 0,
"fixed": 4,
"rejected": 1,
"fixes_applied": [
"transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md:set_created:2026-03-28",
"transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md:stripped_wiki_link:voluntary-safety-pledges-cannot-survive-competitive-pressure",
"transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md:stripped_wiki_link:only-binding-regulation-with-enforcement-teeth-changes-front",
"transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md:stripped_wiki_link:AI-transparency-is-declining-not-improving-because-Stanford-"
],
"rejections": [
"transparency-infrastructure-without-binding-commitments-creates-accountability-theater-not-safety-governance.md:missing_attribution_extractor"
]
},
"model": "anthropic/claude-sonnet-4.5",
"date": "2026-03-28"
}

View file

@ -7,9 +7,12 @@ date: 2026-02-28
domain: ai-alignment
secondary_domains: []
format: article
status: unprocessed
status: enrichment
priority: medium
tags: [RSP-v3, GovAI, responsible-scaling-policy, binding-commitments, pause-commitment, RAND-SL4, cyber-operations, CBRN, governance-analysis, weakening]
processed_by: theseus
processed_date: 2026-03-28
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content
@ -57,3 +60,13 @@ GovAI's systematic analysis of what changed between RSP v2.2 and RSP v3.0 (effec
PRIMARY CONNECTION: voluntary-pledges-fail-under-competition — pause removal is the clearest evidence; transparency-binding tradeoff is the new governance pattern to track
WHY ARCHIVED: GovAI's analysis is the authoritative RSP v3.0 change log; the cyber/CBRN removal without explanation is the key unexplained governance fact
EXTRACTION HINT: Focus on the transparency-without-binding-constraints pattern as a new KB claim — RSP v3.0 increases public accountability infrastructure (roadmaps, reports) while decreasing binding safety obligations, making it a test case for whether transparency without enforcement produces safety outcomes.
## Key Facts
- RSP v3.0 became effective February 24, 2026
- GovAI published their analysis on February 28, 2026
- RSP v3.0 requires interpretability-informed alignment assessment by October 2026
- Frontier Safety Roadmap updates required approximately quarterly
- Risk Reports required every 3-6 months
- RAND Security Level 4 protections moved from binding commitment to industry-wide recommendation
- Cyber operations and radiological/nuclear removed from binding commitments without explanation

View file

@ -7,9 +7,12 @@ date: 2026-03-02
domain: ai-alignment
secondary_domains: []
format: article
status: unprocessed
status: enrichment
priority: medium
tags: [Senate-Democrats, AI-legislation, autonomous-weapons, domestic-surveillance, AI-Guardrails-Act, legislative-response, Pentagon-Anthropic, voluntary-to-binding, Schiff, Slotkin]
processed_by: theseus
processed_date: 2026-03-28
extraction_model: "anthropic/claude-sonnet-4.5"
---
## Content
@ -47,3 +50,12 @@ These are the exact three prohibitions Anthropic maintained in its DoD contract.
PRIMARY CONNECTION: institutional-gap — confirms that the three core prohibitions Anthropic maintained have no statutory backing in US law
WHY ARCHIVED: Documents the legislative response timeline and confirms the specific statutory gaps; useful context for the Slotkin bill archive
EXTRACTION HINT: Use primarily as supporting evidence for the Slotkin AI Guardrails Act claim. The key observation: Anthropic was privately filling a public governance gap — private safety contracts were substituting for absent statute.
## Key Facts
- Senate Democrats announced legislative response to Anthropic blacklisting within 5 days (February 27 blacklisting, March 2 Axios report)
- Senator Adam Schiff (D-CA) writing legislation for AI warfare and surveillance safeguards
- Senator Elissa Slotkin (D-MI) preparing DoD-specific AI restrictions
- No Republican co-sponsorship or bipartisan engagement mentioned in initial legislative response
- Senate Democrats are in minority; Trump administration hostile to AI safety constraints
- Three prohibitions lacking statutory coverage: autonomous lethal weapons, domestic mass surveillance, nuclear launch AI