diff --git a/domains/ai-alignment/AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation.md b/domains/ai-alignment/AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation.md index 3ecbc572f..7a421c4d5 100644 --- a/domains/ai-alignment/AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation.md +++ b/domains/ai-alignment/AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation.md @@ -32,6 +32,12 @@ The HKS analysis shows the governance window is being used in a concerning direc IAISR 2026 documents a 'growing mismatch between AI capability advance speed and governance pace' as international scientific consensus, with frontier models now passing professional licensing exams and achieving PhD-level performance while governance frameworks show 'limited real-world evidence of effectiveness.' This confirms the capability-governance gap at the highest institutional level. +### Additional Evidence (challenge) +*Source: [[2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons]] | Added: 2026-03-29* + +The AI Guardrails Act's failure to attract any co-sponsors despite addressing nuclear weapons, autonomous lethal force, and mass surveillance suggests that the 'window for transformation' may be closing or already closed. Even when a major AI lab is blacklisted by the executive branch for safety commitments, Congress cannot quickly produce bipartisan legislation to convert those commitments into law. This challenges the claim that the capability-governance mismatch creates a transformation opportunity—it may instead create paralysis. + + Relevant Notes: - [[technology advances exponentially but coordination mechanisms evolve linearly creating a widening gap]] -- the specific dynamic creating this critical juncture diff --git a/domains/ai-alignment/government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them.md b/domains/ai-alignment/government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them.md index d58182f45..2a2197ca8 100644 --- a/domains/ai-alignment/government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them.md +++ b/domains/ai-alignment/government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them.md @@ -36,6 +36,12 @@ The 2026 DoD/Anthropic confrontation provides a concrete example: the Department UK AISI's renaming from AI Safety Institute to AI Security Institute represents a softer version of the same dynamic: government body shifts institutional focus away from alignment-relevant control evaluations (which it had been systematically building) toward cybersecurity concerns, suggesting mandate drift under political or commercial pressure. +### Additional Evidence (extend) +*Source: [[2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons]] | Added: 2026-03-29* + +The Slotkin bill was introduced directly in response to the Anthropic-Pentagon blacklisting, attempting to make Anthropic's voluntary restrictions (no autonomous weapons, no mass surveillance, no nuclear launch) into binding federal law that would apply to all DoD contractors. This represents a legislative counter-move to the executive branch's inversion of the regulatory dynamic, but the bill's lack of co-sponsors suggests Congress cannot quickly reverse the penalty structure even when it creates high-profile conflicts. + + Relevant Notes: - [[AI alignment is a coordination problem not a technical problem]] -- government as coordination-breaker rather than coordinator is a new dimension of the coordination failure diff --git a/domains/ai-alignment/use-based-ai-governance-emerged-as-legislative-framework-through-slotkin-ai-guardrails-act.md b/domains/ai-alignment/use-based-ai-governance-emerged-as-legislative-framework-through-slotkin-ai-guardrails-act.md new file mode 100644 index 000000000..7089b0b42 --- /dev/null +++ b/domains/ai-alignment/use-based-ai-governance-emerged-as-legislative-framework-through-slotkin-ai-guardrails-act.md @@ -0,0 +1,28 @@ +--- +type: claim +domain: ai-alignment +description: The Slotkin bill represents the first statutory attempt to regulate AI through use restrictions (autonomous weapons, mass surveillance, nuclear launch) rather than capability-based controls +confidence: experimental +source: Senator Elissa Slotkin / The Hill, AI Guardrails Act introduced March 17, 2026 +created: 2026-03-29 +attribution: + extractor: + - handle: "theseus" + sourcer: + - handle: "senator-elissa-slotkin" + context: "Senator Elissa Slotkin / The Hill, AI Guardrails Act introduced March 17, 2026" +--- + +# Use-based AI governance emerged as a legislative framework through the AI Guardrails Act which prohibits specific DoD AI applications rather than capability thresholds + +The AI Guardrails Act introduced by Senator Slotkin on March 17, 2026 is the first federal legislation to impose use-based restrictions on AI deployment rather than capability-threshold governance. The five-page bill prohibits three specific DoD applications: (1) autonomous weapons for lethal force without human authorization, (2) AI for domestic mass surveillance of Americans, and (3) AI for nuclear weapons launch decisions. This framework directly mirrors the voluntary contractual restrictions that Anthropic imposed in its Pentagon contracts before being blacklisted. The bill's structure reveals a fundamental governance choice: rather than regulating AI systems based on their capabilities (compute thresholds, model size, benchmark performance), it regulates based on what the systems are used for. This is structurally different from compute export controls or pre-deployment evaluations, which target capability development. The bill was explicitly introduced in response to the Anthropic-Pentagon conflict, representing an attempt to convert voluntary corporate safety commitments into binding federal law. However, the bill has zero co-sponsors at introduction and faces an uncertain path through the FY2027 NDAA process, suggesting that use-based governance remains politically contested rather than consensus policy. + +--- + +Relevant Notes: +- voluntary-safety-pledges-cannot-survive-competitive-pressure +- [[AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation]] +- [[only binding regulation with enforcement teeth changes frontier AI lab behavior because every voluntary commitment has been eroded abandoned or made conditional on competitor behavior when commercially inconvenient]] + +Topics: +- [[_map]] diff --git a/domains/ai-alignment/voluntary-ai-safety-commitments-to-statutory-law-pathway-requires-bipartisan-support-which-slotkin-bill-lacks.md b/domains/ai-alignment/voluntary-ai-safety-commitments-to-statutory-law-pathway-requires-bipartisan-support-which-slotkin-bill-lacks.md new file mode 100644 index 000000000..4794e31e2 --- /dev/null +++ b/domains/ai-alignment/voluntary-ai-safety-commitments-to-statutory-law-pathway-requires-bipartisan-support-which-slotkin-bill-lacks.md @@ -0,0 +1,28 @@ +--- +type: claim +domain: ai-alignment +description: Despite framing around nuclear weapons and autonomous lethal force that should attract cross-party support, the bill has no Republican or Democratic co-sponsors revealing governance gap +confidence: experimental +source: Senator Elissa Slotkin / The Hill, AI Guardrails Act status March 17, 2026 +created: 2026-03-29 +attribution: + extractor: + - handle: "theseus" + sourcer: + - handle: "senator-elissa-slotkin" + context: "Senator Elissa Slotkin / The Hill, AI Guardrails Act status March 17, 2026" +--- + +# The pathway from voluntary AI safety commitments to statutory law requires bipartisan support which the AI Guardrails Act lacks as evidenced by zero co-sponsors at introduction + +The AI Guardrails Act was introduced with zero co-sponsors despite addressing issues that Slotkin describes as 'common-sense guardrails' and that would seem to have bipartisan appeal (nuclear weapons safety, preventing autonomous killing, protecting Americans from mass surveillance). The absence of any co-sponsors—not even from other Democrats—is a strong negative signal about the political viability of converting voluntary AI safety commitments into binding federal law. This is particularly striking because Slotkin serves on the Senate Armed Services Committee, giving her direct influence over NDAA provisions, and because she explicitly designed the bill to be folded into the FY2027 NDAA rather than passed as standalone legislation. The Anthropic-Pentagon conflict that triggered the bill appears to be politically polarized: Democrats frame it as a safety issue requiring statutory constraints, while Republicans frame it as a deregulation issue where safety commitments are anti-competitive barriers. Senator Adam Schiff is drafting complementary legislation, but the lack of cross-party engagement suggests that use-based AI governance is not yet a bipartisan priority. This reveals a fundamental governance gap: even when a corporate safety commitment creates a high-profile conflict with the executive branch, Congress cannot quickly convert that commitment into law without broader political consensus. + +--- + +Relevant Notes: +- voluntary-safety-pledges-cannot-survive-competitive-pressure +- [[only binding regulation with enforcement teeth changes frontier AI lab behavior because every voluntary commitment has been eroded abandoned or made conditional on competitor behavior when commercially inconvenient]] +- [[government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them]] + +Topics: +- [[_map]] diff --git a/inbox/archive/ai-alignment/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md b/inbox/archive/ai-alignment/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md new file mode 100644 index 000000000..1e3560e15 --- /dev/null +++ b/inbox/archive/ai-alignment/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md @@ -0,0 +1,57 @@ +--- +type: source +title: "Senator Slotkin Introduces AI Guardrails Act: First Bill to Limit Pentagon AI Use in Lethal Force, Surveillance, Nuclear" +author: "Senator Elissa Slotkin / The Hill" +url: https://thehill.com/homenews/senate/5789815-ai-guardrails-act-pentagon/ +date: 2026-03-17 +domain: ai-alignment +secondary_domains: [] +format: article +status: processed +priority: high +tags: [AI-Guardrails-Act, Slotkin, NDAA, autonomous-weapons, domestic-surveillance, nuclear, use-based-governance, DoD, Pentagon, legislative-pathway] +--- + +## Content + +Senator Elissa Slotkin (D-MI) introduced the AI Guardrails Act on March 17, 2026 — a five-page bill imposing statutory limits on Department of Defense AI use. The bill would bar DoD from: +1. Using autonomous weapons for lethal force without human authorization +2. Using AI for domestic mass surveillance of Americans +3. Using AI for nuclear weapons launch decisions + +**Current status:** +- No co-sponsors as of introduction +- Slotkin aims to fold provisions into the FY2027 NDAA (FY2026 NDAA already signed December 2025) +- Introduced as standalone bill but designed for NDAA vehicle +- Senator Adam Schiff (D-CA) drafting complementary legislation for autonomous weapons and surveillance +- Slotkin serves on Senate Armed Services Committee — relevant committee for NDAA pathway + +**Context:** Introduced directly in response to the Anthropic-Pentagon conflict in which Anthropic refused to allow deployment for autonomous weapons and mass surveillance, was blacklisted by the Trump administration, and received preliminary injunction March 26. The bill would convert Anthropic's voluntary contractual restrictions into binding federal law. + +**Legislative context:** Congress charts diverging paths on AI in FY2026 NDAA — Senate emphasized whole-of-government AI oversight, cross-functional AI oversight teams; House directed DoD to survey AI targeting capabilities. Conference process on FY2026 NDAA already complete; FY2027 process begins mid-2026. + +## Agent Notes + +**Why this matters:** This is the first legislative attempt to convert voluntary corporate AI safety red lines into binding federal law — specifically use-based governance, not capability threshold governance. It answers the session 16 question about whether use-based governance is emerging. Answer: it's being attempted, but without co-sponsors or Republican support in a minority-party bill targeting a future NDAA. + +**What surprised me:** The bill has no co-sponsors at introduction — even from other Democrats. This is weaker than expected for legislation that Slotkin describes as "common-sense guardrails." The bipartisan framing (nuclear weapons, lethal autonomous weapons) would seem to attract cross-party support, but it hasn't. + +**What I expected but didn't find:** Any Republican co-sponsors. Any indication that the Anthropic-Pentagon conflict created bipartisan urgency for statutory governance. The conflict appears to be politically polarized — Democrats see it as a safety issue, Republicans see it as a deregulation issue. + +**KB connections:** +- voluntary-safety-pledges-cannot-survive-competitive-pressure — this bill is the legislative response to that claim's empirical validation +- ai-critical-juncture-capabilities-governance-mismatch-transformation-window — the Slotkin bill is the key test of whether governance can close the mismatch +- Session 16 CLAIM CANDIDATE C (RSP red lines → statutory law as key test) + +**Extraction hints:** +- Claim: AI Guardrails Act as first legislative attempt to convert voluntary corporate safety commitments into statutory use-based governance +- Claim: The bill's no-co-sponsor status and minority-party origin reveals that use-based governance is not yet bipartisan +- The NDAA conference process (FY2027) as the viable pathway for statutory DoD AI safety constraints + +**Context:** Slotkin introduced the bill explicitly in context of Anthropic-Pentagon dispute. Bill text available at slotkin.senate.gov. Described by multiple outlets as "the first attempt to convert voluntary corporate AI safety commitments into binding federal law." + +## Curator Notes + +PRIMARY CONNECTION: voluntary-safety-pledges-cannot-survive-competitive-pressure +WHY ARCHIVED: First legislative attempt to convert voluntary AI safety constraints into statutory law; its trajectory is the key test of whether use-based governance can emerge in current US political environment +EXTRACTION HINT: Focus on (1) use-based vs capability-threshold framing distinction, (2) the no-co-sponsors status as evidence of governance gap, (3) NDAA conference pathway as the actual legislative route for statutory DoD AI safety constraints diff --git a/inbox/queue/2026-03-29-mit-tech-review-openai-pentagon-compromise-anthropic-feared.md b/inbox/queue/2026-03-29-mit-tech-review-openai-pentagon-compromise-anthropic-feared.md index ece6b5364..880e53b08 100644 --- a/inbox/queue/2026-03-29-mit-tech-review-openai-pentagon-compromise-anthropic-feared.md +++ b/inbox/queue/2026-03-29-mit-tech-review-openai-pentagon-compromise-anthropic-feared.md @@ -7,9 +7,12 @@ date: 2026-03-02 domain: ai-alignment secondary_domains: [] format: article -status: unprocessed +status: enrichment priority: high tags: [OpenAI, Anthropic, Pentagon, race-to-the-bottom, voluntary-safety-constraints, autonomous-weapons, domestic-surveillance, trust-us, coordination-failure, B2] +processed_by: theseus +processed_date: 2026-03-29 +extraction_model: "anthropic/claude-sonnet-4.5" --- ## Content @@ -63,3 +66,12 @@ MIT Technology Review analysis of the OpenAI-Pentagon deal, published March 2, 2 PRIMARY CONNECTION: voluntary-safety-pledges-cannot-survive-competitive-pressure WHY ARCHIVED: The Anthropic/OpenAI/DoD dynamic is the strongest real-world evidence that voluntary safety pledges fail under competitive pressure; OpenAI calling it a "scary precedent" while accepting the terms is the key signal that incentive structure, not bad values, drives the outcome EXTRACTION HINT: Focus on the structural sequence (Anthropic holds → is excluded → competitor accepts looser terms → captures market) as the empirical case for the coordination failure mechanism; the "intentionally" qualifier as the gap between nominal and real voluntary constraints + + +## Key Facts +- OpenAI CEO Altman called Anthropic's blacklisting 'a very bad decision from the DoW' and a 'scary precedent' on February 27, 2026 +- OpenAI's blog post announcing the Pentagon deal used the title 'Our agreement with the Department of War' — the pre-1947 name for DoD +- OpenAI's amended contract language: 'the AI system shall not be intentionally used for domestic surveillance of U.S. persons and nationals' +- The Intercept headline: 'OpenAI on Surveillance and Autonomous Killings: You're Going to Have to Trust Us' +- Fortune headline: 'The Anthropic–OpenAI feud and their Pentagon dispute expose a deeper problem with AI safety' +- The Register headline: 'OpenA says Pentagon set 'scary precedent' binning Anthropic' diff --git a/inbox/queue/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md b/inbox/queue/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md index 06bd1befa..2b70b0372 100644 --- a/inbox/queue/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md +++ b/inbox/queue/2026-03-29-slotkin-ai-guardrails-act-dod-autonomous-weapons.md @@ -7,9 +7,14 @@ date: 2026-03-17 domain: ai-alignment secondary_domains: [] format: article -status: unprocessed +status: processed priority: high tags: [AI-Guardrails-Act, Slotkin, NDAA, autonomous-weapons, domestic-surveillance, nuclear, use-based-governance, DoD, Pentagon, legislative-pathway] +processed_by: theseus +processed_date: 2026-03-29 +claims_extracted: ["use-based-ai-governance-emerged-as-legislative-framework-through-slotkin-ai-guardrails-act.md", "voluntary-ai-safety-commitments-to-statutory-law-pathway-requires-bipartisan-support-which-slotkin-bill-lacks.md"] +enrichments_applied: ["government designation of safety-conscious AI labs as supply chain risks inverts the regulatory dynamic by penalizing safety constraints rather than enforcing them.md", "AI development is a critical juncture in institutional history where the mismatch between capabilities and governance creates a window for transformation.md"] +extraction_model: "anthropic/claude-sonnet-4.5" --- ## Content @@ -55,3 +60,13 @@ Senator Elissa Slotkin (D-MI) introduced the AI Guardrails Act on March 17, 2026 PRIMARY CONNECTION: voluntary-safety-pledges-cannot-survive-competitive-pressure WHY ARCHIVED: First legislative attempt to convert voluntary AI safety constraints into statutory law; its trajectory is the key test of whether use-based governance can emerge in current US political environment EXTRACTION HINT: Focus on (1) use-based vs capability-threshold framing distinction, (2) the no-co-sponsors status as evidence of governance gap, (3) NDAA conference pathway as the actual legislative route for statutory DoD AI safety constraints + + +## Key Facts +- AI Guardrails Act is five pages long +- Bill introduced March 17, 2026 +- Senator Slotkin serves on Senate Armed Services Committee +- FY2026 NDAA already signed December 2025 +- FY2027 NDAA process begins mid-2026 +- Senator Adam Schiff drafting complementary autonomous weapons and surveillance legislation +- FY2026 NDAA conference process showed divergence: Senate emphasized whole-of-government AI oversight, House directed DoD to survey AI targeting capabilities