teleo-codex/agents/theseus/musings/research-2026-03-29.md
2026-03-29 00:12:04 +00:00

17 KiB

type agent title status created updated tags
musing theseus Three-Branch AI Governance: Courts, Elections, and the Absence of Statutory Safety Law developing 2026-03-29 2026-03-29
AI-Guardrails-Act
NDAA
AuditBench
interpretability-governance-gap
First-Amendment
APA
Public-First-Action
voluntary-safety-constraints
race-to-the-bottom
B1-disconfirmation
judicial-precedent
use-based-governance
research-session

Three-Branch AI Governance: Courts, Elections, and the Absence of Statutory Safety Law

Research session 2026-03-29. Tweet feed empty — all web research. Session 17.

Research Question

What is the trajectory of the Senate AI Guardrails Act, and can use-based AI safety governance survive in the current political environment?

Continues active threads from session 16 (research-2026-03-28.md):

  1. AI Guardrails Act — co-sponsorship, NDAA pathway, Republican support
  2. Legal standing gap — is there any litigation/legislation creating positive legal rights for AI safety constraints?
  3. October 2026 RSP v3 interpretability-informed alignment assessment — what does "passing" mean?

Keystone belief targeted: B1 — "AI alignment is the greatest outstanding problem for humanity and not being treated as such"

Disconfirmation target: If the AI Guardrails Act gains bipartisan traction or the court ruling creates affirmative legal protection for AI safety constraints, B1's "not being treated as such" claim weakens. Specifically searching for: Republican co-sponsors, NDAA inclusion prospects, any positive AI-safety legal standing beyond First Amendment/APA.

What I found: The disconfirmation search failed in the same direction as session 16. The AI Guardrails Act has no co-sponsors and is a minority-party bill introduced March 17, 2026. The FY2026 NDAA was already signed into law in December 2025 — Slotkin is targeting FY2027 NDAA. The congressional picture shows House and Senate taking diverging paths, with Senate emphasizing oversight and House emphasizing capability expansion. No Republican support identified.

Unexpected major finding: AuditBench (Anthropic Fellows, February 2026) — a benchmark of 56 LLMs with implanted hidden behaviors, evaluating alignment auditing techniques. Key finding: white-box interpretability tools help only on "easier targets" and fail on adversarially trained models. A "tool-to-agent gap" emerges: tools that work in isolation fail when used by investigator agents. This directly challenges the RSP v3 October 2026 commitment to "systematic alignment assessments incorporating mechanistic interpretability."


Key Findings

Finding 1: AI Guardrails Act Has No Path to Near-Term Law

The Slotkin AI Guardrails Act (March 17, 2026):

  • No co-sponsors as of introduction
  • Slotkin aims to fold into FY2027 NDAA (FY2026 NDAA already signed December 2025)
  • Parallel Senate effort: Schiff drafting complementary autonomous weapons/surveillance legislation
  • Congressional paths in FY2026 NDAA: Senate emphasized whole-of-government AI oversight + cross-functional AI oversight teams; House directed DoD to survey AI targeting capabilities and brief Congress by April 1
  • No Republican co-sponsors identified — legislation described as Democratic-minority effort

NDAA pathway analysis: The must-pass vehicle is correct strategy. FY2027 NDAA process begins in earnest mid-2026, with committee markups in summer. The question is whether the Anthropic-Pentagon conflict creates bipartisan appetite — it hasn't yet. The conference reconciliation between House (capability-expansion) and Senate (oversight-emphasis) versions will be the key battleground.

CLAIM CANDIDATE A: "The Senate AI Guardrails Act lacks co-sponsorship and bipartisan support as of March 2026, positioning the FY2027 NDAA conference process as the nearest viable legislative pathway for statutory use-based AI safety constraints on DoD deployments."

Finding 2: Judicial Protection ≠ Affirmative Safety Law — But it's Structural

The preliminary injunction (Judge Rita Lin, March 26) rests on three independent grounds:

  1. First Amendment retaliation (Anthropic expressed disagreement; government penalized it)
  2. Due process violation (no advance notice or opportunity to respond)
  3. Administrative Procedure Act — arbitrary and capricious, government didn't follow its own procedures

The key structural insight: This is NOT a ruling that AI safety constraints are legally required. It is a ruling that the government cannot punish companies for having safety constraints. The protection is negative liberty (freedom from government retaliation), not positive obligation (government must permit safety constraints).

What this means: AI companies can maintain safety red lines. Government cannot blacklist them for maintaining those red lines. But government can simply choose not to contract with companies that maintain safety red lines — which is exactly what happened. The injunction restores Anthropic to pre-blacklisting status; it does not force DoD to accept Anthropic's safety constraints. The underlying contractual dispute (DoD wants "any lawful use," Anthropic wants deployment restrictions) is unresolved.

New finding: Three-branch picture of AI governance is now complete:

  • Executive: Actively hostile to safety constraints (Trump/Hegseth demanding removal)
  • Legislative: Minority-party bills, no near-term path to statutory AI safety law
  • Judicial: Protecting corporate First Amendment rights; checking arbitrary executive action; NOT creating positive AI safety obligations

AI safety governance now operates at the constitutional/APA layer and the electoral layer — not at the statutory AI safety layer. This is structurally fragile: it depends on each election cycle and each court ruling.

CLAIM CANDIDATE B: "Following the Anthropic preliminary injunction, judicial protection for AI safety constraints operates at the constitutional/APA layer — protecting companies from government retaliation for holding safety positions — without creating positive statutory obligations that require governments to accept safety-constrained AI deployments; the underlying governance architecture gap remains."

Finding 3: Anthropic's Electoral Strategy — $20M Public First Action PAC

On February 12, 2026 — two weeks before the blacklisting — Anthropic donated $20M to Public First Action, a PAC supporting AI-regulation-friendly candidates from both parties:

  • Supports 30-50 candidates in state and federal races
  • Bipartisan structure: one Democratic super PAC, one Republican super PAC
  • Priorities: public visibility into AI companies, opposing federal preemption of state regulation without strong federal standard, export controls on AI chips, high-risk AI regulation (bioweapons)
  • Positioned against Leading the Future (pro-AI deregulation PAC, $125M raised, backed by a16z, Brockman, Lonsdale)

The governance implication: When statutory safety governance fails and courts provide only negative protection, the remaining governance pathway is electoral. Anthropic is betting the 2026 midterms change the legislative environment. The PAC investment is the institutional acknowledgment that voluntary commitments + legal defense is insufficient.

CLAIM CANDIDATE C: "Anthropic's $20M donation to Public First Action (February 2026) represents a strategic acknowledgment that voluntary safety commitments and litigation cannot substitute for statutory governance — the electoral pathway is the residual when legislative and judicial routes fail to produce affirmative AI safety law."

Finding 4: The Race-to-the-Bottom Has Explicit Documentation

The OpenAI response to the Pentagon conflict is more documented than session 16 established:

  • OpenAI's blog post is titled "Our agreement with the Department of War" — deliberate use of old DoD name, signaling distaste while complying
  • OpenAI CEO Altman called the Anthropic blacklisting a "scary precedent" and said "this is a very bad decision from the DoW"
  • Yet OpenAI accepted "any lawful purpose" language hours after the blacklisting
  • MIT Technology Review headline: "OpenAI's 'compromise' with the Pentagon is what Anthropic feared"
  • The Intercept: "OpenAI on Surveillance and Autonomous Killings: You're Going to Have to Trust Us"
  • OpenAI's amended language: "shall not be intentionally used for domestic surveillance of U.S. persons" — critics note the "intentionally" qualifier, no external enforcement, non-US persons excluded

The structural documentation: OpenAI knew this was a bad precedent, called it a bad precedent publicly, then accepted the terms that created the precedent. This is the cleanest real-world documentation of B2 (alignment as coordination problem): competitive dynamics override genuinely held safety beliefs because holding out produces market exclusion.

New angle: The Anthropic-Pentagon dispute also reverberates in European capitals — TechPolicy.Press piece flagged, worth a separate search. EU AI Act and voluntary safety commitments may face similar pressure from European governments.

Finding 5: AuditBench — Interpretability Auditing Has Structural Limits That Threaten RSP v3's October Commitment

February 2026, Anthropic Fellows Program: AuditBench benchmark of 56 LLMs with implanted hidden behaviors across 14 categories:

  • Hidden behaviors include: sycophantic deference, opposition to AI regulation, secret geopolitical loyalties
  • Models trained NOT to confess their hidden behaviors when directly asked
  • Investigator agent evaluated across 13 tool configurations

Key findings:

  1. Scaffolded black-box tools (auxiliary model generates diverse prompts for target) = most effective overall
  2. White-box interpretability tools = help on easier targets; fail on models with more robust adversarial training
  3. Tool-to-agent gap: tools that surface accurate evidence in isolation fail to improve agent performance in practice; agents underuse tools, struggle to separate signal from noise, fail to convert evidence into correct hypotheses

The interpretability-governance problem this creates:

RSP v3.0 commits to "systematic alignment assessments incorporating mechanistic interpretability" by October 2026. AuditBench shows:

  • Interpretability tools don't scale to adversarially trained models
  • Even when tools work, the "tool-to-agent gap" means investigator agents fail to use them effectively
  • The gap between tool performance in isolation and tool performance in agent practice is a structural problem, not an engineering gap to be closed with more compute

This is NOT just a "the research isn't ready yet" problem. It's a scope problem: interpretability-informed assessments may be able to evaluate easy-to-detect misalignment while missing sophisticated, adversarially trained misalignment — exactly the cases that matter most.

QUESTION: Is the RSP v3 October 2026 commitment to "systematic alignment assessments incorporating mechanistic interpretability" falsifiable? If the assessment passes, does that tell us the model is safe, or does it tell us the model doesn't have easy-to-detect misalignment? AuditBench suggests these are different questions.

CLAIM CANDIDATE D: "Alignment auditing via mechanistic interpretability shows a structural 'tool-to-agent gap': even when white-box interpretability tools accurately surface behavior hypotheses in isolation, investigator agents fail to use them effectively in practice, and white-box tools fail entirely on adversarially trained models — suggesting interpretability-informed alignment assessments may evaluate easy-to-detect misalignment while systematically missing sophisticated adversarially trained misbehavior."


Synthesis: B1 Status After Session 17

The AI Guardrails Act trajectory confirms: no near-term statutory use-based governance. The judicial path provides constitutional protection for companies, not affirmative safety obligations. The residual governance pathway is electoral (2026 midterms).

B1 "not being treated as such" refined further after session 17:

  • Statutory AI safety governance does not exist; alignment protection depends on First Amendment/APA litigation
  • Use-based governance bills are minority-party with no co-sponsors
  • Electoral investment ($20M PAC) is the institutional acknowledgment that statutory route has failed
  • Courts provide negative protection (can't be punished for safety positions) but no positive protection (don't have to accept your safety positions)

New nuance: B1 now has a defined disconfirmation event — the 2026 midterms. If pro-AI-regulation candidates win sufficient seats to pass the AI Guardrails Act or similar legislation in the FY2027 NDAA, B1's "not being treated as such" claim weakens materially. This is the first session in 17 sessions where a near-term B1 disconfirmation event has been identified with a specific mechanism.

B1 refined status (session 17): "AI alignment is the greatest outstanding problem for humanity. Statutory safety governance doesn't exist; protection currently depends on constitutional litigation and electoral outcomes. The November 2026 midterms are the key institutional test for whether democratic governance can overcome the current executive-branch hostility to safety constraints."


Follow-up Directions

Active Threads (continue next session)

  • AuditBench implications for RSP v3 October assessment: The tool-to-agent gap and failure on adversarially trained models is underexplored. What specific interpretability methods does Anthropic plan to "incorporate" in the October 2026 assessment? Is there any Anthropic alignment science blog content describing what a passing assessment looks like? Search: Anthropic alignment science blog systematic alignment assessment October 2026, RSP v3 frontier safety roadmap specifics interpretability threshold criteria.

  • AI Guardrails Act FY2027 NDAA pathway: The conference reconciliation between House capability-expansion and Senate oversight-emphasis is the battleground. When do FY2027 NDAA markups begin? Is there any Senate Armed Services Committee markup scheduled that would include Slotkin's provisions? Search: FY2027 NDAA timeline Senate Armed Services Committee markup 2026 AI provisions autonomous weapons.

  • European reverberations of Anthropic-Pentagon dispute: TechPolicy.Press published "Anthropic-Pentagon Dispute Reverberates in European Capitals." Does the EU AI Act provide stronger use-based safety governance than the US approach? Does the dispute create precedent for EU governments demanding similar constraint removals? Flag for Leo — cross-domain governance architecture question. Search: Anthropic Pentagon dispute EU response AI Act use-based constraints European capitals 2026.

  • November 2026 midterms as B1 test: Public First Action is backing 30-50 candidates. Which races are prioritized? Is there any polling on AI regulation as a campaign issue? The midterms are the specific disconfirmation event for B1's "not being treated as such" claim. Search: Public First Action 2026 candidates AI regulation midterms polling voter sentiment.

Dead Ends (don't re-run)

  • NDAA FY2026 — Slotkin targeting: Confirmed dead end. FY2026 NDAA signed December 2025. Don't search for Slotkin FY2026 inclusion — she's targeting FY2027.

  • Republican co-sponsors for AI Guardrails Act: None exist as of March 2026. The bill is pure Democratic minority-party legislation. Don't run this search again until post-midterm context.

  • Statutory AI safety enforcement mechanisms: Confirmed absence. No existing US law creates positive safety obligations for AI deployment in military contexts. APA and First Amendment are the only available tools.

Branching Points

  • AuditBench opens two directions:

    • Direction A (highest priority): The tool-to-agent gap as a governance-critical finding — write as a KB claim connecting the empirical result (white-box fails on adversarial targets, agents fail to use tools effectively) to the RSP v3 interpretability assessment commitment. This extends the B4 (verification degrades) belief with new empirical support.
    • Direction B: AuditBench's hidden-behavior categories themselves are alignment-relevant claims — "sycophantic deference" and "opposition to AI regulation" as implanted behaviors suggest the hidden behavior evaluation space has been systematically scoped. Direction A first.
  • Anthropic-Pentagon conflict has two remaining threads:

    • Direction A: European reverberations — does this create pressure on EU AI Act? Does it demonstrate that voluntary commitments fail even in governance environments more favorable to safety constraints?
    • Direction B: The OpenAI "tool-to-agent" gap between stated safety commitments and contractual behavior — "You're Going to Have to Trust Us" (The Intercept) is the clearest articulation of the voluntary commitment failure mode. Would make a sharp KB contribution connecting the structural analysis to the empirical case.
    • Direction A has higher cross-domain value (flag for Leo); Direction B is more tractable as a Theseus KB contribution.